Jan 2, 2024
Had a doubt that when we fine tune a model say like mistral on certain insteuctions, and we also tokenize well by setting proper bos & eos and padding =eos. Still when the model is trained using LoRA, it isn’t able to properly generate the eos token. Hence it keeps generating words endlessly. Note: I did confirm to add eos token during training too.