--

Had a doubt that when we fine tune a model say like mistral on certain insteuctions, and we also tokenize well by setting proper bos & eos and padding =eos. Still when the model is trained using LoRA, it isn’t able to properly generate the eos token. Hence it keeps generating words endlessly. Note: I did confirm to add eos token during training too.

--

--

Niranjan Akella
Niranjan Akella

Written by Niranjan Akella

A scientist by heart and a passionate researcher in my field of expertise. I love philosophy.

Responses (1)