Replies: 5 comments 3 replies
-
The learning rate is something that YOU are controlling. In your setup it's essentially zero because you are using the noam scheduler with a very low base learning rate. Either disable noam decay, or increase your base learning rate factor. The provided llama2 finetuning recipe is far from perfect, but you can see there it's not using noam decay. Alternatively, the "standard" transformer recipes (wmt17, wiki103) use noam decay, but with a higher base learning rate. As I already advised earlier, you might want to start from a working setup, and then start changing things iteratively (e.g. standard llama2 finetuning recipe, then change to a "simple" translation task, then to your low resource task). |
Beta Was this translation helpful? Give feedback.
-
Post here the first 5 lines of your /src-tgt-prompts.enty file |
Beta Was this translation helpful? Give feedback.
-
Taking the second example (the first one is tricky with just short prompt)
For inference remive the space before user and assistant. Having said that, it may not work properly since the only tahitian the model will know is from your finetuning data so it is not so easy to learn in this context but give it a try. |
Beta Was this translation helpful? Give feedback.
-
also use those:
|
Beta Was this translation helpful? Give feedback.
-
just train about 2000 steps and post the log |
Beta Was this translation helpful? Give feedback.
-
Hi @francoishernandez, @vince62s,
Iʻve been trying, and trying, and trying to fine-tune TowerInstruct-7B-v0.1 with my bilingual datasets.
My prompt format for fine-tuning and validation files:
My prompt format for inference:
The output from the inference:
My fine-tuning config file:
I follow the Eole NLP TowerInstruct-7B-v0.1 recipen which I guess would work for the v0.2 version. During training, the Learning Rate is flat as a pancake, the accuracy stays in a flat range and wonʻt improve substantially. Then I merge the LoRA weights with the provided command and run inference.
See training logs here.
Please tell me what Iʻm doing wrong.
Beta Was this translation helpful? Give feedback.
All reactions