We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi,
I tried to train below models with LorEFT for instruction fine tuning of text2SQL-
But these models are not able to load to GPU. I tried this code base to load the model and finetune: https://github.com/stanfordnlp/pyreft/blob/main/examples/loreft/train.py
Command to run the script in above mentioned link: python train.py --task "instruct" --model "meta-llama/Llama-2-7b-hf" --data_dir /datasets --train_dataset /datasets/text_2_sql/train_final.json --seed 42 --layers 'all' --rank 4 --position "f10+l10" --epochs 16 --save_model --intervention_type "LoreftIntervention" --gradient_accumulation_steps 4 --batch_size 4 --output_dir /finetuned_models/reft_models/test_layers_all_rank4_fl1010_epochs_16_lr_5e3 --lr 5e-3 --schedule "cosine" --warmup_ratio 0.03 --logging_steps 10
python train.py --task "instruct" --model "meta-llama/Llama-2-7b-hf" --data_dir /datasets --train_dataset /datasets/text_2_sql/train_final.json --seed 42 --layers 'all' --rank 4 --position "f10+l10" --epochs 16 --save_model --intervention_type "LoreftIntervention" --gradient_accumulation_steps 4 --batch_size 4 --output_dir /finetuned_models/reft_models/test_layers_all_rank4_fl1010_epochs_16_lr_5e3 --lr 5e-3 --schedule "cosine" --warmup_ratio 0.03 --logging_steps 10
Am I missing anything here? Could someone look into this issue for further model support?
The text was updated successfully, but these errors were encountered:
Hey, it is because the script is for Llama models! Not because of the library, i think. For all other types, please add them to this dict: https://github.com/stanfordnlp/pyreft/blob/main/examples/loreft/train.py#L57
Sorry, something went wrong.
frankaging
No branches or pull requests
Hi,
I tried to train below models with LorEFT for instruction fine tuning of text2SQL-
But these models are not able to load to GPU.
I tried this code base to load the model and finetune: https://github.com/stanfordnlp/pyreft/blob/main/examples/loreft/train.py
Command to run the script in above mentioned link:
python train.py --task "instruct" --model "meta-llama/Llama-2-7b-hf" --data_dir /datasets --train_dataset /datasets/text_2_sql/train_final.json --seed 42 --layers 'all' --rank 4 --position "f10+l10" --epochs 16 --save_model --intervention_type "LoreftIntervention" --gradient_accumulation_steps 4 --batch_size 4 --output_dir /finetuned_models/reft_models/test_layers_all_rank4_fl1010_epochs_16_lr_5e3 --lr 5e-3 --schedule "cosine" --warmup_ratio 0.03 --logging_steps 10
Am I missing anything here?
Could someone look into this issue for further model support?
The text was updated successfully, but these errors were encountered: