You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We would like to evaluate the model performance for various LLM fine tuning approaches and compare them with the standard benchmarks. An experiment we would like to try is:
Compare the full cartesian product of fine tuning for the Granite model (medium model) with relevant combinations:
{small, medium, large models} x {no pre-training + full supervised training, full supervised fine-tuning, LoRA, RAG, LoRA + RAG etc.} x {synthetic, no synthetic}. We can omit combinations that may not be relevant for our use case.
Benchmarks we can compare against (obtained from ChatGPT, we should validate these numbers with relevant published papers):
The text was updated successfully, but these errors were encountered:
hemajv
changed the title
Comparison of different LLM fine tuning methods against standard benchmarks
Comparison of different LLM fine tuning methods for Granite model against standard benchmarks
Jul 8, 2024
We would like to evaluate the model performance for various LLM fine tuning approaches and compare them with the standard benchmarks. An experiment we would like to try is:
Compare the full cartesian product of fine tuning for the Granite model (medium model) with relevant combinations:
{small, medium, large models} x {no pre-training + full supervised training, full supervised fine-tuning, LoRA, RAG, LoRA + RAG etc.} x {synthetic, no synthetic}. We can omit combinations that may not be relevant for our use case.
Benchmarks we can compare against (obtained from ChatGPT, we should validate these numbers with relevant published papers):
The text was updated successfully, but these errors were encountered: