Skip to content

Commit

Permalink
potentially fix tokenize_and_chunk
Browse files Browse the repository at this point in the history
  • Loading branch information
Guitaricet committed Jul 24, 2023
1 parent 6fd724d commit 2f6227b
Show file tree
Hide file tree
Showing 5 changed files with 4,377 additions and 7 deletions.
69 changes: 65 additions & 4 deletions README.dev.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,8 +2,40 @@ Some script to check that the most common training reigmes work.

```
torchrun --nproc-per-node 2 torchrun_main.py \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_gpt2_512 \
--tokenizer gpt2 \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_EleutherAI_pythia-1.4b_512 \
--model_name_or_path EleutherAI/pythia-1.4b \
--use_peft \
--relora 10 \
--model_revision step1000 \
--batch_size 4 \
--total_batch_size 96 \
--lr 5e-4 \
--max_length 512 \
--eval_every 20 \
--save_every 20 \
--num_training_steps 40 \
--distributed_type ddp \
--optimizer adam_zero \
--tags debug
torchrun --nproc-per-node 2 torchrun_main.py \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_EleutherAI_pythia-1.4b_512 \
--model_name_or_path EleutherAI/pythia-1.4b \
--model_revision step1000 \
--batch_size 6 \
--total_batch_size 96 \
--lr 5e-4 \
--max_length 512 \
--eval_every 2 \
--save_every 10 \
--num_training_steps 20 \
--distributed_type ddp \
--tags debug,fsdp_debug
torchrun --nproc-per-node 2 torchrun_main.py \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_t5-base_512 \
--model_config configs/llama_250m.json \
--batch_size 24 \
--total_batch_size 96 \
Expand All @@ -16,6 +48,35 @@ torchrun --nproc-per-node 2 torchrun_main.py \
--tags debug,fsdp_debug
torchrun --nproc-per-node 2 torchrun_main.py \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_t5-base_512 \
--model_config configs/llama_250m.json \
--batch_size 24 \
--total_batch_size 96 \
--lr 5e-4 \
--max_length 512 \
--eval_every 2 \
--save_every 10 \
--num_training_steps 20 \
--distributed_type fsdp \
--tags debug,fsdp_debug
torchrun --nproc-per-node 2 torchrun_main.py \
--dataset_path preprocessed_data/wikitext_wikitext-2-v1_gpt2_512 \
--model_config configs/llama_250m_50K.json \
--batch_size 24 \
--total_batch_size 96 \
--lr 5e-4 \
--max_length 512 \
--eval_every 2 \
--save_every 10 \
--num_training_steps 20 \
--distributed_type ddp \
--dtype float32 \
--tags debug,fsdp_debug
torchrun --nproc-per-node 2 torchrun_main.py \
--model_config configs/llama_250m.json \
--batch_size 24 \
Expand Down Expand Up @@ -59,7 +120,7 @@ torchrun --nproc-per-node 2 torchrun_main.py \
--num_training_steps 20000 \
--save_every 5000 \
--eval_every 5000 \
--continue_from checkpoints/llama_250m-2023-06-09-11-29-56/model_5000 \
--warmed_up_model checkpoints/llama_250m-2023-06-09-11-29-56/model_5000 \
--distributed_type fsdp \
--tags debug,fsdp_debug
Expand All @@ -81,7 +142,7 @@ torchrun --nproc-per-node 2 torchrun_main.py \
--num_training_steps 20000 \
--save_every 5000 \
--eval_every 5000 \
--continue_from checkpoints/llama_250m-2023-06-09-11-29-56/model_5000 \
--warmed_up_model checkpoints/llama_250m-2023-06-09-11-29-56/model_5000 \
--distributed_type fsdp \
--tags debug,fsdp_debug
Expand Down
Loading

0 comments on commit 2f6227b

Please sign in to comment.