This repo includes pertinent codebase for developing a miniaturized GPT model. It follows GPT-2 architecture.
- Include text data
- run:
python data/data/prepare.py
python train.py --batch_size=32 --wandb_log=True
python sample.py --out_dir=out-wiki
This repo includes pertinent codebase for developing a miniaturized GPT model. It follows GPT-2 architecture.
python data/data/prepare.py
python train.py --batch_size=32 --wandb_log=True
python sample.py --out_dir=out-wiki