Skip to content

Latest commit

 

History

History
18 lines (10 loc) · 338 Bytes

README.md

File metadata and controls

18 lines (10 loc) · 338 Bytes

nugpt

This repo includes pertinent codebase for developing a miniaturized GPT model. It follows GPT-2 architecture.

Commands

Data Engineering

  • Include text data
  • run: python data/data/prepare.py

Model Training

  • python train.py --batch_size=32 --wandb_log=True

Inference

  • python sample.py --out_dir=out-wiki