🎈 A series of lightweight GPT models featuring TinyGPT Base (~51M params) and TinyGPT2 (~95M params). Fast, creative text generation trained on whimsical stories.
Stars
16
Forks
1
Watchers
16
Open Issues
0
Overall repository health assessment
No package.json found
This might not be a Node.js project
42
commits
feat: update README with TinyGPT2.1 and TinyGPT2-DPO details, enhancing model descriptions and training configurations
edf93a8View on GitHubfeat: implement Direct Preference Optimization (DPO) training with dataset handling and evaluation
bac2014View on GitHubfeat: update model configurations for TinyGPT2 and TinyGPT2-SFT with new paths and descriptions
35deb0eView on GitHubfeat: add training loss curve image and hardware info
457af15View on GitHubfeat: add TinyGPT2-SFT model support and update related documentation
7f18bf6View on GitHubfeat: enhance inference and training scripts with improved prompt formatting and EOS handling
a59d39fView on GitHubfeat: add TinyGPT2 model implementation and update related configurations
86061eeView on GitHubfeat: enhance GPT model with support for cached generation and improved position handling
ba80a51View on GitHubfeat: enhance training script with evaluation and checkpointing features
67f8d65View on GitHubfeat: add script for model training using liger kernels
12dfd96View on GitHub