Found 162 repositories(showing 30)
sleepingcat4
code to train a gpt-2 model to train it on tiny stories dataset according to the TinyStories paper
ChaitanyaK77
This Repository provides a Jupyter Notebook for building a small language model from scratch using 'TinyStories' dataset. Covers data preprocessing, BPE tokenization, binary storage, GPU memory management, and training a Transformer in PyTorch. Generate sample stories to test your model. Ideal for learning NLP and PyTorch.
Ce-daros
No description available
raymond-van
Reproducing GPT on the TinyStories dataset
sirohikartik
Tinystories version of gpt with custom inference engine
crabcode
Small indentation-based CYOA engine
noanabeshima
A TinyStories LM with SAEs and transcoders
clankur
a transformer implemented primarily using einops and trained on the tinystories dataset
tanaydesai
small language models based on decoder-only transformers pretrained on TinyStories from scratch.
gabrielolympie
Repo aiming at benchmarking several variant of transformer architectures for pre training, on datasets TinyStories and Pints 1.5
RobertKirk
Code for the TinyStories experiments from "Mechanistically analyzing the effects of fine-tuning on procedurally defined tasks".
annasoligo
A replication of Anthropic's Sleeper Agents and Stage-wise Model Diffing on a TinyStories language model.
AIDeveloperGroup
- Finetuning Llama-2-7b using QLoRA using TinyStories dataset - This finetuning was conducted on the free version of Google collab with less than 15 Gb VRAM thanks to QLora and Parameter-Efficient Fine-Tuning.
dishant2009
Implementation of TinyStories Paper, my first exposure to small language models
rrenaud
A reimplementation of the TinyStories paper for generating a stories dataset in non-English languages
kavyamali
0.2M and 2.5M parameter TinyStories model trained on GRU with and without attention.
HafsaaOuifak
Finetuning Llama-2-7b using QLoRA using TinyStories dataset
EN10
Train and run a small Llama 2 model from scratch on the TinyStories dataset.
SauravP97
Train Tiny Stories dataset from the paper - "TinyStories: How Small Can Language Models Be and Still Speak Coherent English?"
vivek-rd
Train GPT style model on tinystories dataset
oelin
A 2M parameter neural language model trained on the TinyStories corpus.
Pondsiders
Building transformers from scratch, one concept at a time. 13 Jupyter notebooks evolving from bag-of-words to mixture of experts, trained on TinyStories.
tchung1970
A CLI story generator powered by TinyStories-1M
vini2309
Small Language Model trained from scratch on TinyStories dataset
UlisseMini
Reproduction of TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
EN10
TinyStories on Modal
springtangent
An attempt to use the huggingface transformers library recreate the TinyStories models.
McMaart
An implementation of small language models for generating children stories. Inspired by "TinyStories" (Eldan and Li, 2023).
hiaripc
(WIP) llama2 scratch ttnn implementation to be trained with TinyStories. Goal: having a little gnome telling about stories on Tenstorrent e75 accelerator :-)
GOPITRINADH3561
🧠 Gemma 3 architecture implemented from scratch & pre-trained on TinyStories | 164.6M params | MQA, Sliding Window, RoPE, GeGLU | Perplexity 5.96 | Every component hand-coded in PyTorch