Found 42 repositories(showing 30)
KushalRegmi61
Building GPT-2 from scratch , classification and instruction fine-tunning.
sumitdotml
learning about large language models by building a gpt-2 from scratch
Danish-Ahmed007
No description available
Abdullah-Eissa
No description available
This project provides code to build, train, and use a 124M-parameter GPT-2 model with GPT-3-inspired configurations (see References and Acknowledgements).
Ryan-W31
Building a GPT-2 Model from scratch!
anik475
Building GPT-2-124M from scratch using Pytorch
Hooper-at-Dev
Building an LLM from scratch with GPT-2 configrations
alexbond008
Building GPT-2 from scratch to understand from first principles how large language models work.
SCCSMARTCODE
A fundamental implementation of the GPT-2 architecture from scratch, designed to provide a clear and thorough understanding of generative pre-trained transformers. This repository focuses on building GPT-2 step by step, explaining the key components and their interactions for text generation and language modeling tasks
bhauryal7
Training LLM from scratch following the book 'Build a Large Language Model' by Sebastian Raschka
piyush-gpt
No description available
Da-nishAhmed
No description available
AlmightySoulking
Learning about building a Large Language model from scratch
Max00358
Building GPT-2 from scratch!
cjzcczc
Building GPT-2 from Scratch
jcentner
Building GPT-2 from scratch
Dinesh-Tejus
Building GPT-2 from scratch
pinakiganguly
Building GPT 2 from scratch
carlk13
Building GPT-2 from scratch
Educational implementation of a GPT-2 style decoder transformer in PyTorch, demonstrating self-attention, transformer blocks and autoregressive text generation.
Addi-11
Building gpt-2 model from scratch
Building a GPT-2 Transformer-Based Model from Scratch
No description available
No description available
harshitkumar009
Building the Architecture and pretraining the GPT 2 from scratch
marcinbogdanski
Collection of Jupyter Notebooks building up to reproduce GPT-2 from scratch.
saisena-ss
Building GPT-2 model and it's tokenizer from scratch - from Andrej Karpathy Youtube videos
AbdellatifSatir
A from-scratch implementation of GPT-2, focusing on understanding and building Transformer-based language models.
ronin-winter
Building an LLM from scratch specifically a small GPT-2 124M parameter model that is pre-trained on books from Project Gutenberg.