Found 187 repositories(showing 30)
zhanshijinwat
Train a 1B LLM with 1T tokens from scratch by personal
FareedKhan-dev
A straightforward method for training your LLM, from downloading data to generating text.
wei-potato
使用deepspeed从头开始训练一个LLM,经过pretrain和sft阶段,验证llm学习知识、理解语言、回答问题的能力
marimo-team
A companion to CMU professor Zico Kolter's Intro to Modern AI. Learn the basics of machine learning, then train your own LLM from scratch.
KastanDay
Multi-model video-to-text by combining embeddings from Flan-T5 + CLIP + Whisper + SceneGraph. The 'backbone LLM' is pre-trained from scratch on YouTube (YT-1B dataset).
timmzimm
No description available
131AIClub
Implement and train a Tiny LLM from scratch!
Trained a 114 million Parameter LLM from Scratch.
LF-Luis
Personal project to learn how to build and pre-train a modern LLM from scratch.
saketd403
Train LLMs such as GPT and LLama from scratch.
Emisaber
A simple but somehow complete(though not enough) project to train a LLM from scratch
merterbak
Train LLM from scratch with SOTA techniques like RoPE, GQA and KV caching.
adriablancafort
A single script to train an LLM from scratch for less than $100 and run inference through a ChatGPT-like Web Interface or a CLI.
viralcode
Train your own LLM from scratch
minpeter
🦐 A minimal pretraining trainer for LLMs — from scratch.
vvasylkovskyi
AI LLM Project that barks trained from scratch
VachanVY
Generative Pretrained Model (GPT) in JAX. A step by step guide to train LLMs on large datasets from scratch
AIAnytime
Train a Hindi LLM from Scratch in pure Pytorch.
Tuziking
A project to pre-train a small LLM from scratch
Usman-Rafique
LLM Forge: Experimental playground for building and training practical LLMs with limited compute resources. Pre-train from scratch, fine-tune, and generate text using modular GPT-style models
akshatjain07065
This repository contains a small-scale Transformer-based LLM built from scratch using PyTorch and Hugging Face. The model is trained on Wikipedia data and deployed via FastAPI.
petermartens98
Lightweight LLM inspired by Qwen3, built from scratch in PyTorch. Full training pipeline with transformer components including RMSNorm, Rotary Position Embeddings (RoPE), Grouped-Query Attention (GQA), and SwiGLU layers. Trained with hybrid Muon + AdamW optimizer, causal masking, efficient batching, and evaluation tools.
mayurnayak1705
No description available
CastorYu
A simplistic script for training your own hybrid llm (using autoregressive model for drafting and diffusion model for refining).
vukrosic
Code & Train Anthropic Claude style LLM from scratch.
nabinkhair42
~58M parameter transformer LLM trained from scratch for Git/GitHub developer assistance.
Bilal-Ahmad6
LLM built from scratch predicting the next word trained on self defined weights.
Jayesh-Dev21
AI web application, using custom data-set to train and build a LLM from scratch to provide assistance for fraud detection and protfolio management
fayazkhan121
This is the official UI for UrduGPT — a custom-built English → Urdu translator powered by a Transformer-based LLM trained from scratch using PyTorch.
This project implements an LLM from scratch, utilizing the transformer mechanism similar to that of GPT-2. The model is trained on a customer service dataset.