🧪 Advanced LLM fine-tuning framework with Reinforcement Learning (GRPO/DPO), Multi-Agent Swarm Training, Adaptive Optimization, and Unsloth integration (2x faster, 70% less VRAM). Train 1.5B-70B+ models on 8GB+ GPUs with QLoRA, PEFT, LUFFY off-policy reasoning, and Search-R1. RAG-enabled with smart chunking.
Stars
2
Forks
0
Watchers
2
Open Issues
0
Overall repository health assessment
No package.json found
This might not be a Node.js project
12
commits
feat(docs): convert mermaid diagrams to SVG with improved contrast
85305d3View on GitHubdocs: Update README with new integrations and mermaid diagrams
5679d2eView on GitHubfeat: Add integrations inspired by PocketFlow, AgentFlow, claude-flow, and Unsloth
f97ac70View on GitHubdocs: translate all remaining Italian comments to English
9158ee8View on GitHubdocs: translate remaining Italian comments to English in Python files
b5a2f74View on GitHubdocs: fix architecture anchor link - remove emoji variation selector
f00d439View on GitHubdocs: update all GPU references to RTX 2070 Super 8GB
89613c4View on GitHubdocs: fix anchor links for GitHub and update model size support
3087defView on GitHubdocs: fix navigation links and update presentation for RTX 2070 Super 8GB and DeepSeek Qwen Distill 1.5B
49589d2View on GitHubdocs: replace benchmarks section with linkedin-benchmarks.png image
10c92dbView on GitHubfix: regenerate architecture diagram and replace training speed with image
86b8b2eView on GitHubdocs: replace Mermaid diagrams with static SVG images for mobile compatibility
f19080cView on GitHub