Found 36,137 repositories(showing 30)
hiyouga
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠
datawhalechina
《开源大模型食用指南》针对中国宝宝量身打造的基于Linux环境快速微调(全参数/Lora)、部署国内外开源大模型(LLM)/多模态大模型(MLLM)教程
huggingface
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
ymcui
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
tloen
Instruct-tune LLaMA on consumer hardware
camenduru
stable diffusion webui colab
lyogavin
AirLLM 70B inference with single 4GB GPU
modelscope
Use PEFT or Full-parameter to CPT/SFT/DPO/GRPO 600+ LLMs (Qwen3.5, DeepSeek-R1, GLM-5, InternLM3, Llama4, ...) and 300+ MLLMs (Qwen3-VL, Qwen3-Omni, InternVL3.5, Ovis2.5, GLM4.5v, Llava, Phi4, ...) (AAAI 2025).
microsoft
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
OpenPipe
Agent Reinforcement Trainer: train multi-step agents for real-world tasks using GRPO. Give your agents on-the-job training. Reinforcement learning for Qwen3.5, GPT-OSS, Llama, and more!
LianjiaTech
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
cloneofsimo
Using Low-rank adaptation to quickly fine-tune diffusion models.
meshtastic
The official firmware for Meshtastic, an open-source, off-grid mesh communication system.
yangjianxin1
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
Lightning-AI
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
Akegarasu
SD-Trainer. LoRA & Dreambooth training scripts & GUI use kohya-ss's trainer, for diffusion model.
Lightricks
Official Python inference and LoRA trainer package for the LTX-2 audio–video generative model.
markqvist
The cryptography-based networking stack for building unstoppable networks with LoRa, Packet Radio, WiFi and everything in between.
transformerlab
The open source research environment for AI researchers to seamlessly train, evaluate, and scale models from local hardware to GPU clusters.
ExpressLRS
High Performance Open Source Radio Control Link
Facico
Chinese-Vicuna: A Chinese Instruction-following LLaMA-based Model —— 一个中文低资源的llama+lora方案,结构参考alpaca
1technophile
MQTT gateway for ESP8266 or ESP32 with bidirectional 433mhz/315mhz/868mhz, Infrared communications, BLE, Bluetooth, beacons detection, mi flora, mi jia, LYWSD02, LYWSD03MMC, Mi Scale, TPMS, BBQ thermometer compatibility & LoRa.
Michael-A-Kuykendall
⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
nunchaku-ai
[ICLR2025 Spotlight] SVDQuant: Absorbing Outliers by Low-Rank Components for 4-Bit Diffusion Models
mymusise
基于ChatGLM-6B + LoRA的Fintune方案
predibase
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
hiyouga
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
aiming-lab
🦞 Just talk to your agent — it learns and EVOLVES 🧬.
agentheroes
Generate, animate and schedule your AI characters 🤖