Found 2,449 repositories(showing 30)
tloen
Instruct-tune LLaMA on consumer hardware
meta-llama
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama model family and using them on various provider services
Lightning-AI
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
pengxiao-song
🎉 Repo for LaWGPT, Chinese-Llama tuned with Chinese Legal knowledge. 基于中文法律知识的大语言模型
OpenGVLab
[ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
DAMO-NLP-SG
[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
unit-mesh
《AI 研发提效:自己动手训练 LoRA》,包含 Llama (Alpaca LoRA)模型、ChatGLM (ChatGLM Tuning)相关 Lora 的训练。训练内容:用户故事生成、测试代码生成、代码辅助生成、文本转 SQL、文本生成代码……
curiousily
Self-paced bootcamp on Generative AI. Tutorials on ML fundamentals, Ollama, LLMs, RAGs, LangChain, LangGraph, Fine-tuning, DSPy & AI Agents (CrewAI), (Using ChatGPT, gpt-oss, Claude, Qwen, Gemma, Llama, Gemini)
Maximilian-Winter
The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM models, execute structured function calls and get structured output. Works also with models not fine-tuned to JSON output and function calls.
jianzhnie
Easy and Efficient Finetuning LLMs. (Supported LLama, LLama2, LLama3, Qwen, Baichuan, GLM , Falcon) 大模型高效量化训练+部署.
ZebangCheng
Emotion-LLaMA: Multimodal Emotion Recognition and Reasoning with Instruction Tuning
zetavg
UI tool for fine-tuning and testing your own LoRA models base on LLaMA, GPT-J and more. One-click run on Google Colab. + A Gradio ChatGPT-like Chat UI to demonstrate your language models.
Shenzhi-Wang
This is the first Chinese chat model specifically fine-tuned for Chinese through ORPO based on the Meta-Llama-3-8B-Instruct model.
Lattice-zjj
OD-FinLLM is a refined model derived from the LLaMA series, with specific enhancements for Chinese financial knowledge. This model is built by fine-tuning LLaMA using a specialized instruction dataset created from publicly available Chinese financial Q&A data and additional web-scraped financial information.
jasonvanf
LLaMA-TRL: Fine-tuning LLaMA with PPO and LoRA
Locutusque
Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free
mzbac
Scripts for fine-tuning Llama2 via SFT and DPO.
Comprehensive toolkit for Reinforcement Learning from Human Feedback (RLHF) training, featuring instruction fine-tuning, reward model training, and support for PPO and DPO algorithms with various configurations for the Alpaca, LLaMA, and LLaMA2 models.
liutiedong
a Fine-tuned LLaMA that is Good at Arithmetic Tasks
git-cloner
llama fine-tuning with lora
juyongjiang
CodeUp: A Multilingual Code Generation Llama-X Model with Parameter-Efficient Instruction-Tuning
teelinsan
Camoscio: An Italian instruction-tuned language model based on LLaMA
bupticybee
Instruct-tune LLaMA on consumer hardware with shareGPT data
NisaarAgharia
Fine-Tuning Falcon-7B, LLAMA 2 with QLoRA to create an advanced AI model with a profound understanding of the Indian legal context.
vinid
ICLR2024 Paper. Showing properties of safety tuning and exaggerated safety.
ZrrSkywalker
Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parameters
poteminr
Instruct LLMs for flat and nested NER. Fine-tuning Llama and Mistral models for instruction named entity recognition. (Instruction NER)
vihangd
Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA
edumunozsala
Fine-tune and quantize Llama-2-like models to generate Python code using QLoRA, Axolot,..
BohdanPetryshyn
Fill-in-the-middle fine-tuning for the Code Llama model 🦙