Found 36 repositories(showing 30)
h2oai
Private chat with local GPT with document, images, video, etc. 100% private, Apache 2.0. Supports oLLaMa, Mixtral, llama.cpp, and more. Demo: https://gpt.h2o.ai/ https://gpt-docs.h2o.ai/
keldenl
A llama.cpp drop-in replacement for OpenAI's GPT endpoints, allowing GPT-powered apps to run off local llama.cpp models instead of OpenAI.
rhohndorf
Uses Auto-GPT with Llama.cpp
docusealco
Ruby FFI bindings for llama.cpp to run open-source LLMs such as GPT-OSS, Qwen 3.5, Gemma 4, and Llama 3 locally with Ruby.
jhud
An easily-trained baby GPT that can stand in for the real thing. Based on Andrej Karpathy's makemore, but set up to mimic a llama-cpp server. This is not production-ready; it's a toy implementation for educational purposes.
unixwzrd
A macOS version of the oobabooga gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and GALACTICA.
tohurtv
Llama.cpp-qt is a Python-based GUI wrapper for the LLama.cpp server, providing a user-friendly interface for configuring and running the server. LLama.cpp is a lightweight implementation of GPT-like models.
kantan-kanto
Local LLM session nodes for ComfyUI using GGUF and llama.cpp, supporting Llama, Mistral, Qwen, DeepSeek, GLM, Gemma, Phi, LLaVA and gpt-oss, enabling both user–model chat and model-to-model dialogue without external runtimes like Ollama.
keldenl
Discord Bot running with GPT. Supports gpt-llama.cpp
yrezehi
Simple lightweight up-to-date llama.cpp & ggml.cpp binding - whisper, gpt-2, LLava, starcoder and Mistral - ONGOING
feers77
Llama.cpp fork with implemented engram technology to run models without GPU. This proof of concept runs gpt-oss:120b only using cpu and ram.
sovit-123
A local RAG + web search pipeline with gpt-oss and other similar scale models powered by llama.cpp
PunithVT
AI-Powered Inference Platform - Deploy OpenAI's GPT-OSS-20B on AWS EC2 with GPU acceleration using llama.cpp.
rick-stevens-ai
Run OpenAI GPT-OSS-120B (116.83B params, 60GB) on SINGLE Intel MAX GPU 1550 using llama.cpp SYCL backend
lindelwemyeza
A simple GPT that runs privately on your computer. Uses python's llama_cpp and Mistral 7b parameter model
tripathiarpan20
Docker image for the Text Generation gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and GALACTICA.
wambugu71
Local gpt in llama.cpp models with chat interface
AlanLiang-314
simple gpt model train with nanoGPT and port to llama.cpp
Manojpatil123
No description available
BrianAshihara
Criação de um ChatBot local com um modelo Generic GPT Unified Format disponibilizado pela Microsoft, para de aprendizado e manuseio de LLM`s. Aprendizado também da biblioteca llama-cpp em Python.
sujal-maheshwari2004
BashAI is a custom Unix-like shell in Python with built-in AI assistance. It supports standard shell features (commands, piping, redirection, jobs) and enhances usability with autocompletion, history, and aliases. The docbot assistant provides real-time help via GPT-4 or offline using Llama.cpp for smarter terminal use.
liangali
No description available
jefripunza
No description available
a llama.cpp openai style web api server
stevenke1981
llama.cpp deployment scripts for GPT-OSS 20B GGUF model (Windows & Linux)
escape-velocity-ai
A general docker container for running openai gpt oss models using llama.cpp
dvrlabs
Basic CLI to local gpt-oss LLM running in llama.cpp. Made with Odin.
VatsaDev
Implementing sampling methods and features from HF transformers/llama.cpp/kobold/etc into Nano-GPT
nefaereti
One-command installer and uninstaller for GPT-OSS 20B HERETIC uncensored AI model. Automatically downloads, verifies, and runs locally on Windows with llama.cpp.
aman-chauhan
Offline paper-reading companion using llama.cpp + GPT-OSS + Python. Helps you locate evidence, summarize sections, and build your own notes while keeping analysis local.