Found 88 repositories(showing 30)
mostlygeek
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
varunvasudeva1
End-to-end documentation to set up your own local & fully private LLM server on Debian. Equipped with chat, web search, RAG, model management, MCP servers, image generation, and TTS.
pwilkin
Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends
Docker/podman container for llama.cpp/vllm/exllamav{2,3} orchestrated using llama-swap
zhiheng-huang
Multi-LLM agent framework with Claude Code-like tools. Use DeepSeek, Claude, GPT, Llama, or any model — same tools, same skills, swap freely. Includes file editing, code search, shell execution, sub-agents, and a production-ready API. The missing SDK between your LLM and your codebase.
meganoob1337
Dynamic LLM model swapping system with Docker, vLLM integration, and GPU acceleration. Supports GGUF & Hugging Face models with automatic swapping and Traefik routing.
claraverse-space
ClaraCore extends llama-swap with intelligent automation, bringing zero-configuration setup to llama.cpp deployments.
bjodah
Proof-of-concept Aider w. local (24GB vram) QwQ+Qwen2.5-Coder using litellm-proxy / llama-swap / llama.cpp
pkeffect
A robust, production-ready Python toolkit to automate the synchronization between a directory of .gguf model files and a llama-swap config.yaml
mostlygeek
homebrew tap for llama-swap
Mootikins
Unified LLM + STT container for AMD Strix Halo (ROCm)
AndrewMobbs
A container for CUDA 12 enabled llama-swap and llama.cpp server built on a distroless base
gpubuilders
Reliable model swapping for any local OpenAI compatible server - llama.cpp, vllm, etc
Exotik850
CLI for creating a configuration to be consumed by llama-swap for existing Ollama model blobs
schutzpunkt
Ansible playbook to configure AMD Strix Halo machines (e.g. Framework Desktop or GMKtec EVO-X2) as local AI inference servers running Fedora 43. Sets up llama.cpp with llama-swap and Open WebUI and downloads GGUF models. With NGINX reverse proxy and TLS via ACME or self-signed certificate.
ai-joe-git
A ComfyUI custom node for llama-swap the hot-swap model manager for llama.cpp.
de-wim
Container iamges for running llama-swap and llama-cpp on AMD hardware
cepa
Local LLM server setup with OpenWebUI and llama-swap + llama.cpp
A NVIDIA PyTorch image based Dockerfile with ComfyUI, vLLM, llama.cpp, llama-swap, OpenWebUI, ollama and a downloader for models.
eleqtrizit
Like Claude Code Router, but intended for use with Anthropic-compatible endpoints (LiteLLM, Llama Swap, vLLM)
BalanceBalls
llama-swap config
0rzech
Custom Llama Swap Container Image
mostlygeek
Documentation Site for llama-swap
mostlygeek
Configuration Examples for llama-swap
pi0neerpat
No description available
llama.cpp and llama-swap installer and updater powershell script
chriswessels
Swiftbar plugin providing MacOS menubar icon for llama-swap
shirayu
Auto-discover GGUF models from directories for llama-swap config
henk717
A simple wrapper for llama-swap to make it easy to use with KoboldCpp
ccebelenski
Auto-configure opencode to use a local llama-swap instance with model and context detection