Found 146 repositories(showing 30)
NadirRouter
Open-source LLM router & AI cost optimizer. Routes simple prompts to cheap/local models, complex ones to premium — automatically. Drop-in OpenAI-compatible proxy for Claude Code, Codex, Cursor, OpenClaw. Saves 40-70% on AI API costs. Self-hosted, no middleman.
thushan
High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model discovery across local and remote inference backends.
erans
LunaRoute is a high-performance local proxy for AI coding assistants like Claude Code, OpenAI Codex CLI, and OpenCode. Get complete visibility into every LLM interaction with zero-overhead passthrough, comprehensive session recording, and powerful debugging capabilities.
LLemonStack
All-in-one local low-code AI agent development platform. Installs and runs n8n, Flowise, Browser-Use, Qdrant, Ollama, and more. Proxies LLM requests through LiteLLM with Langfuse for observability.
pushinspektor856
An ultra-fast C++ daemon proxy that replaces the official GitHub Copilot endpoint, allowing you to use completely free local or open-source LLMs inside VS Code and JetBrains. Unlock advanced AI code completion with zero subscription fees and sub-millisecond latency.
xelektron
A local proxy that strips web pages down to clean text before they enter your AI agent's context window. 704K tokens → 2.6K tokens. No LLM required.
RamiBotAI
RamiBot v3.8.0 is a local-first AI security operations platform integrating multi-LLM support, a dynamic red/blue team skill pipeline, MCP tool orchestration, Docker terminal access, Tor proxy management, and an auto-integrated Kali-based tool server (rami-kali) for controlled, extensible offensive and defensive workflows
abra5umente
local proxy that lets claude (and other LLMs) grab youtube transcripts
lanesket
Know what you spend, see what you send. Lightweight local proxy that logs every LLM call — costs, tokens, full prompts and responses.
lynxai-team
Local LLM proxy, DevOps friendly
jmamda
A local reverse proxy that records every LLM request/response to SQLite. No cloud, no data leaving your machine.
simple10
Full stack observability for local AI agent development. Provides a llm-proxy to use as LLM provider base URL, logs traces to opik.
agenticloops-ai
Profile AI agents by intercepting LLM API traffic through a local MITM proxy. Understand how agents work: prompts, tools, MCP, token usage, costs, and timing — all in a real-time web UI.
eserie-fox
Lightweight Python CLI for managing FRP instances, with local stdio MCP support for LLM-assisted proxy maintenance.
MoonlightByte
The Pythagora LLM Proxy is a project that helps Pythagora utilize local Large Language Models (LLMs) hosted on LM Studio without relying on the OpenAI API
i3T4AN
PowerShell proxy that makes LM Studio models available with the OpenAI API format. Lets editors like Cursor or CodeGPT use local LLMs without paid keys.
rickcrawford
A local proxy that converts websites and APIs to clean Markdown. Convert HTML pages, JSON APIs, and dynamic sites. Get token counts for LLM budgeting.
GuthL
Local MITM proxy that keeps secrets out of LLM traffic.
dabit3
Local proxy to intercept, log, and analyze LLM API traffic with real-time cost tracking
latent-variable
LinoGPT: OpenWebUI + LM Studio + MCP integration. Use local LLMs with web search, HuggingFace, and other MCP tools through a lightweight proxy.
particlesector
Reverse proxy for monitoring and debugging local LLM agents (Ollama). Real-time dashboard, request logging, and performance metrics in a single binary
hellas-ai
local encrypted p2p llm inference proxy/gateway
pborenstein
Lightweight local LLM proxy with SQLite cost tracking & interactive model comparison. Unified OpenAI-compatible API for multiple providers.
DevianceLe
A powerful proxy server that connects Cursor IDE to your local Ollama instance, enabling seamless integration with local LLMs.
xuanzhi33
A lightweight local AI gateway that securely proxies LLM API requests with automatic key injection and CORS support.
dreygur
A Rust implementation of the Model Context Protocol (MCP) remote proxy that bridges local MCP clients (IDEs/LLMs) with remote MCP servers via HTTP/SSE transport.
Felix-Pedro
A simple fast api that acts as a proxy between your AFFiNE self-hosted instance and you local LLM endpoints, alowing for easy model switch and tweking.
dhannusch
Local proxy server that intercepts and visualizes all LLM interactions from Claude Code. Get real-time visibility into prompts, responses, tool calls, and token usage without modifying Claude Code itself.
vxcozy
Local OpenAI/Anthropic-compatible API proxy that delegates to the Codex/Claude CLI - use your Codex/Claude Max subscriptions in AI-aware IDEs
Yash-Singh1
No description available