Back to search
Docker setup for llama.cpp server with router mode, supporting multiple GGUF models with lazy loading
Stars
1
Forks
0
Watchers
1
Open Issues
0
Overall repository health assessment
No package.json found
This might not be a Node.js project
5
commits
fix: auto-download improvements and add qwen35-9b-opus model
915cfe7View on GitHubfeat: auto-download missing models from HuggingFace on startup
a71eb28View on GitHubllama.cpp Docker setup with router mode, model presets, and auto-offload
9737303View on GitHub