A llama.cpp drop-in replacement for OpenAI's GPT endpoints, allowing GPT-powered apps to run off local llama.cpp models instead of OpenAI.
Stars
596
Forks
67
Watchers
596
Open Issues
30
Overall repository health assessment
^1.0.2^2.8.5^4.18.2^0.0.1-security^1.1.8^4.0.2^0.12.7^6.2.8^4.6.2^2.0.222.8.7174
commits
6
commits
6
commits
2
commits
1
commits
revert inference engine changes and restore sibling folder structure for ggml
cb46d55View on GitHubremove inference engine support. split it out from gpt-llama.cpp
94d440bView on GitHubmake sure completion routes are ending conversations properly
995aadcView on GitHubadd ngl and support for short and long versions of the variables
942ed88View on GitHub