Found 97 repositories(showing 30)
Demystify AI agents by building them yourself. Local LLMs, no black boxes, real understanding of function calling, memory, and ReAct patterns.
withcatai
Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level
pguso
Demystify RAG by building it from scratch. Local LLMs, no black boxes - real understanding of embeddings, vector search, retrieval, and context-augmented generation.
withcatai
Run AI ✨ assistant locally! with simple API for Node.js 🚀
MinusZoneAI
基于llama.cpp的一些和提示词相关的节点,目前包括美化提示词和类似clip-interrogator的图片反推 | Use llama.cpp to assist in generating some nodes related to prompt words, including beautifying prompt words and image recognition similar to clip-interrogator
FranckyB
Prompt Manager for ComfyUI, with integration with llama.cpp for prompt generation. Allowing users to generate and save prompts, as well as extract prompts from existing images and video.. The Advance node also allows saving Lora Stacks allong with prompts. Giving the user the option of saving the complete Recipe.
kijai
Simple interface for llama-cpp as ComfyUI node
Build an AI communication analyzer from scratch to understand how AI products actually work. Learn prompt engineering, reasoning pipelines, and local LLM integration using Node.js - no frameworks, no abstractions, just fundamentals
SurfaceData
A demonstration Electron app using node-llama-cpp
dspasyuk
Llama.cui is a small llama.cpp-based chat application for Node.js
mybigday
Node.js binding of llama.cpp
tib0
Local Llama project, L³ is an electron app that runs llama 3 models locally
HyperMink
Scalable AI Inference Server for CPU and GPU with Node.js | Utilizes llama.cpp and parts of llamafile C/C++ core under the hood.
iBz-04
run llms and slms on your hardware & browser
Harry-Ross
A Next.js chat app to use Llama 2 locally using node-llama-cpp
kantan-kanto
Local LLM session nodes for ComfyUI using GGUF and llama.cpp, supporting Llama, Mistral, Qwen, DeepSeek, GLM, Gemma, Phi, LLaVA and gpt-oss, enabling both user–model chat and model-to-model dialogue without external runtimes like Ollama.
sebagallo
ComfyUI custom node that acts as a llama-cpp-python wrapper, with support for vision models. It allows the user to generate text responses from prompts using llama.cpp.
Node-RED Flows for OpenAI API compatible endpoints calling llama.cpp
T8mars
Simple interface for llama-cpp as ComfyUI node,support qwen3.5,gemma 4
arseniy0924
Web UI for orchestrating distributed llama.cpp RPC GPU clusters with auto node discovery, telemetry, and one-click deployment.
pictorialink
Custom nodes for ComfyUI QWen3 8b running based on llama.cpp, which only support the CUDA framework and do not support MPS.
janole
The friendly and powerful desktop AI chatbot supporting both local and cloud AI models
ai-joe-git
A ComfyUI custom node for llama-swap the hot-swap model manager for llama.cpp.
curiousjp
A thin wrapper for llama-cpp-python to be called from a comfyui node
TheeValcode
A TypeScript implementation of a Retrieval-Augmented Generation (RAG) module using node-llama-cpp for embeddings and SQLite with vector extensions for efficient similarity search.
LukeRouleau
A utility to log the structure of a GGML GGUF graph after it is constructed. This can be used to analyze the tensor sizes into and out of different nodes in the compute graph. It includes my fork of llama.cpp which adds the graph logging functionality.
hammercui
致敬原作: qmd (https://github.com/tobi/qmd) - TypeScript实现的混合搜索引擎 Python重写版本 - 专为Windows稳定性优化和更高质量的检索体验而生 为什么重写? 原版qmd使用node-llama-cpp在Windows上存在严重的稳定性问题(随机崩溃)。本项目改用transformers + PyTorch技术栈
shailendra-repo
No description available
tib0
Node Llama Cpp wrapper for Node JS
aawadat
https://github.com/aawadat/node-llama-cpp.git