Found 14 repositories(showing 14)
Simple Qwen3-VL gguf model loader for Comfy-UI.
GGUFloader
GGUF Loader with its Agentic Mode, and floating button, ai Models | Open Source & Offline. Mistral, Deepseek, llama, gemma, qwen
AidenTran900
A C++/Python machine learning library built from scratch. Features classic ML algorithms and a GGUF-compatible inference loader for transformers.
CarapaceUDE
llama.cpp fork: Qwen 3.5 hybrid GGUF + loader fixes; syncs with ggml-org/llama.cpp
zihaomu
No description available
winternewt
No description available
tiny LLM loader
Lolik612
The Loader for gguf models
Zenthrose
Universal vulkan gguf loader. Will load v1, v2, and v3 gguf files, all quantized formats
A fast API booty-licious back-end for running GGUF models with Llama.cpp
GGUFloader
Official blog and knowledge hub for GGUF Loader — guides, model reviews, and local AI tips.
ItsMehRAWRXD
RawrXD IDE — Pure C++20/MASM64 Win32 IDE with three-layer hotpatching, agentic framework, GGUF model loader
r-vage
Smart Language Model Loader for ComfyUI — unified vision-language AI with QwenVL, Mistral, Mistral3 / Ministral3, Florence-2 via Transformers, Docker (vLLM/SGLang/Ollama/llama.cpp), and GGUF backends. nVidia+AMD/ROCm support.
james04nesbitt
A standalone C++20 inference engine for Gemma-3 1B built from scratch, featuring a complete transformer forward pass and GGUF loader. It maximizes throughput and minimizes memory using SIMD-accelerated GEMM kernels (Google Highway), PagedAttention, FlashAttention, and outlier-aware INT8 KV cache quantization.
All 14 repositories loaded