Found 25 repositories(showing 25)
alisafaya
Neurocache: A library for augmenting language models with external caching mechanisms
eneswritescode
Deterministic LLM caching layer with context optimization, LRU eviction, model-aware token counting, and concurrency-safe request deduplication.
Velmurugan256
Web+Electorn model
hugomn
No description available
hemansnation
A lightweight, plug-and-play local memory module for LLM agents. Works offline, remembers past conversations and tasks, integrates into any Python project.
ZigRazor
Adaptive Memory-Aware Caching
TailwindMechanics
No description available
aryansharma2k2
No description available
Namaniit297
Designed a high-performance, configurable cache simulator in C++ for the ANN/SNN Loads. Supports tuning of cache parameters, replacement policies, and trace-based analysis. Enabled architectural insights for embedded systems.
dshrutika079-design
No description available
BogdanFloris
A self-hosted CDN and predictive cache built from scratch using Raft Consensus for cache metadata
archakamk
A Transformer-based code autocompletion engine with a FastAPI backend and VSCode extension, trained on open-source Python code for intelligent, low-latency code suggestions.
NWylynko
No description available
Velmurugan256
No description available
LukeMainwaring
A cache for your mind - AI chat grounded in what you actually know
pateldivyesh1323
Smart caching manager
rizwan199811
Reduce LLM API costs and speed up responses by caching completions with NeuroCache’s intelligent, provider-agnostic caching layer.
Saad7890-web
Memory for AI systems. Speed for everything else.
Neuro-Cache
A brain-inspired adaptive caching framework that learns what to keep or forget. Public demo version — commercial release coming December 2025.
iakshkhurana
Second Brain Application
Shilin237
No description available
JoaoCarlosHallak
No description available
TailwindMechanics
No description available
TailwindMechanics
No description available
Ha-r-i
NeuroCache is an intelligent cache that uses learning-based algorithms (like heuristics or RL) to predict and evict less useful items. Unlike static methods like LRU/LFU, it adapts to usage patterns, reducing latency, cache misses, and bandwidth in dynamic systems like browsers or CDNs.
All 25 repositories loaded