Honorable mention
NVIDIA research-oriented toolkit for LLM KV-cache compression to stretch context within fixed VRAM budgets.
llmkv-cachecompressioninferencetaaft-repositories
Filter by platform, license text, maturity, maintenance cadence, and editorial tags like privacy-focused or self-hosted. Search matches names, summaries, tags, and use cases.
1 tool match your filters
NVIDIA research-oriented toolkit for LLM KV-cache compression to stretch context within fixed VRAM budgets.