14-stage Fusion Pipeline for LLM token compression — reversible compression, AST-aware code analysis, intelligent content routing. Zero LLM inference cost. MIT licensed.
-
Updated
Mar 18, 2026 - Python
14-stage Fusion Pipeline for LLM token compression — reversible compression, AST-aware code analysis, intelligent content routing. Zero LLM inference cost. MIT licensed.
The official repo for "LLoCo: Learning Long Contexts Offline"
Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)
🦞 龙虾饼 - 智能上下文压缩系统,让 AI 记忆永不溢出
Stop re-explaining your codebase to AI. Infinite speed memory + code graph for Claude Code & Codex CLI. 17 MCP tools, subagent protocol, hybrid search, TUI dashboard, crash recovery. Save 80-200K+ tokens/session.
Rolling context compression for Claude Code — never hit the context wall. Auto-compresses old messages while keeping recent context verbatim. Zero config, zero latency. Works as a Claude Code plugin.
Biological code organization system with 1,029+ production-ready snippets - 95% token reduction for Claude/GPT with AI-powered discovery & offline packs
Awesome list of paper on vision-based context compression
LLM context compression proxy — 40-70% token savings, zero code changes
🌶️ OpenClaw skill that reduces token consumption by 30-60% through context compression and smart optimization
SYMB Session Intelligence | SYMB-Mem | context compression for Claude Code. All Data Is Important.
OpenClaw low-memory optimization guide for resource-constrained servers (2GB RAM)
Zed extension for Headroom — context compression for AI agents
Exploring Context Compression techniques for token reduction. Fine-tuning LLMs for efficient text compression and reduced inference costs, analyzing the trade-offs with Q&A accuracy.
Squeeze verbose LLM agent tool output down to only the relevant lines
Agent memory runtime: short/long-term context, vector persistence, compression, and personalization primitives.
Exploring artificial compressed languages to improve efficiency, context usage, and cross-lingual unification in LLMs
Detecting silent pivot substitution in LLMs under context compression
Capture and compact long Google AI Studio sessions with multimodal image OCR and context handoff artifacts.
Add a description, image, and links to the context-compression topic page so that developers can more easily learn about it.
To associate your repository with the context-compression topic, visit your repo's landing page and select "manage topics."