The official repo for "LLoCo: Learning Long Contexts Offline"
-
Updated
Jun 15, 2024 - Python
The official repo for "LLoCo: Learning Long Contexts Offline"
Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)
Stop re-explaining your codebase to AI. Infinite speed memory + code graph for Claude Code & Codex CLI. 17 MCP tools, subagent protocol, hybrid search, TUI dashboard, crash recovery. Save 80-200K+ tokens/session.
Biological code organization system with 1,029+ production-ready snippets - 95% token reduction for Claude/GPT with AI-powered discovery & offline packs
Awesome list of paper on vision-based context compression
Rolling context compression for Claude Code — never hit the context wall. Auto-compresses old messages while keeping recent context verbatim. Zero config, zero latency. Works as a Claude Code plugin.
天将|LLM 智能体运行时资源控制引擎:Token 预算控制、成本优化、稳定性 Guard、多工具 Agent 调度
OpenClaw low-memory optimization guide for resource-constrained servers (2GB RAM)
Exploring Context Compression techniques for token reduction. Fine-tuning LLMs for efficient text compression and reduced inference costs, analyzing the trade-offs with Q&A accuracy.
Exploring artificial compressed languages to improve efficiency, context usage, and cross-lingual unification in LLMs
LLM context compression proxy — 40-70% token savings, zero code changes
Detecting silent pivot substitution in LLMs under context compression
Retriever, Summarizer, Reader for LLM ODQA(Open-Domain Question Answering) to increase Information Density
Contextomizer is an ultra-fast, deterministic library for transforming bloated tool outputs, raw APIs, documents, and messy logs into perfectly optimized context for AI Agents 🤖🚀
a technique for compressing verbose AI tool call outputs into concise summaries, reducing token consumption
Capture and compact long Google AI Studio sessions with multimodal image OCR and context handoff artifacts.
Infinite context for AI assistants using semantic compression and retrieval with Gemini
Agent memory runtime: short/long-term context, vector persistence, compression, and personalization primitives.
Add a description, image, and links to the context-compression topic page so that developers can more easily learn about it.
To associate your repository with the context-compression topic, visit your repo's landing page and select "manage topics."