The official repo for "LLoCo: Learning Long Contexts Offline"
-
Updated
Jun 15, 2024 - Python
The official repo for "LLoCo: Learning Long Contexts Offline"
Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)
Stop re-explaining your codebase to AI. Infinite speed memory + code graph for Claude Code & Codex CLI. 17 MCP tools, subagent protocol, hybrid search, TUI dashboard, crash recovery. Save 80-200K+ tokens/session.
Biological code organization system with 1,029+ production-ready snippets - 95% token reduction for Claude/GPT with AI-powered discovery & offline packs
Awesome list of paper on vision-based context compression
OpenClaw low-memory optimization guide for resource-constrained servers (2GB RAM)
天将|LLM 智能体运行时资源控制引擎:Token 预算控制、成本优化、稳定性 Guard、多工具 Agent 调度
Exploring Context Compression techniques for token reduction. Fine-tuning LLMs for efficient text compression and reduced inference costs, analyzing the trade-offs with Q&A accuracy.
Retriever, Summarizer, Reader for LLM ODQA(Open-Domain Question Answering) to increase Information Density
Exploring artificial compressed languages to improve efficiency, context usage, and cross-lingual unification in LLMs
LLM context compression proxy — 40-70% token savings, zero code changes
Detecting silent pivot substitution in LLMs under context compression
a technique for compressing verbose AI tool call outputs into concise summaries, reducing token consumption
Infinite context for AI assistants using semantic compression and retrieval with Gemini
Agent memory runtime: short/long-term context, vector persistence, compression, and personalization primitives.
Add a description, image, and links to the context-compression topic page so that developers can more easily learn about it.
To associate your repository with the context-compression topic, visit your repo's landing page and select "manage topics."