1. 📘 Topic and Domain: Memory-augmented large language models (LLMs) with a focus on developing a lightweight and efficient memory system called LightMem.
2. 💡 Previous Research and New Ideas: Based on existing memory systems for LLMs and the Atkinson-Shiffrin human memory model, proposing a new three-stage memory architecture with sensory memory, short-term memory, and long-term memory with sleep-time updates.
3. ❓ Problem: Addressing the high computational overhead and inefficiencies in existing LLM memory systems while maintaining performance, particularly in handling long-context and multi-turn interactions.
4. 🛠️ Methods: Implements three key components: (1) Pre-compression sensory memory to filter redundant information, (2) Topic-aware short-term memory for semantic grouping, and (3) Sleep-time update mechanism for long-term memory maintenance with offline parallel updates.
5. 📊 Results and Evaluation: On LongMemEval benchmark, LightMem outperformed baselines by 2.70%-9.65% in QA accuracy while reducing token usage by 32×-117×, API calls by 17×-177×, and runtime by 1.67×-12.45× across GPT and Qwen backbones.