Memory-Augmented Architecture for Long-Term Context Handling in Large Language Models
Memory-Augmented Architecture for Long-Term Context Handling in Large Language Models
Large Language Models face significant challenges in maintaining coherent interactions over extended dialogues due to their limited contextual memory. This limitation often leads to fragmented exchanges and reduced relevance in responses, diminishing user experience. To address these issues, we propose a memory-augmented architecture that dynamically retrieves, updates, and prunes relevant information from past interactions, ensuring effective long-term context handling. Experimental results demonstrate that our solution significantly improves contextual coherence, reduces memory overhead, and enhances response quality, showcasing its potential for real-time applications in interactive systems.
Haseeb Ullah Khan Shinwari、Muhammad Usama
计算技术、计算机技术
Haseeb Ullah Khan Shinwari,Muhammad Usama.Memory-Augmented Architecture for Long-Term Context Handling in Large Language Models[EB/OL].(2025-06-23)[2025-07-21].https://arxiv.org/abs/2506.18271.点此复制
评论