|国家预印本平台
首页|Efficient Tuning of Large Language Models for Knowledge-Grounded Dialogue Generation

Efficient Tuning of Large Language Models for Knowledge-Grounded Dialogue Generation

Efficient Tuning of Large Language Models for Knowledge-Grounded Dialogue Generation

来源:Arxiv_logoArxiv
英文摘要

Large language models (LLMs) demonstrate remarkable text comprehension and generation capabilities but often lack the ability to utilize up-to-date or domain-specific knowledge not included in their training data. To address this gap, we introduce KEDiT, an efficient method for fine-tuning LLMs for knowledge-grounded dialogue generation. KEDiT operates in two main phases: first, it employs an information bottleneck to compress retrieved knowledge into learnable parameters, retaining essential information while minimizing computational overhead. Second, a lightweight knowledge-aware adapter integrates these compressed knowledge vectors into the LLM during fine-tuning, updating less than 2\% of the model parameters. The experimental results on the Wizard of Wikipedia and a newly constructed PubMed-Dialog dataset demonstrate that KEDiT excels in generating contextually relevant and informative responses, outperforming competitive baselines in automatic, LLM-based, and human evaluations. This approach effectively combines the strengths of pretrained LLMs with the adaptability needed for incorporating dynamic knowledge, presenting a scalable solution for fields such as medicine.

Bo Zhang、Hui Ma、Dailin Li、Jian Ding、Jian Wang、Bo Xu、HongFei Lin

语言学

Bo Zhang,Hui Ma,Dailin Li,Jian Ding,Jian Wang,Bo Xu,HongFei Lin.Efficient Tuning of Large Language Models for Knowledge-Grounded Dialogue Generation[EB/OL].(2025-04-10)[2025-07-16].https://arxiv.org/abs/2504.07754.点此复制

评论