Distilling Tool Knowledge into Language Models via Back-Translated Traces
Distilling Tool Knowledge into Language Models via Back-Translated Traces
Large language models (LLMs) often struggle with mathematical problems that require exact computation or multi-step algebraic reasoning. Tool-integrated reasoning (TIR) offers a promising solution by leveraging external tools such as code interpreters to ensure correctness, but it introduces inference-time dependencies that hinder scalability and deployment. In this work, we propose a new paradigm for distilling tool knowledge into LLMs purely through natural language. We first construct a Solver Agent that solves math problems by interleaving planning, symbolic tool calls, and reflective reasoning. Then, using a back-translation pipeline powered by multiple LLM-based agents, we convert interleaved TIR traces into natural language reasoning traces. A Translator Agent generates explanations for individual tool calls, while a Rephrase Agent merges them into a fluent and globally coherent narrative. Empirically, we show that fine-tuning a small open-source model on these synthesized traces enables it to internalize both tool knowledge and structured reasoning patterns, yielding gains on competition-level math benchmarks without requiring tool access at inference.
Haige Bo、Changran Hu、Guohao Li、Xingyue Huang、Xianglong Hu、Zifeng Ding、Yuan He、Rishabh、Waleed Alzarooni、Ziyu Ye、Wendong Fan、Bailan He
计算技术、计算机技术数学
Haige Bo,Changran Hu,Guohao Li,Xingyue Huang,Xianglong Hu,Zifeng Ding,Yuan He,Rishabh,Waleed Alzarooni,Ziyu Ye,Wendong Fan,Bailan He.Distilling Tool Knowledge into Language Models via Back-Translated Traces[EB/OL].(2025-06-23)[2025-07-09].https://arxiv.org/abs/2506.19171.点此复制
评论