|国家预印本平台
首页|Deciphering Trajectory-Aided LLM Reasoning: An Optimization Perspective

Deciphering Trajectory-Aided LLM Reasoning: An Optimization Perspective

Deciphering Trajectory-Aided LLM Reasoning: An Optimization Perspective

来源:Arxiv_logoArxiv
英文摘要

We propose a novel framework for comprehending the reasoning capabilities of large language models (LLMs) through the perspective of meta-learning. By conceptualizing reasoning trajectories as pseudo-gradient descent updates to the LLM's parameters, we identify parallels between LLM reasoning and various meta-learning paradigms. We formalize the training process for reasoning tasks as a meta-learning setup, with each question treated as an individual task, and reasoning trajectories serving as the inner loop optimization for adapting model parameters. Once trained on a diverse set of questions, the LLM develops fundamental reasoning capabilities that can generalize to previously unseen questions. Extensive empirical evaluations substantiate the strong connection between LLM reasoning and meta-learning, exploring several issues of significant interest from a meta-learning standpoint. Our work not only enhances the understanding of LLM reasoning but also provides practical insights for improving these models through established meta-learning techniques.

Junnan Liu、Hongwei Liu、Linchen Xiao、Shudong Liu、Taolin Zhang、Zihan Ma、Songyang Zhang、Kai Chen

计算技术、计算机技术

Junnan Liu,Hongwei Liu,Linchen Xiao,Shudong Liu,Taolin Zhang,Zihan Ma,Songyang Zhang,Kai Chen.Deciphering Trajectory-Aided LLM Reasoning: An Optimization Perspective[EB/OL].(2025-05-26)[2025-06-13].https://arxiv.org/abs/2505.19815.点此复制

评论