|国家预印本平台
| 注册
首页|Strengthening Programming Comprehension in Large Language Models through Code Generation

Strengthening Programming Comprehension in Large Language Models through Code Generation

Strengthening Programming Comprehension in Large Language Models through Code Generation

来源:Arxiv_logoArxiv
英文摘要

Large language models (LLMs) have recently shown impressive results on diverse code-related tasks, benefiting from large-scale training and instruction tuning. However, studies reveal that their grasp of fundamental programming concepts, such as data flow and control flow, remains shallow, leading to fragile performance when code requires deeper reasoning. This limitation restricts the practical adoption of LLMs in real-world software development. To address this issue, this work introduces a counterfactual code augmentation framework combined with concept-aware tuning, designed to guide LLMs toward stronger conceptual understanding. Comprehensive evaluation across multiple models and benchmarks demonstrates the effectiveness of the proposed approach.

Xiaoning Ren、Qiang Hu、Wei Ma、Yan Li、Yao Zhang、Lingxiao Jiang、Yinxing Xue

计算技术、计算机技术

Xiaoning Ren,Qiang Hu,Wei Ma,Yan Li,Yao Zhang,Lingxiao Jiang,Yinxing Xue.Strengthening Programming Comprehension in Large Language Models through Code Generation[EB/OL].(2025-08-18)[2025-09-09].https://arxiv.org/abs/2508.12620.点此复制

评论