VisualCoder: Guiding Large Language Models in Code Execution with Fine-grained Multimodal Chain-of-Thought Reasoning
VisualCoder: Guiding Large Language Models in Code Execution with Fine-grained Multimodal Chain-of-Thought Reasoning
Predicting program behavior and reasoning about code execution remain significant challenges in software engineering, particularly for large language models (LLMs) designed for code analysis. While these models excel at understanding static syntax, they often struggle with dynamic reasoning tasks. We introduce VisualCoder, a simple yet effective approach that enhances code reasoning by integrating multimodal Chain-of-Thought (CoT) reasoning with a visual Control Flow Graph (CFG). By aligning code snippets with their corresponding CFGs, VisualCoder provides deeper insights into execution flows. We address challenges in multimodal CoT integration through a reference mechanism, ensuring consistency between code and its execution path, thereby improving performance in program behavior prediction, error detection, and output generation.
Dung Duy Le、Cuong Chi Le、Tien N. Nguyen、Nghi D. Q. Bui、Huy Nhat Phan、Hoang-Chau Truong-Vinh
计算技术、计算机技术
Dung Duy Le,Cuong Chi Le,Tien N. Nguyen,Nghi D. Q. Bui,Huy Nhat Phan,Hoang-Chau Truong-Vinh.VisualCoder: Guiding Large Language Models in Code Execution with Fine-grained Multimodal Chain-of-Thought Reasoning[EB/OL].(2024-10-30)[2025-08-02].https://arxiv.org/abs/2410.23402.点此复制
评论