ChatPaper.aiChatPaper

多语言思维能否增强大语言模型的推理能力?

Could Thinking Multilingually Empower LLM Reasoning?

April 16, 2025
作者: Changjiang Gao, Xu Huang, Wenhao Zhu, Shujian Huang, Lei Li, Fei Yuan
cs.AI

摘要

先前的研究表明,大型语言模型存在显著的“英语偏向性”,即在任务以英语呈现时,其表现往往更优。有趣的是,我们观察到在某些推理任务中,使用其他特定语言反而能获得优于英语的表现。然而,这一现象尚未得到充分探索。本文旨在探讨多语言在推理任务中的潜力上限,提出多语言推理相较于仅用英语推理,不仅有望显著提升(约10个Acc@k点),而且展现出更强的鲁棒性(对翻译质量和语言选择的变动具有容忍度)。除了分析这一上限背后的原因及实现过程中面临的挑战外,我们还发现,常见的答案选择方法因其局限性和偏见,无法触及这一上限。这些发现为未来研究如何充分利用多语言推理在大型语言模型中的潜力,指明了方向。
English
Previous work indicates that large language models exhibit a significant "English bias", i.e. they often perform better when tasks are presented in English. Interestingly, we have observed that using certain other languages in reasoning tasks can yield better performance than English. However, this phenomenon remains under-explored. In this paper, we explore the upper bound of harnessing multilingualism in reasoning tasks, suggesting that multilingual reasoning promises significantly (by nearly 10 Acc@k points) and robustly (tolerance for variations in translation quality and language choice) higher upper bounds than English-only reasoning. Besides analyzing the reason behind the upper bound and challenges in reaching it, we also find that common answer selection methods cannot achieve this upper bound, due to their limitations and biases. These insights could pave the way for future research aimed at fully harnessing the potential of multilingual reasoning in LLMs.

Summary

AI-Generated Summary

PDF252April 21, 2025