Institutional Repository of Key Laboratory of Behavioral Science, CAS
Revealing the Parallel Multilingual Learning within Large Language Models | |
Yongyu Mu1; Peinan Feng1; Zhiquan Cao1; Yuzhang Wu1; Bei Li1; Chenglong Wang1; Tong Xiao1,2; Kai Song3; Tongran Liu4![]() | |
2024 | |
通讯作者邮箱 | lixiaoyumu9@gmail.com, xiaotong@mail.neu.edu.cn ; zhujingbo@mail.neu.edu.cn |
会议名称 | EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference |
会议录名称 | EMNLP 2024 - 2024 Conference on Empirical Methods in Natural Language Processing, Proceedings of the Conference |
页码 | 6976-6997 |
会议日期 | 2024 |
会议地点 | 不详 |
摘要 | Large language models (LLMs) can handle multilingual and cross-lingual text within a single input; however, previous works leveraging multilingualism in LLMs primarily focus on using English as the pivot language to enhance language understanding and reasoning. Given that multiple languages are a compensation for the losses caused by a single language's limitations, it's a natural next step to enrich the model's learning context through the integration of the original input with its multiple translations. In this paper, we start by revealing that LLMs learn from Parallel Multilingual Input (PMI). Our comprehensive evaluation shows that PMI enhances the model's comprehension of the input, achieving superior performance than conventional in-context learning (ICL). Furthermore, to explore how multilingual processing affects prediction, we examine the activated neurons in LLMs. Surprisingly, involving more languages in the input activates fewer neurons, leading to more focused and effective neural activation patterns. This neural reaction coincidently mirrors the neuroscience insight about synaptic pruning, highlighting a similarity between artificial and biological 'brains'. Our parallel multilingual data and code could be found at |
收录类别 | EI |
语种 | 英语 |
文献类型 | 会议论文 |
条目标识符 | https://ir.psych.ac.cn/handle/311026/48602 |
专题 | 中国科学院行为科学重点实验室 |
作者单位 | 1.NLP Lab, School of Computer Science and Engineering, Northeastern University, Shenyang, China 2.NiuTrans Research, Shenyang, China 3.Bytedance, Seattle, United States 4.CAS Key Laboratory of Behavioral Science, Institute of Psychology, CAS, Beijing, China |
推荐引用方式 GB/T 7714 | Yongyu Mu,Peinan Feng,Zhiquan Cao,et al. Revealing the Parallel Multilingual Learning within Large Language Models[C],2024:6976-6997. |
条目包含的文件 | ||||||
文件名称/大小 | 文献类型 | 版本类型 | 开放类型 | 使用许可 | ||
Revealing the Parall(3295KB) | 会议论文 | 开放获取 | CC BY-NC-SA | 浏览 下载 |
除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。
修改评论