论文标题
元转移学习,用于代码开关语音识别
Meta-Transfer Learning for Code-Switched Speech Recognition
论文作者
论文摘要
当今世界上越来越多的人说是多种语言的混合语言。但是,由于资源有限,收集混合语言数据所需的费用和大量努力,建立用于代码转换的语音识别系统仍然很困难。因此,我们提出了一种新的学习方法,元转移学习,以通过明智地从高资源单语言数据集中提取信息,以低资源环境中的代码切换语音识别系统进行学习。我们的模型学会了识别单个语言,并转移它们,以便通过对代码切换数据进行优化来更好地识别混合语言语音。基于实验结果,我们的模型在语音识别和语言建模任务上的表现优于现有基准,并且收敛速度更快。
An increasing number of people in the world today speak a mixed-language as a result of being multilingual. However, building a speech recognition system for code-switching remains difficult due to the availability of limited resources and the expense and significant effort required to collect mixed-language data. We therefore propose a new learning method, meta-transfer learning, to transfer learn on a code-switched speech recognition system in a low-resource setting by judiciously extracting information from high-resource monolingual datasets. Our model learns to recognize individual languages, and transfer them so as to better recognize mixed-language speech by conditioning the optimization on the code-switching data. Based on experimental results, our model outperforms existing baselines on speech recognition and language modeling tasks, and is faster to converge.