计算机科学
语言模型
判别式
正规化(语言学)
适应(眼睛)
秩(图论)
语音识别
域适应
绩效改进
人工智能
机器学习
模式识别(心理学)
物理
光学
经济
组合数学
分类器(UML)
数学
运营管理
作者
Yu Yu,Chao-Han Huck Yang,Jari Kolehmainen,Prashanth Gurunath Shivakumar,Yile Gu,Sungho Ryu Roger Ren,Qi Luo,Aditya Gourav,I‐Ming Chen,Yi-Chieh Liu,Tuan Dinh,Ankur Gandhe Denis Filimonov,Shalini Ghosh,Andreas Stolcke,Ariya Rastow,Ivan Bulyko
标识
DOI:10.1109/asru57964.2023.10389632
摘要
We propose a neural language modeling system based on low-rank adaptation (LoRA) for speech recognition output rescoring. Although pretrained language models (LMs) like BERT have shown superior performance in second-pass rescoring, the high computational cost of scaling up the pretraining stage and adapting the pretrained models to specific domains limit their practical use in rescoring. Here we present a method based on low-rank decomposition to train a rescoring BERT model and adapt it to new domains using only a fraction (0.08%) of the pretrained parameters. These inserted matrices are optimized through a discriminative training objective along with a correlation-based regularization loss. The proposed low-rank adaptation RescoreBERT (LoRB) architecture is evaluated on LibriSpeech and internal datasets with decreased training times by factors between 5.4 and 3.6.
科研通智能强力驱动
Strongly Powered by AbleSci AI