Syllable-Based Sequence-to-Sequence Speech Recognition with the Transformer in Mandarin Chinese
Shiyu Zhou1,2; Linhao Dong1,2; Shuang Xu1; Bo Xu1
2018
会议日期2018
会议地点印度的海德拉巴
关键词Asr Multi-head Attention Syllable Based Acoustic Modeling Sequence-to-sequence
期号2018
英文摘要

Sequence-to-sequence attention-based models have recently
shown very promising results on automatic speech recognition
(ASR) tasks, which integrate an acoustic, pronunciation and
language model into a single neural network. In these models,
the Transformer, a new sequence-to-sequence attentionbased
model relying entirely on self-attention without using
RNNs or convolutions, achieves a new single-model state-ofthe-
art BLEU on neural machine translation (NMT) tasks. Since
the outstanding performance of the Transformer, we extend
it to speech and concentrate on it as the basic architecture of
sequence-to-sequence attention-based model on Mandarin Chinese
ASR tasks. Furthermore, we investigate a comparison between
syllable based model and context-independent phoneme
(CI-phoneme) based model with the Transformer in Mandarin
Chinese. Additionally, a greedy cascading decoder with the
Transformer is proposed for mapping CI-phoneme sequences
and syllable sequences into word sequences. Experiments on
HKUST datasets demonstrate that syllable based model with
the Transformer performs better than CI-phoneme based counterpart,
and achieves a character error rate (CER) of 28.77%,
which is competitive to the state-of-the-art CER of 28.0% by
the joint CTC-attention based encoder-decoder network.

会议录Interspeech
文献子类国际会议
语种英语
开始日期2018
结束日期2018
内容类型会议论文
源URL[http://ir.ia.ac.cn/handle/173211/22392]  
专题数字内容技术与服务研究中心_听觉模型与认知计算
通讯作者Shiyu Zhou
作者单位1.Institute of Automation, Chinese Academy of Sciences
2.University of Chinese Academy of Sciences
推荐引用方式
GB/T 7714
Shiyu Zhou,Linhao Dong,Shuang Xu,et al. Syllable-Based Sequence-to-Sequence Speech Recognition with the Transformer in Mandarin Chinese[C]. 见:. 印度的海德拉巴. 2018.
个性服务
查看访问统计
相关权益政策
暂无数据
收藏/分享
所有评论 (0)
暂无评论
 

除非特别说明,本系统中所有内容都受版权保护,并保留所有权利。


©版权所有 ©2017 CSpace - Powered by CSpace