Skip to main content

Hierarchical Softmax for End-to-End Low-resource Multilingual Speech Recognition

Qianying Liu (Kyoto University); Zhuo Gong (The University of Tokyo); Zhengdong Yang (Kyoto University); Yuhang Yang (School of Information Science and Engineering, Xinjiang University, China); Sheng Li (National Institute of Information & Communications Technology (NICT)); Chenchen Ding (); Nobuaki Minematsu (The University of Tokyo); Hao Huang (Xinjiang University); Fei Cheng (Kyoto University); Chenhui Chu (Kyoto University); Sadao Kurohashi (Kyoto University)

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
06 Jun 2023

Low-resource speech recognition has been long-suffering from insufficient training data. In this paper, we propose an approach that leverages neighboring languages to improve low-resource scenario performance, founded on the hypothesis that similar linguistic units in neighboring languages exhibit comparable term frequency distributions, which enables us to construct a Huffman tree for performing multilingual hierarchical Softmax decoding. This hierarchical structure enables cross-lingual knowledge sharing among similar tokens, thereby enhancing low-resource training outcomes. Empirical analyses demonstrate that our method is effective in improving the accuracy and efficiency of low-resource speech recognition.

More Like This

  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00
  • SPS
    Members: Free
    IEEE Members: $11.00
    Non-members: $15.00