論文

査読有り 筆頭著者
2012年9月

Automatic Allocation of Training Data for Speech Understanding Based on Multiple Model Combinations

IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS
  • Kazunori Komatani
  • ,
  • Mikio Nakano
  • ,
  • Masaki Katsumaru
  • ,
  • Kotaro Funakoshi
  • ,
  • Tetsuya Ogata
  • ,
  • Hiroshi G. Okuno

E95D
9
開始ページ
2298
終了ページ
2307
記述言語
英語
掲載種別
研究論文(学術雑誌)
DOI
10.1587/transinf.E95.D.2298
出版者・発行元
IEICE-INST ELECTRONICS INFORMATION COMMUNICATIONS ENG

The optimal way to build speech understanding modules depends on the amount of training data available. When only a small amount of training data is available, effective allocation of the data is crucial to preventing overfitting of statistical methods. We have developed a method for allocating a limited amount of training data in accordance with the amount available. Our method exploits rule-based methods for when the amount of data is small, which are included in our speech understanding framework based on multiple model combinations, i.e., multiple automatic speech recognition (ASR) modules and multiple language understanding (LU) modules, and then allocates training data preferentially to the modules that dominate the overall performance of speech understanding. Experimental evaluation showed that our allocation method consistently outperforms baseline methods that use a single ASR module and a single LU module while the amount of training data increases.

リンク情報
DOI
https://doi.org/10.1587/transinf.E95.D.2298
Web of Science
https://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcAuth=JSTA_CEL&SrcApp=J_Gate_JST&DestLinkType=FullRecord&KeyUT=WOS:000309043000016&DestApp=WOS_CPL
ID情報
  • DOI : 10.1587/transinf.E95.D.2298
  • ISSN : 0916-8532
  • eISSN : 1745-1361
  • Web of Science ID : WOS:000309043000016

エクスポート
BibTeX RIS