ダウンロード数: 231

このアイテムのファイル:
ファイル 記述 サイズフォーマット 
j.csl.2017.11.001.pdf758.4 kBAdobe PDF見る/開く
完全メタデータレコード
DCフィールド言語
dc.contributor.authorMirzaei, Maryam Sadaten
dc.contributor.authorMeshgi, Kouroshen
dc.contributor.authorKawahara, Tatsuyaen
dc.contributor.alternative河原, 達也ja
dc.date.accessioned2019-04-16T00:41:22Z-
dc.date.available2019-04-16T00:41:22Z-
dc.date.issued2018-05-
dc.identifier.issn0885-2308-
dc.identifier.urihttp://hdl.handle.net/2433/240843-
dc.description.abstractThis paper addresses the viability of using Automatic Speech Recognition (ASR) errors as the predictor of difficulties in speech segments, thereby exploiting them to improve Partial and Synchronized Caption (PSC), which we have proposed to train second language (L2) listening skill by encouraging listening over reading. The system uses ASR technology to make word-level text-to-speech synchronization and generates a partial caption. The baseline system determines difficult words based on three features: speech rate, word frequency and specificity. While it encompasses most of the difficult words, it does not cover a wide range of features that hinder L2 listening. Therefore, we propose the use of ASR systems as a model of L2 listeners and hypothesize that ASR errors can predict challenging speech segments for these learners. Among different cases of ASR errors, annotation results suggest the usefulness of four categories of homophones, minimal pairs, negatives, and breached boundaries for L2 listeners. A preliminary experiment with L2 learners focusing on these four categories of the ASR errors revealed that these cases highlight the problematic speech regions for L2 listeners. Based on the findings, the PSC system is enhanced to incorporate these kinds of useful ASR errors. An experiment with L2 learners demonstrated that the enhanced version of PSC is not only preferable, but also more helpful to facilitate the L2 listening process.en
dc.format.mimetypeapplication/pdf-
dc.language.isoeng-
dc.publisherElsevier BVen
dc.rights© 2018. This manuscript version is made available under the CC-BY-NC-ND 4.0 license http://creativecommons.org/licenses/by-nc-nd/4.0/.en
dc.rightsThe full-text file will be made open to the public on 1 May 2020 in accordance with publisher's 'Terms and Conditions for Self-Archiving'.en
dc.rightsThis is not the published version. Please cite only the published version.en
dc.rightsこの論文は出版社版でありません。引用の際には出版社版をご確認ご利用ください。ja
dc.subjectComputer-assisted language learningen
dc.subjectSecond language listening skillen
dc.subjectAutomatic speech recognitionen
dc.subjectPartial and synchronized captionen
dc.titleExploiting Automatic Speech Recognition Errors to Enhance Partial and Synchronized Caption for Facilitating Second Language Listeningen
dc.typejournal article-
dc.type.niitypeJournal Article-
dc.identifier.jtitleComputer Speech and Language-
dc.identifier.volume49-
dc.identifier.spage17-
dc.identifier.epage36-
dc.relation.doi10.1016/j.csl.2017.11.001-
dc.textversionauthor-
dc.addressGraduate School of Informatics, Kyoto Universityen
dc.addressGraduate School of Informatics, Kyoto Universityen
dc.addressGraduate School of Informatics, Kyoto Universityen
dcterms.accessRightsopen access-
datacite.date.available2020-05-01-
出現コレクション:学術雑誌掲載論文等

アイテムの簡略レコードを表示する

Export to RefWorks


出力フォーマット 


このリポジトリに保管されているアイテムはすべて著作権により保護されています。