Back

The Korean Speech Recognition Sentences: A Large Corpus for Evaluating Semantic Context and Language Experience in Speech Perception.

  • Published In: Journal of Speech, Language & Hearing Research, 2023, v. 66, n. 9. P. 3399 1 of 3

  • Database: Academic Search Ultimate 2 of 3

  • Authored By: Jieun Song; Byungjun Kim; Minjeong Kim; Iverson, Paul 3 of 3

Abstract

Purpose: The aim of this study was to develop and validate a large Korean sentence set with varying degrees of semantic predictability that can be used for testing speech recognition and lexical processing. Method: Sentences differing in the degree of final-word predictability (predictable, neutral, and anomalous) were created with words selected to be suitable for both native and nonnative speakers of Korean. Semantic predictability was evaluated through a series of cloze tests in which native (n = 56) and nonnative (n = 19) speakers of Korean participated. This study also used a computer language model to evaluate final-word predictabilities; this is a novel approach that the current study adopted to reduce human effort in validating a large number of sentences, which produced results comparable to those of the cloze tests. In a speech recognition task, the sentences were presented to native (n = 23) and nonnative (n = 21) speakers of Korean in speech-shaped noise at two levels of noise. Results: The results of the speech-in-noise experiment demonstrated that the intelligibility of the sentences was similar to that of related English corpora. That is, intelligibility was significantly different depending on the semantic condition, and the sentences had the right degree of difficulty for assessing intelligibility differences depending on noise levels and language experience. Conclusions: This corpus (1,021 sentences in total) adds to the target languages available in speech research and will allow researchers to investigate a range of issues in speech perception in Korean. Supplemental Material: https://doi.org/10.23641/asha.24045582 [ABSTRACT FROM AUTHOR]

Additional Information

  • Source:Journal of Speech, Language & Hearing Research. 2023/09, Vol. 66, Issue 9, p3399
  • Document Type:Article
  • Subject Area:Computer Science
  • Publication Date:2023
  • ISSN:1092-4388
  • DOI:10.1044/2023_JSLHR-23-00137
  • Accession Number:171950380
  • Copyright Statement:Copyright of Journal of Speech, Language & Hearing Research is the property of American Speech-Language-Hearing Association and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)

Looking to go deeper into this topic? Look for more articles on EBSCOhost.