Impact of ASR performance on free speaking language assessment


Type
Conference Object
Change log
Authors
Knill, KM 
Gales, MJF 
Kyriakopoulos, Konstantinos  ORCID logo  https://orcid.org/0000-0002-7659-4541
Malinin, A 
Ragni, A 
Abstract

In free speaking tests candidates respond in spontaneous speech to prompts. This form of test allows the spoken language proficiency of a non-native speaker of English to be assessed more fully than read aloud tests. As the candidate's responses are unscripted, transcription by automatic speech recognition (ASR) is essential for automated assessment. ASR will never be 100% accurate so any assessment system must seek to minimise and mitigate ASR errors. This paper considers the impact of ASR errors on the performance of free speaking test auto-marking systems. Firstly rich linguistically related features, based on part-of-speech tags from statistical parse trees, are investigated for assessment. Then, the impact of ASR errors on how well the system can detect whether a learner's answer is relevant to the question asked is evaluated. Finally, the impact that these errors may have on the ability of the system to provide detailed feedback to the learner is analysed. In particular, pronunciation and grammatical errors are considered as these are important in helping a learner to make progress. As feedback resulting from an ASR error would be highly confusing, an approach to mitigate this problem using confidence scores is also analysed.

Description
Keywords
speech recognition, spoken language assessment
Journal Title
Proceedings of the Annual Conference of the International Speech Communication Association, INTERSPEECH
Conference Name
Interspeech 2018
Journal ISSN
2308-457X
1990-9772
Volume Title
2018-September
Publisher
ISCA
Sponsorship
Cambridge Assessment (unknown)