Automated Scoring of Speaking Tasks in the Test of English‐for‐Teaching (TEFT™)

This research report presents a summary of research and development efforts devoted to creating scoring models for automatically scoring spoken item responses of a pilot administration of the Test of English-for-Teaching (TEFT™) within the ELTeach™ framework. The test consists of items for all four language modalities: reading, listening, writing, and speaking. This report only addresses the speaking items, which elicit responses ranging from highly predictable to semipredictable speech from nonnative English teachers or teacher candidates. We describe the components of the system for automated scoring, comprising an automatic speech recognition (ASR) system, a set of filtering models to flag nonscorable responses, linguistic measures relating to the various construct subdimensions, and multiple linear regression scoring models for each item type. Our system is set up to simulate a hybrid system whereby responses flagged as potentially nonscorable by any component of the filtering model are routed to a human rater, and all other responses are scored automatically by our system.

[1]  Jack Mostow,et al.  A Prototype Reading Coach that Listens , 1994, AAAI.

[2]  Jian Cheng,et al.  Validating automated speaking tests , 2010 .

[3]  Su-Youn Yoon,et al.  Vocabulary Profile as a Measure of Vocabulary Sophistication , 2012, BEA@NAACL-HLT.

[4]  Su-Youn Yoon,et al.  Non-scorable Response Detection for Automated Speaking Proficiency Assessment , 2011, BEA@ACL.

[5]  L. Boves,et al.  Quantitative assessment of second language learners' fluency: comparisons between read and spontaneous speech. , 2002, The Journal of the Acoustical Society of America.

[6]  Anne Burns,et al.  English-for-Teaching: Rethinking Teacher Proficiency in the Classroom. , 2015 .

[7]  Maxine Eskénazi,et al.  An overview of spoken language technology for education , 2009, Speech Commun..

[8]  Klaus Zechner,et al.  Automated Content Scoring of Spoken Responses in an Assessment for Teachers of English , 2013, BEA@NAACL-HLT.

[9]  Ashish Verma,et al.  Sensei: Spoken language assessment for call center agents , 2007, 2007 IEEE Workshop on Automatic Speech Recognition & Understanding (ASRU).

[10]  Su-Youn Yoon,et al.  Assessment of ESL Learners' Syntactic Competence Based on Similarity Measures , 2012, EMNLP-CoNLL.

[11]  Su-Youn Yoon,et al.  Application of Structural Events Detected on ASR Outputs for Automated Speaking Assessment , 2012, INTERSPEECH.

[12]  Xiaoming Xi,et al.  Automatic scoring of non-native spontaneous speech in tests of spoken English , 2009, Speech Commun..

[13]  Su-Youn Yoon,et al.  Acoustic Feature-based Non-scorable Response Detection for an Automated Speaking Proficiency Assessment , 2012, INTERSPEECH.