What is it about?

We use a large language model to analyze transcripts of student speech. Students responded using natural speech to describe a STEM vocabulary word or use it in a sentence. We scored student responses on a range of three (no understanding, some understanding, and mastered).

Featured Image

Why is it important?

We find that a language model can score students near identically to a trained human rater. This can greatly increase the scalability of assessing students from verbal responses. This could have far reaching ramifications in a number of fields where assessment of speech is time consuming.

Perspectives

This publication builds on a growing trend in my lab and others where machine learning has the ability to reliably replicate instructor assessment. While much research needs to be completed, I am incredibly impressed with PhD student Zhongdi Wu and his diligence in evaluating the work.

Eric Larson
Southern Methodist University

Read the Original

This page is a summary of: Towards Scalable Vocabulary Acquisition Assessment with BERT, July 2023, ACM (Association for Computing Machinery),
DOI: 10.1145/3573051.3596170.
You can read the full text:

Read

Contributors

The following have contributed to this page