“Choices? That's the dream”: challenges and opportunities in non-speech information closed-captioning

  • Lloyd May
  • , Michael Clemens
  • , Khang Dang
  • , Keita Ohshiro
  • , Sripathi Sridhar
  • , Pauline Wee
  • , Magdalena Fuentes
  • , Sooyeon Lee
  • , Mark Cartwright

Research output: Contribution to journalArticlepeer-review

Abstract

Introduction: Access to non-speech information (NSI) in video content is essential to creating accessible and engaging video content, particularly for D/deaf and Hard-of-Hearing (DHH) audiences. In this paper we present an overview of the current state of NSI captioning research, professional practice, and user preferences. Methods: We utilized a comprehensive review approach that combined a systematic literature review methodology with a mixed-methods survey and interview study. 1276 papers were screened with 36 eligible for the final inductive best fit analysis. 168 DHH participants completed an online survey and 15 participated in semi-structured interviews. Additionally, 5 professional captioners participated in semi-structured interviews. Results and discussion: We offer systematic insights into the current challenges related to NSI captioning faced by DHH users and professional captioners, trends in recent NSI captioning research, as well as opportunities for future work that enhance user agency, utilize integrated research methodologies, and broaden community involvement.

Original languageEnglish (US)
Article number1575176
JournalFrontiers in Computer Science
Volume7
DOIs
StatePublished - 2025
Externally publishedYes

All Science Journal Classification (ASJC) codes

  • Computer Science (miscellaneous)
  • Human-Computer Interaction
  • Computer Vision and Pattern Recognition
  • Computer Science Applications

Keywords

  • accessibility
  • closed-captioning
  • deaf and hard-of-hearing
  • music
  • non-speech information
  • sound effects
  • subtitles

Fingerprint

Dive into the research topics of '“Choices? That's the dream”: challenges and opportunities in non-speech information closed-captioning'. Together they form a unique fingerprint.

Cite this