Enhancing the accessibility of hearing impaired to video content through fully automatic dynamic captioning

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In this paper we introduce an automatic subtitle positioning approach designed to enhance the video accessibility of deaf and hearing impaired people to multimedia documents. By using a dynamic subtitle and captioning approach, which exploits various computer vision techniques, including face detection, tracking and recognition, video temporal segmentation into shots and scenes and active speaker recognition, we are able to position each video subtitle segment in the near vicinity of the active speaker. The experimental evaluation performed on 30 video elements validates our approach with average F1-scores superior to 92%.

Original languageEnglish
Title of host publication2019 7th E-Health and Bioengineering Conference, EHB 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781728126036
DOIs
Publication statusPublished - 1 Nov 2019
Externally publishedYes
Event7th IEEE International Conference on E-Health and Bioengineering, EHB 2019 - Iasi, Romania
Duration: 21 Nov 201923 Nov 2019

Publication series

Name2019 7th E-Health and Bioengineering Conference, EHB 2019

Conference

Conference7th IEEE International Conference on E-Health and Bioengineering, EHB 2019
Country/TerritoryRomania
CityIasi
Period21/11/1923/11/19

Keywords

  • Active speaker recognition
  • Deaf and hearing impaired people
  • Dynamic subtitle

Fingerprint

Dive into the research topics of 'Enhancing the accessibility of hearing impaired to video content through fully automatic dynamic captioning'. Together they form a unique fingerprint.

Cite this