Ultra-low bitrate video conferencing using deep image animation

Goluck Konuko, Giuseppe Valenzise, Stéphane Lathuilière

Research output: Contribution to journalConference articlepeer-review

Abstract

In this work we propose a novel deep learning approach for ultra-low bitrate video compression for video conferencing applications. To address the shortcomings of current video compression paradigms when the available bandwidth is extremely limited, we adopt a model-based approach that employs deep neural networks to encode motion information as keypoint displacement and reconstruct the video signal at the decoder side. The overall system is trained in an end-to-end fashion minimizing a reconstruction error on the encoder output. Objective and subjective quality evaluation experiments demonstrate that the proposed approach provides an average bitrate reduction for the same visual quality of more than 80% compared to HEVC.

Original languageEnglish
Pages (from-to)4210-4214
Number of pages5
JournalICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings
Volume2021-June
DOIs
Publication statusPublished - 1 Jan 2021
Event2021 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2021 - Virtual, Toronto, Canada
Duration: 6 Jun 202111 Jun 2021

Keywords

  • Deep learning
  • Model-based compression
  • Video compression
  • Video conferencing

Fingerprint

Dive into the research topics of 'Ultra-low bitrate video conferencing using deep image animation'. Together they form a unique fingerprint.

Cite this