Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleNovember 2022
A Deep Dive Into Neural Synchrony Evaluation for Audio-visual Translation
ICMI '22: Proceedings of the 2022 International Conference on Multimodal InteractionPages 642–647https://doi.org/10.1145/3536221.3556621We present a comprehensive analysis of the neural audio-visual synchrony evaluation tool SyncNet. We assess the agreement of SyncNet scores vis-a-vis human perception and whether we can use these as a reliable metric for evaluating audio-visual lip-...
- keynoteMarch 2021
"Now You're Speaking My Language": Towards a Seamless Localized Product Experience
CHIIR '21: Proceedings of the 2021 Conference on Human Information Interaction and RetrievalPage 3https://doi.org/10.1145/3406522.3444756Increased internet access and improvements in localization processes means that more people have access to information and entertainment than ever before. However, many people still find it difficult to navigate online and offline resources in their ...
- short-paperDecember 2020
See me Speaking? Differentiating on Whether Words are Spoken On Screen or Off to Optimize Machine Dubbing
ICMI '20 Companion: Companion Publication of the 2020 International Conference on Multimodal InteractionPages 130–134https://doi.org/10.1145/3395035.3425640Dubbing is the art of finding a translation from a source into a target language that can be lip-synchronously revoiced, i. e., that makes the target language speech appear as if it was spoken by the very actors all along. Lip synchrony is essential for ...
- research-articleJuly 2019
Text-based editing of talking-head video
- Ohad Fried,
- Ayush Tewari,
- Michael Zollhöfer,
- Adam Finkelstein,
- Eli Shechtman,
- Dan B Goldman,
- Kyle Genova,
- Zeyu Jin,
- Christian Theobalt,
- Maneesh Agrawala
ACM Transactions on Graphics (TOG), Volume 38, Issue 4Article No.: 68, Pages 1–14https://doi.org/10.1145/3306346.3323028Editing talking-head video to change the speech content or to remove filler words is challenging. We propose a novel method to edit talking-head video based on its transcript to produce a realistic output video in which the dialogue of the speaker has ...
- research-articleJuly 2018
Deep video portraits
- Hyeongwoo Kim,
- Pablo Garrido,
- Ayush Tewari,
- Weipeng Xu,
- Justus Thies,
- Matthias Niessner,
- Patrick Pérez,
- Christian Richardt,
- Michael Zollhöfer,
- Christian Theobalt
ACM Transactions on Graphics (TOG), Volume 37, Issue 4Article No.: 163, Pages 1–14https://doi.org/10.1145/3197517.3201283We present a novel approach that enables photo-realistic re-animation of portrait videos using only an input video. In contrast to existing approaches that are restricted to manipulations of facial expressions only, we are the first to transfer the full ...
- ArticleOctober 1995
Progressive video coding for storage applications
This paper addresses the problem of video coding for storage applications. Storage applications require several features to a coding scheme such as robustness to dubbing, trick modes and the possibility of progressive transmission with an exact rate ...