AI Article Synopsis

  • In virtual reality, a new system has been developed for generating talking face videos that accurately incorporates eye blinking, which is often overlooked in existing methods.
  • The system uses a combination of identity, audio, and blinking features to create realistic facial movements and allows for independent control of eye blinks.
  • Experimental results demonstrate that this approach can effectively produce lifelike talking faces with precise timing and patterns of blinking, enhancing the overall communication experience.

Article Abstract

In virtual reality, talking face generation is committed to using voice and face images to generate real face speech videos to improve the communication experience in the case of limited user information exchange. In a real video, blinking is an action often accompanied by speech, and it is also one of the indispensable actions in real face speech videos. However, the current methods either do not pay attention to the generation of eye movements, or cannot control the blinking in the generated results. To this end, this article proposes a novel system which produces vivid talking face with controllable eye blinks driven by the joint features including identity feature, audio feature, and blink feature. In order to disentangle the blinking action, we designed three independent features to individually drive the main components in the generated frame, namely the facial appearance, mouth movements, and eye movements. Through the adversarial training of the identity encoder, we filter out the information of the eye state from the identity feature, thereby strengthening the independence of the blinking feature. We introduced the blink score as the leading information of the blink feature, and through training, the value can be consistent with human perception to form a complete and independent control of the eyes. Experimental results on multiple datasets show that our method can not only reproduce real talking faces, but also ensure that the blinking pattern and time are fully controllable.

Download full-text PDF

Source
http://dx.doi.org/10.1109/TVCG.2022.3199412DOI Listing

Publication Analysis

Top Keywords

talking face
12
eye movements
12
face controllable
8
controllable eye
8
blinking feature
8
real face
8
face speech
8
speech videos
8
blinking action
8
identity feature
8

Similar Publications

Clinical Manifestations.

Alzheimers Dement

December 2024

Centre for Brain Research, Indian Institute of Science, Bangalore, Karnataka, India.

Background: Dementia, a global health challenge, drives the need for comprehensive understanding. Longitudinal cohort studies are vital, yet maintaining follow-up in dementia cohorts poses challenges. This study explores challenges in follow-up, refines protocols, and develops strategies that can elevate dementia research quality.

View Article and Find Full Text PDF

Nasalance is a valuable clinical biomarker for hypernasality. It is computed as the ratio of acoustic energy emitted through the nose to the total energy emitted through the mouth and nose (eNasalance). A new approach is proposed to compute nasalance using Convolutional Neural Networks (CNNs) trained with Mel-Frequency Cepstrum Coefficients (mfccNasalance).

View Article and Find Full Text PDF

Speech change is a biometric marker for Parkinson's disease (PD). However, evaluating speech variability across diverse languages is challenging. We aimed to develop a cross-language algorithm differentiating between PD patients and healthy controls using a Taiwanese and Korean speech data set.

View Article and Find Full Text PDF

This case report discusses a unique presentation of an artery of Percheron (AOP) infarct resulting in rapidly resolving internuclear ophthalmoplegia (INO) without classical signs. This is the case of a 70-year-old male patient who presented to a community Emergency Department following acute code stroke activation. Physical exam and imaging studies including non-contrast CT, CT angiography, CT perfusion, and MRI were performed.

View Article and Find Full Text PDF

Objective: In preterm and very low birth weight (VLBW) infants, attention-related problems have been found to be more pronounced and emerge later as academic difficulties that may persist into school age. In response, based on three attention networks: alerting, orienting, and executive attention, we examined the development of attention functions at 42 months (not corrected for prematurity) as a follow-up study of VLBW ( = 23) and normal birth weight (NBW:  = 48) infants.

Method: The alerting and orienting attention networks were examined through an overlap task with or without warning signal.

View Article and Find Full Text PDF

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!