Skip to main content

Learning realistic lip motions for humanoid face robots.

Publication ,  Journal Article
Hu, Y; Lin, J; Goldfeder, JA; Wyder, PM; Cao, Y; Tian, S; Wang, Y; Wang, J; Wang, M; Zeng, J; Mehlman, C; Wang, Y; Zeng, D; Chen, B; Lipson, H
Published in: Science robotics
January 2026

Lip motion represents outsized importance in human communication, capturing nearly half of our visual attention during conversation. Yet anthropomorphic robots often fail to achieve lip-audio synchronization, resulting in clumsy and lifeless lip behaviors. Two fundamental barriers underlay this challenge. First, robotic lips typically lack the mechanical complexity required to reproduce nuanced human mouth movements; second, existing synchronization methods depend on manually predefined movements and rules, restricting adaptability and realism. Here, we present a humanoid robot face designed to overcome these limitations, featuring soft silicone lips actuated by a 10-degree-of-freedom mechanism. To achieve lip synchronization without predefined movements, we used a self-supervised learning pipeline based on a variational autoencoder (VAE) combined with a facial action transformer, enabling the robot to autonomously infer more realistic lip trajectories directly from speech audio. Our experimental results suggest that this method outperforms simple heuristics like amplitude-based baselines in achieving more visually coherent lip-audio synchronization. Furthermore, the learned synchronization successfully generalizes across multiple linguistic contexts, enabling robot speech articulation in 10 languages unseen during training.

Duke Scholars

Published In

Science robotics

DOI

EISSN

2470-9476

ISSN

2470-9476

Publication Date

January 2026

Volume

11

Issue

110

Start / End Page

eadx3017

Related Subject Headings

  • Speech
  • Robotics
  • Movement
  • Motion
  • Lip
  • Learning
  • Language
  • Humans
  • Face
  • Equipment Design
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Hu, Y., Lin, J., Goldfeder, J. A., Wyder, P. M., Cao, Y., Tian, S., … Lipson, H. (2026). Learning realistic lip motions for humanoid face robots. Science Robotics, 11(110), eadx3017. https://doi.org/10.1126/scirobotics.adx3017
Hu, Yuhang, Jiong Lin, Judah Allen Goldfeder, Philippe M. Wyder, Yifeng Cao, Steven Tian, Yunzhe Wang, et al. “Learning realistic lip motions for humanoid face robots.Science Robotics 11, no. 110 (January 2026): eadx3017. https://doi.org/10.1126/scirobotics.adx3017.
Hu Y, Lin J, Goldfeder JA, Wyder PM, Cao Y, Tian S, et al. Learning realistic lip motions for humanoid face robots. Science robotics. 2026 Jan;11(110):eadx3017.
Hu, Yuhang, et al. “Learning realistic lip motions for humanoid face robots.Science Robotics, vol. 11, no. 110, Jan. 2026, p. eadx3017. Epmc, doi:10.1126/scirobotics.adx3017.
Hu Y, Lin J, Goldfeder JA, Wyder PM, Cao Y, Tian S, Wang Y, Wang J, Wang M, Zeng J, Mehlman C, Zeng D, Chen B, Lipson H. Learning realistic lip motions for humanoid face robots. Science robotics. 2026 Jan;11(110):eadx3017.

Published In

Science robotics

DOI

EISSN

2470-9476

ISSN

2470-9476

Publication Date

January 2026

Volume

11

Issue

110

Start / End Page

eadx3017

Related Subject Headings

  • Speech
  • Robotics
  • Movement
  • Motion
  • Lip
  • Learning
  • Language
  • Humans
  • Face
  • Equipment Design