This Robot Learned to Talk by Watching Humans on YouTube

This Robot Learned to Talk by Watching Humans on YouTube

Scientists have created a robot that learns lip movements by watching humans rather than following preset rules. The breakthrough could help future robots feel more natural and emotionally engaging.

When people speak face to face, a surprisingly large share of attention is directed toward the movement of the lips. Robots, however, have struggled for decades to reproduce this basic part of communication. Even the most advanced humanoid machines often rely on stiff, exaggerated mouth motions that feel cartoonish, assuming they have a face at all.

Humans place enormous weight on facial expression, especially the mouth. An awkward step or clumsy hand motion is easy to ignore, but even a small error in facial movement quickly stands out. This sensitivity is part of what researchers call the “Uncanny Valley,” where robots appear unsettling instead of lifelike. Poor lip motion is a major reason many robots feel emotionless or eerie, but researchers say that barrier may finally be weakening.

A Robot That Learns Facial Motion

On January 15, researchers at Columbia Engineering announced a significant advance. They built a robot that can learn how to move its lips for speaking and singing rather than relying on preprogrammed rules. In a study published in Science Robotics, the team showed the robot forming words in multiple languages and even singing a track from its AI generated debut album “hello world_.”

The robot gained this ability through observation instead of instructions. It first learned how to control its own face by watching its reflection in a mirror, gradually understanding how its 26 facial motors shaped different expressions. After that, it studied hours of YouTube videos to observe how human mouths move during speech and song.

Read more

اپنا تبصرہ بھیجیں