Still images come to life in an AI model developed by Samsung at its Moscow research lab, ZDNET reports. Researchers claim with 32 or more reference frames (about a second of video), the results can be highly realistic.
With as little as a single frame, speech-like sequences can be generated from reference images, such as a rare photo.
Samsung suggests several applications: improved video calls, better player presence in multi-player games, and more believable character special effects in film and television.
Others say it adds to the arsenal for deep fakes, deceptive imagery intended to misrepresent statements or actions by well-known personalities, including politicians and authority figures.
The system works by extracting features from the source picture and mapping them to expressions and movements drawn from a huge database of images. It uses as AI technology known as generative adversarial networks, often more simply called GANS.
The research paper calls the approach ‘few-shot adversarial learning.’
OUR TAKE
- This is a great illustration of the two-edged sword of many AI advances. The same technology benefiting ‘video presence’ for consumer products also makes deepfakes easier to create
Samsung uses AI to transform photos into talking head videos
ZDNET | May 23, 2019 | by Campbell Kwan
SEE ALSO
- New AI can create fake videos of people from a single picture
BIG THINK | May 24, 2019 | by Paul Ratner | ‘The target face will do anything done by the source face, which can be any talking head.’ - ‘Few-Shot Adversarial Learning of Realistic Neural Talking Head Models’ | Original research paper