Microsoft Research Asia released a new paper introducing VASA, a framework for generating lifelike talking faces. The researchers presented their model, dubbed VASA 1, that can generate realistic videos based only on a single static image and a speech audio clip. The full paper is available at arXiv . The results are impressive and beat all previous tools that use generative artificial intelligence to produce realistic deepfakes. What is particularly interesting about VASA 1 is the overall a

Read the full article at Neowin