Google’s generative video model Veo 3 has a subtitles problem

MIT Technology Review - AI
Jul 15, 2025 14:40
Rhiannon Williams
1 views
airesearchtechnology

Summary

Google’s new generative video model, Veo 3, introduced the ability to create sounds and dialogue alongside hyperrealistic video clips, quickly attracting attention from creatives. However, the model struggles with generating accurate subtitles, highlighting ongoing challenges in synchronizing audio and text in AI-generated content. This limitation points to the need for further advancements in multimodal AI systems for seamless video production.

As soon as Google launched its latest video-generating AI model at the end of May, creatives rushed to put it through its paces. Released just months after its predecessor, Veo 3 allows users to generate sounds and dialogue for the first time, sparking a flurry of hyperrealistic eight-second clips stitched together into ads, ASMR videos,…

Related Articles

I've been coding with AI for two years. Here is what I've learned

Hacker News - AIJul 17

In this article, the author reflects on two years of coding with AI tools, noting significant productivity gains and the ability to tackle more complex projects. However, they emphasize that human oversight remains crucial, as AI-generated code can introduce subtle errors. The piece highlights the growing importance of AI as a coding assistant, while underscoring the need for developers to maintain strong foundational skills.

7 AI features coming to iOS 26 that I can't wait to use (and how you can try them)

ZDNet - Artificial IntelligenceJul 17

Apple is introducing seven new AI-powered features in iOS 26, including the standout "Hold Assist," which promises to save users significant time. These enhancements reflect Apple's commitment to integrating advanced AI into everyday tasks, signaling increased competition and innovation in the mobile AI space.

Cheating? Or the acumen of modern programming? FOSS, "AI", and human conscience

Hacker News - AIJul 17

The article explores the ethical debate around using AI tools and open-source software (FOSS) in programming, questioning whether leveraging such technologies constitutes cheating or simply reflects modern programming practices. It highlights the evolving role of human conscience in determining the boundaries of acceptable AI-assisted work. The discussion underscores the need for clearer guidelines as AI becomes increasingly integrated into software development.