With Synthesia 3.0, the next era of video is here

October 1, 2025

Create AI videos with 230+ avatars in 140+ languages.

Try Free AI Video
Get Started for FREE
Get started

For nearly a century, video has stayed the same. The format is still a one-way broadcast - recorded once and played back, over and over. Updating audio or visual elements isn’t easy, and you can’t interact, or measure your learning.

AI is about to change that. Instead of watching a static video, you’ll be able to interact in real time: talk to it, ask a question, or let the video ask you one. It will be highly personalized, with the full business context needed to train teams, and solve problems.

This is an entirely new media format, and today we’re excited to show you Synthesia 3.0, our AI video platform reimagined for the future of video, and some of the new products based on this vision.

Watch the video below for more information and read on for a quick breakdown of what’s new in Synthesia 3.0:

Introducing Video Agents

Video Agents are the first step toward video becoming a two-way conversation. They can be inserted at any point in your video - right in the Synthesia editor - to begin a real-time conversation with the viewer. And you can combine them with any other type of content, like static video and interactive scenes.

Video Agents can talk, listen, and act in real time, bringing a level of two-way interactivity that feels like a live conversation. They can run training sessions, screen job candidates, and even guide customers through learning experiences. 

But the real breakthrough is in a new category of data. Video Agents operate with specific knowledge of your business. That means they can capture data in real time, feed it back into your systems, and take any action you assign. With a full understanding of your business, this functionality can easily automate and scale repetitive processes - giving your team time back to work on what matters most.

New features coming to the Synthesia platform

To make Video Agents possible, we’re bringing new capabilities across the entire Synthesia platform. Here’s an overview of the most significant updates:

Avatars

Our new Express-2 avatars move and talk like professional speakers, with facial expressions, perfect lip sync, and natural hand and body gestures. We’re also adding many new features to our Avatar offering to allow for more customization:

You’ll now be able to create entirely new Avatars with a single prompt and place them into any environment. The lighting, depth, and perspective look so real that they blend in naturally, as if they were filmed on location.

You can prompt avatars in an office, at the beach or a construction site. You can also prompt an outfit to match that scene.

With Veo 3 integration, you can now prompt B-roll footage of your avatar demonstrating a task, or taking a specific action - like walking, driving, cooking, or anything else you can imagine.

You can create Personal Avatars more easily, AI versions that look and sound like you: rather than having to upload a video, you can just upload a single image, and have access to all the same customization features described above.

Voices

Our Express-Voice proprietary voice model creates your perfect voice clone in seconds, matching your tone and preserving your dialect, accent, and rhythm. 

Interactivity

Get viewers to engage and take action with clickable video elements like call-to-actions and branching.

Dubbing

Translations with frame-accurate lip sync. Upload a video, and it will automatically translate into 30+ languages, and localize, with perfect voice and lip sync.

We’re also working on additional features that will be coming to Synthesia 3.0 in 2026:

Copilot is your very own professional video editor. It writes your script in seconds, connects to knowledge bases, and suggests visual elements.

Courses reimagines workplace learning by designing interactive experiences that combine Avatars, Video Agents, and Interactivity, and measure skill development.

The future of video

A century ago, the first TV broadcast transformed communication. 

With Synthesia 3.0, we’re transforming it again.

In the future, we won’t be limited by the constraints of working with cameras. Soon, static video will be a relic of the past. We won’t just watch, we’ll interact. With every video, we’ll engage more, understand much better, and learn much faster.

At Synthesia, we’re incredibly excited to build this new video format with you.

About the author

CEO & Co-Founder of Synthesia

Victor Riparbelli

Meet Victor Riparbelli, one of the visionary authors contributing to our blog. With a rich background in technology entrepreneurship spanning over a decade, Victor co-founded Synthesia in 2017 alongside Prof. Matthias Niessner, Prof. Lourdes Agapito, and Steffen Tjerrild. A true pioneer in the field, Victor marries technical expertise, academic knowledge, and entrepreneurial spirit to redefine the possibilities of AI. His commitment to ethical and responsible AI innovation is at the heart of his contributions. Explore Victor's insights as he guides us through the exciting world of AI video creation and enterprise solutions.

Go to author's profile
Get started

Make videos with AI avatars in 140+ languages

Try out our AI Video Generator

Create a free AI video
Create free AI video
Create free AI video
Unmute

Trusted by 50,000+ teams.

faq

Frequently asked questions