10 AI Tools We Actually Use in Video Production

10 AI Tools We Actually Use in Video Production

We test a lot of AI tools. Most get used once, slow us down, or fall apart the second you push them past a demo. Those get deleted. These ten stayed. Not because they’re flashy. Because they solve real production problems: Faster iteration Tighter turnarounds Outputs that would have cost 10× more two years ago This is not a hype list. This is how these tools show up inside a working video studio in Sydney and Hong Kong — what they’re good at, where they break, and how to use them without producing slop. Important context: Tools change weekly. This is what we’re using right now because it holds up under real briefs, real clients, and real pressure.

We test a lot of AI tools. Most get used once, slow us down, or fall apart the second you push them past a demo. Those get deleted. These ten stayed. Not because they’re flashy. Because they solve real production problems: Faster iteration Tighter turnarounds Outputs that would have cost 10× more two years ago This is not a hype list. This is how these tools show up inside a working video studio in Sydney and Hong Kong — what they’re good at, where they break, and how to use them without producing slop. Important context: Tools change weekly. This is what we’re using right now because it holds up under real briefs, real clients, and real pressure.

Ed Holmes

Ed Holmes

Feb 19, 2026

How We Think About AI Tools

Before the list, one thing matters more than the tool itself:

AI doesn’t make decisions. It multiplies them.

If you don’t know what you’re trying to say, these tools will just help you say nothing faster.

We keep tools that:

  • Improve realism or control

  • Reduce turnaround without killing quality

  • Make versioning and localisation practical

  • Solve a specific production bottleneck

If it’s fun but unreliable, it doesn’t survive the workflow.

The 10 AI Video Production Tools We Actually Use

1. Nano Banana Pro


Image generation that holds together


What it does:

High-quality image generation, manipulation, background swaps, style transfer, upscaling.


What it’s good for:

  • Style frames

  • Storyboards

  • Character consistency

  • Product consistency

  • Transforming existing imagery without collapse


Why we use it:

  • Images hold up at high resolution

  • Style transfer feels intentional

  • You can refine instead of regenerate blindly


Where it breaks:

Still images only. Movement lives elsewhere.


How it shows up in real work:

When a product or character needs to live in multiple worlds, Nano Banana lets us push environments while keeping the hero element stable. That makes client decisions faster and cheaper.

2. Kling


Default for realistic AI video


What it does:

Text-to-video and image-to-video generation.


What it’s good for:

  • Realistic lighting

  • Natural motion

  • Camera language


Why we use it:

  • Movement doesn’t immediately scream AI

  • Handles multi-element scenes well

  • Strong quality-to-speed ratio


Where it breaks:

Lip sync needs supervision. Character consistency requires direction.


How it shows up in real work:

Cutaways, environments, motion moments. We curate hard and only keep clips with weight.

3. Veo3


When lip sync actually matters


What it does:

Video generation with improved speech synchronisation.


What it’s good for:

Talking moments where credibility matters.


Why we use it:

  • Best lip sync we’ve tested

  • Handles longer dialogue beats cleanly


Where it breaks:

More expensive and slower. Not worth it unless speech is central.


How it shows up in real work:

Used when a line needs to land cleanly. Shots are still designed to avoid tight close-ups. Direction finishes the last 20%.

4. ElevenLabs


Voice generation that saves versioning


What it does:

AI voice generation across languages and accents.


What it’s good for:

  • Voiceover

  • Narration

  • Localisation

  • Scratch tracks


Why we use it:

  • Natural delivery

  • Fast iteration

  • Multi-language versions without chaos


Where it breaks:

Repeated use of the same voice becomes noticeable.


How it shows up in real work:

We keep edits moving with AI VO while approvals evolve. Human VO replaces it once scripts lock.

5. Suno


Music without licensing headaches


What it does:

AI-generated music based on mood and duration.


What it’s good for:

  • Social content

  • Corporate films

  • Internal videos

  • Event content

Why we use it:

  • Fast

  • Royalty-free

  • Easy to generate pacing variations


Where it breaks:

Not for emotionally defining or brand-critical work.


How it shows up in real work:

Functional music beds that support the story without becoming the story.

6. Topaz Labs


Rescue footage you’d otherwise bin


What it does:

Upscaling, denoising, sharpening, stabilisation.


What it’s good for:

Footage that matters emotionally but fails technically.


Why we use it:

  • Cleans phone footage

  • Saves older material

  • Upscales without destroying detail


Where it breaks:

Cannot fix fundamentally broken footage.


How it shows up in real work:

When something must stay in the cut, Topaz stops it from being distracting.

7. HeyGen

Avatars for consistency and speed


What it does:

AI avatar presenters with lip sync and multilingual output.


What it’s good for:

  • Training

  • Explainers

  • Corporate delivery


Why we use it:

  • Same presenter across languages

  • Predictable and repeatable


Where it breaks:

Not for emotional storytelling.


How it shows up in real work:

Straightforward delivery where clarity beats charisma.

8. Higgsfield

The switchboard


What it does:

Aggregates multiple video models in one interface.


What it’s good for:

  • Fast exploration

  • A/B testing engines

  • Direction-setting


Why it’s essential:

Most teams waste days jumping between tools. Higgsfield turns that into an afternoon.


Where it breaks:
Not a finishing tool. Taste still required.


How it shows up in real work:

We test widely, kill 90%, lock rules early, then scale production.

9. Claude


Scripts, structure, translation


What it does:

Writing, structuring, organising language.


What it’s good for:

  • Scripts

  • Treatments

  • Captions

  • Versioning


Why we use it:

  • Turns vague briefs into structure

  • Speeds up blank-page phase


Where it breaks:

Left alone, it becomes generic.


How it shows up in real work:

Used to think faster. Always rewritten by humans.

10. Descript


Editing without misery


What it does:

Edit video and audio by editing text.


What it’s good for:

  • Interviews

  • Webinars

  • Podcasts

  • Dialogue-heavy content


Why we use it:

  • Fast structural edits

  • Clean filler removal

  • Easy cutdowns


Where it breaks:

Not for complex cinematic finishing.


How it shows up in real work:

Find the story fast. Finish properly elsewhere.

How These Tools Actually Work Together


Typical AI-enhanced workflow:


Brief
→ Claude structures the idea
Look development → Nano Banana + Higgsfield
Video generation → Kling or Veo3
Audio → ElevenLabs + Suno
Cleanup → Topaz
Fast edits → Descript
Final polish → Human hands


AI handles the middle 70%.

The first 15% (direction) and last 15% (finish) remain human.

What We Don’t Use AI For

  • Final colour grade

  • Complex sound design

  • Hero emotional moments

  • Anything where “almost right” becomes brand risk

The Hard Truth


These tools do not make you better.

They make you faster.

If you use them to produce more mediocre work, you will just get there quicker.

We use AI to remove friction — so we can spend more time deciding what actually matters:

  • Which option has a pulse?

  • Which one deserves to exist?


That is still director work.

And it always will be.

Check more blogs

A quick overview of how we work together to make your edit best in class!