How to integrate
AI into Video production in 2024?

How we handle traditional video production is on the cusp of a creative & technological evolution. 

We're trying not to sensationalise this, or add further hysteria to the overhyped AI train, but as someone who has spent 25 years working in visual content, video creation, videoSEO, and filmmaking - this truly will be a transformative year for our industry.

As we dive headfirst into 2024, it's crucial that marketing leaders, CMO's, agency producers, and production companies, understand and embrace the potential value of Video GenAI.

There are loads of aspects of videoAI that are 'not quite there yet' - but we're mapping the exponential improvement of GenAI, and video is just the next aspect that's ready for revolution.

Here's some of the key features we think we're going to see this year:


  1. AI and SEO: Revolutionizing Video Discovery
  2. AI in 'functional' Video Content
  3. Multimodal Models: Next-Level Video Understanding
  4. The Emergence of AI-Driven Video Generators
  5. AI-Powered Live Streaming Experiences
  6. The Rise of Headless Channels on YouTube
  7. Enhanced Realism through AI-Driven Photogrammetry
  8. Redefining Virtual Production with AI

& AI driven search

Computer vision has been used in video and visual SEO content analytics for years - however in 2024 this capability will become available to more people!

The integration of AI in computer vision is a game-changer for SEO, as well as for in visual and video content. AI will enable teams to develop faster and more commercially trained detection models. 

VideoSEO will work on 2 fronts:

  1. Better visual mark-up, with automatic tagging and categorisation of owned content, making content more searchable and more quickly accessible. 
  2. More rapid and regular analysis of published visual content, allowing for content strategists to understand better search intent aligned to images and videos. 

It's a new way of doing 'function' video content!

Functional content is everywhere. From SEO informational to B2B onboarding, to educational 'how to' content. And videos are the perfect way of conveying these concepts quickly. As VideoAI becomes more controllable and more consistent - this 'option' will become more popular with brands, agencies, and production outfits.

In 2024 we're going to see Brands and Agencies adopt hybrid production workflows with functional and informational video content streams.

Leaning into a mix of AI processes to create talking heads, animation, AI Broll - all pulled together using AI augmented production processes..

These functional pieces can be hyper local, global, and all rendered in near realtime. 

We may also see a resurgence of interactive functional videos. Where targeted visuals, scenes, and messaging can be truly individual.

Multimodal is the next level in visual understanding

In AI circles these days, the term “large language model” (LLM) is frequently used to describe any advanced AI model tool.

This made sense for the past few years, as most AI tools were focused on text-only models, across multiple industries and actions. But this hasn't yielded the best results for motion AI as of yet? Why?

Well...Imagine we see an image of a ball, with blue sky, and the background of a park with trees. And we asked a Human and an AI to decide what would happen next.....

The human will assume a few things about the potential action from commons sense of the world, and the context of the image (things like gravity, ball material, height, etc...). The AI would simply represent predictive knowledge of the context (have I seen this action before) or a linear movement, instructed by the input. 

And this is how VideoAI tools have been working in 2023. 

But in 2024 modern models will be multimodal engines - so, trained on text, images, video, 3D, photogrammetry, audio, music, physical parameters and more - and precisely tuned on variables that serve to enhance these specific industries. They're far more than just language models, they are General Model Worlds (GMW). And that's what we need for effective commercial use cases. 

The saturation of AI-Driven Video Generators

Platforms like Runway, Pika, Google Gemini, and Stable Video Diffusion are at the forefront of AI in video generation.

With MidJourney entering the space in Feb, we are likely to see an uplift in demand, use, and expectations.

These platforms will offer advanced tools for creating high-quality video content at a fraction of the current time and cost. We're expecting AI generators to provide customizable video templates (camera movements, segmentation, FPS, etc), as well as automated editing, and even script-based video creation, making the process more efficient and accessible.

We are also going to see brands and agencies seek to control their video, image, and content pipelines - allowing for more control and hybridisation with other creative technologies and integrated tech stacks. We're supporting inhouse studios integrate modular Video and imagery pipelines into their existing process maps, and PIM toolkits. It's a brave, exciting new world! 

AI-Powered 'Live Streaming' Experiences

LCM's and near realtime output generation is opening the door to some truly amazing creative applications for streamers and live event creators.

AI will bring a new dimension by offering more interactive and personalized experiences for creators and viewers. YouTube announced they will be releasing a complete collection of new AI influenced filters, tools, and settings that will augment directly into their YouTube Studio suite. 

The Nvidia Broadcast toolkits is the first step in empowering the end user with AI enhanced realtime streaming techniques, such as the amazing eye alignment tool, as well as AI video denoising, graphical generation and overlays, and digital tracking features.

Twitch has been working with a number of creators and developers to generate a utopia of streaming technologies such as: Live voice and audio management, chat screen automation, live graphics applications, automatic shorts, clipping, and redistribution,  and even live music generation at certain times! 

This is mirroring the expectations of live sports broadcasting toolkits!

Increased faceless channels across YouTube

Faceless channels are a great way of building (and understanding) niche audiences.

AI will enable creators to generate more niche, engaging, and varied content. The integration will be deeper than the content level.

In embracing these AI-driven changes, it's important to view AI not just as a tool, but as a collaborative partner in the creative process. The integration of AI into video production promises a future of enhanced creativity, efficiency, effectiveness, and inclusivity.

AI augmented, radiance fields

Photogrammetry will make significant progress in 2024 - in video production and XR experiences.

We're extremely excited by the opportunities and potential of adding AI & photogrammetry. As an augmentation to geometry based digital environments, photogrammetry is acts as a 360 virtual space, generated by millions of points in 3D space. This method of visual representation is much lighter (file size) than traditional geometry, and can more quickly represent real world natural environments, weather systems, and landscapes. AI tools can augment and alter these environments in post processing. AI photogrammetry will help revolutionise virtual production, set extensions, and green screen digital replacement.

Applying AI to the Virtual Production pipeline

Perhaps some of the most interesting advancements in virtual production will be AI integration. Whether that's AI directly influencing content on the volumes. We have been testing and experimenting with AI 2D background plates for months. Tools such as Cuebric have helped bring simple and stable UI's to VP infrastructure - and we cant wait to help refine this workflow more. 

OperationalAI functions helping enhance and improve various phases in the virtual production process. We will be integrating AI speed to illustration to help augment on pre-production and art direction.   

As generative AI models become bigger, it's inevitable that environmental impact is going to influence how we train future models.

How can small agencies integrate AI into their working pipeline without risking personality?

What does "VideoAI" actually mean? What's the landscape for videoAI companies in 2024?

It works in a similar way to diffusion, but instead of making a single full image - it creates the content in multiple phases.

Salesforce & McKinsey say that 86% Marketers expect GenAI will be in their commercial activities this year.

How will GenAI influence visual creative marketing content, processes, & activities in 2024?

The future of videoSEO is undoubtedly going to be shaped by AI in both analysis and optimisation.

In this article we'll delve into where the Virtual Production industry will develop in 2024.

Fancy a coffee and a chat?

If you have a rough brief, drop us a line and we'd be happy to talk through it.

Innovation discovery:

We're great at discovering growth too. Let's chat about a potential collab.

How to integrate AI into Video in 2024?
This website uses cookies to improve your experience. By using this website you agree to our Data Protection Policy.
Read more