Deepfakevideos enable an Asian news broadcaster bolster

Expansion Strategy with Minimal Investment

The Client

An Asian broadcaster with 1.3 billion viewers across 173 countries

An Asian broadcaster with channels spanning entertainment, news, and sports wanted to expand its reach. They needed an out-of-the-box solution to provide current affairs programs to an audience viewing content in 12 plus regional languages.

Building a Vision AI solution can be daunting and expensive.

Skills 0%
Internal governance 0%
Technology innovation 0%

Executive Summary

Industry Overview

Changing Consumer Behavior

In the post-pandemic ecosystem, with changing consumer habits, the industry is likely to focus on cost-efficiency, revenue enhancement opportunities, and profit protection with greater technology integration. For a 4-5 year period, the revenue growth globally is projected to be at 4.5% CAGR. Asia is expecting growth at 17% CAGR, and India at 11% CAGR over the same period amounting to INR 4.5 Trillion by 2023.

Business Challenge

Expansion Strategy

This prominent English news brand wanted to expand into the fast-growing regional market and establish themselves as a premier news source regionally. They were looking to broadcast news programs that focused on local events and test their regional expansion strategy by re-using newsroom footage with AI-powered synthetic speech and lip-sync.

The Akaike Edge

Lab-on-hire with pre-built solution accelerators

Akaike fills the supply gaps across the AI spectrum

Lab-on-hire 0%
Solution accelerators 0%
Rapid prototyping 0%


TTS and Video Synthesis using deep neural networks

Step 1.

Select target anchor panel from archives

The broadcaster had more than 260,000 hours of video in its archives. Focusing on the re-usability of the client’s media assets, from the available video footage, a few of the newsroom’s panel of anchors were selected.

Step 2.

A blended AI cocktail recipe

Post video selection, an AI recipe was whipped up for image synthesis and automated lip synchronization which blended Computer Vision, Deep Learning and GAN (Generative Adversarial Networks) technology.

Step 3.

Expressive speech using deep neural networks

The custom solution converted written text to natural-sounding speech. This was achieved by using deep neural networks trained on human speech to create human-like expressive speech. The target speech segment was then accurately adapted to a video with a speaking face using GAN.

Research shows that 87% of Vision AI
projects do not yield expected results

either owing to training data insufficiencies stalling the project, or being too slow to deploy. Our AI experts can help you accelerate in data sparse environments.

Have an Idea?

Let’s talk.

Akaike helps businesses and communities crack vast volumes of unstructured data using Vision AI to extract actionable insights, thereby improving the way they work, live, learn, and play.

We use cookies to give you the best experience.

Get in Touch