• There are no suggestions because the search field is empty.

IBC2021 Accelerator: Smart Remote Production For Real-Time Animation

Jan 10, 2022 9:26:25 AM

In the latest publication of IBCDaily, an article appeared about the "Smart Remote Production for Real-Time Animation" Accelerator, which Respeecher was involved in.

Together with VRT, Rai - Radiotelevisione Italiana, Yle, RTE, European Broadcasting Union, Digital Domain, RADiCAL, NVIDIA and Epic Games/Unreal Engine, we will compete against other accelerators on this year's IBC event for the "Media Innovation Project of the year Award".

For this we showcase and compare two different workflows of a low budget, remote pipeline, using off the shelf tools and the power of smartphones and AI.

Read the full article below.
 
 

Accelerating the animation pipeline without increasing the amount of kit required is the principal aim of the Smart Remote Production For Real-Time Animation Accelerator project, which is bringing together leading broadcasters and vendors to leverage the latest technological developments of markerless motion capture and speech-driven facial animation to drive CG performances in the Unreal real-time render engine.

“We don’t usually work directly with vendors, so this is very interesting to us,” says RAI’s Roberto Iacoviello.

Indeed, Italy’s RAI is just one of a strong broadcast contingent of Accelerator Champions that also include RTÉ, VRT, YLE, and the EBU, working alongside the Entertainment Technology Center at USC, Digital Domain, and Unreal Engine developer Epic Games.

With participants including RADiCAL and Respeecher, with guidance from IBC Accelerator supporter Nvidia, this Accelerator represents a powerful conglomeration of talent and expertise that is looking to create the most effective, low-cost pipeline that they can to take material from script to 3D character in a real-time, distributed workflow environment.

Low cost and accessible

One of the key elements to this Accelerator is the phrase ‘low-cost’. As well as testing the feasibility of using vocal performance and body posture to drive 3D avatars from remotely connected locations, its stated aim is to do all this while using minimal equipment.

Even more so, according to one of the project leads, RTÉ’s Ultan Courtney, the wish is to utilize technology that exists less in production facilities and more in the pockets of the people that work there and their audiences.

“As opposed to having studios doing volumetric capture and other high-end techniques, we want to look at a workflow that engages millions of people,” he says.

“Most households have a smartphone, for example, so we’re working on that basis and what is the most accessible, highest tech we can work with to tell stories and communicate.”

This changes the technological emphasis of performance capture markedly. As opposed to a high technology solution involving wearing marker-based motion capture suits in a precisely delineated and monitored capture volume, the tech emphasis is shifted to the AI processing of a standard video signal.

The AI tracks the performance and converts that into an avatar, mapping limb and facial movements onto a virtual character.

“This is the difference between now and a few years ago; there is a lot of AI technology that can help us at a low budget,” explains Iacoviello. “There are three aspects where AI has made a lot of progress: marker-less capture, emotional response and speech animation.”

Voice cloning

Speech is one of the areas where this Accelerator is using interesting new tools. Respeecher is software that effectively clones voices using Deep Learning techniques and has already been used on The Mandalorian to create the voice of a young Luke Skywalker and by the NFL to recreate the voice of deceased football coach Vince Lombardi at the 2021 Super Bowl. The process is fascinating, and on the voice cloning FAQ page, you can discover how the de-aging of voice works, as well as the entire voice cloning process.

“In this project the idea is to make one source speaker be able to speak in several different voices for the POC piece,” says company co-founder Alex Serdiuk.

“To date we’ve offered white glove service where we need to be involved, but this uses our new self-serve Voice Marketplace. There are many challenges for us with this as we lose a lot of control over the recording conditions and performance levels of the actors involved, but it is important to be able to democratise these tools so that they can be used in the sort of environments that ideally would require just an iPhone.”

Of course, one of the challenges any Accelerator team faces is that while the tools may be available to make the POC a success, getting them to work together in a designated workflow is not always easy.

Sometimes they are lucky and the work has already been done. For instance, once motion has been captured using RADiCAL and uploaded to the cloud, RADiCAL provides an animated FBX file which can be used in both Omniverse and Unreal to retarget the animation to CG characters. 

In this case, as an illustration, the wrnch CaptureStream iOS-based AI capture software can already export into Nvidia’s Omniverse real-time simulation and collaboration platform via the wrnch AI Pose Estimator. 

However, other areas of the workflow still need work. Despite Omniverse’s success at becoming a universal solvent for 3D tools, there are always challenges.

For example, can the team’s Pose Estimation workflow be easily connected to the speech-to-facial animation workflow that the team is working on? And can the resulting characters then be imported from Omniverse and integrated into Unreal environments?

“We are pushing things forward but not everything is designed to be opened up to other uses. There are a few steps forward and back, and that’s where the Accelerator helps; having a target forces us to crash through all these problems,” says Courtney.

Multiple use cases

As the RAI team - which also includes Alberto Ciprian and Davide Zappia - points out, the TV market is not the only one that will be interested in the concept of realistic realtime character animation.

Virtual influencers, for example, become a distinct possibility, while its use in high-end pre-production or even to drive real-time performances in mixed reality LED capture volumes is fairly compelling. “When you go into the digital world the use cases are limited only by your imagination,” says Zappia. 

Paola Sunna, senior project manager at EBU Technology & Innovation, says: “I was really happy to see so many broadcasters involved in this project. What I really hope already is that we keep going on this project even after IBC is over. We are learning so many things.”

For more information on the IBC Accelerator Media Innovation Programme, supported by Nvidia, visit the IBC website.

Smart Remote Production For Real-Time Animation

Champions: RTÉ, EBU, RAI, VRT, YLE, ETC/USC, Digital Domain, Unreal/Epic Games

Participants: Respeecher, RADiCAL

Image of Orysia Khimiak
Orysia Khimiak

PR and Comms Manager

For the past 9 years, have been engaged in Global PR of early stage and AI startups, in particular Reface, Allset, and now Respeecher. Clients were featured in WSJ, Forbes, Mashable, the Verge, Tech Crunch, and Financial Times. For over a year, I Orysia been conducting PR Basics course on Projector. During the war, became more actively involved as a fixer and worked with the BBC, Guardian and The Times.

Related Articles
Respeecher's Commitment to Ethical Voice Cloning: Key Initiatives

Jun 13, 2023 9:49:20 AM

Ethical voice cloning goes beyond getting written consent from voice owners and working...

Respeecher Champions Ethical Synthetic Media with PAI's Responsible Practices Framework

Apr 10, 2023 8:56:07 AM

At Respeecher, we strive to revolutionize voice cloning for content creators everywhere....

Code[ish] Podcast: The Ethical and Technical Side of Deep Fakes featuring Respeecher

Jan 28, 2021 8:02:06 AM

We’ve recently been invited to talk about deep fakes at Code[ish], a podcast created by...