LOS ANGELES» “Fear not! For I shall lead ye to riches beyond your wildest dreams!”
Inside a tiny recording booth in downtown Los Angeles, John Peck waited for a verdict from the voiceover engineer: Did the line sound pirate-y enough? Try again, the engineer suggested, perhaps with more throaty emphasis on “wildest.” It might make the animated character Peck was voicing — a buccaneer with a peg leg — a tiny bit funnier.
Peck, 33, cleared his throat and gave it a whirl, prompting chuckles from the production team. A couple of clicks on a laptop later, and an artificial intelligence tool synced Peck’s voice with a cartoon pirate’s mouth movements. The character was destined for an episode of “StEvEn & Parker,” a YouTube series about rapscallion brothers that attracts 30 million unique viewers weekly.
Just a few years ago, lip-syncing a minute of animation could take up to four hours. An animator would listen to an audio track and laboriously adjust character mouths frame by frame. But Peck’s one-minute scene took 15 minutes for the AI tool to sync, including time spent by an artist to refine a few spots by hand.
Toonstar, the startup behind “StEvEn & Parker,” uses AI throughout the production process — from honing story lines to generating imagery to dubbing dialogue for overseas audiences. “By leaning into the technology, we can make full episodes 80% faster and 90% cheaper than industry norms,” said John Attanasio, a Toonstar founder.
“This is how you build the next generation of hot intellectual property,” Attanasio added excitedly.
AI entrepreneurs love to pontificate about how AI is going to transform — transform! — the process of developing and producing movies and television shows. In the ears of Hollywood traditionalists, AI enthusiasts can morph into cartoon pirates arriving to plunder. Those irksome writers, visual effects specialists and sound editors who slow everything down and make everything so expensive with their precious creativity? Bots can replace most of them.
So far, though, not much has changed. The biggest studios continue to run in much the same way that they did before OpenAI and other artificial intelligence technology burst onto the scene. Human brains are still much better at creating art than virtual ones. Disney, Universal, Warner Bros. and the like are also trying to sort through major concerns about how generative AI software is built, how copyright holders are compensated and how unions might react.
But one thing has become clear: No part of the entertainment business has more to lose — and gain — from AI than animation.
The $420 billion global animation industry (movies, television cartoons, games, animé) has long been dominated by computergenerated imagery; Walt Disney Animation hasn’t released a hand-drawn film since 2011.
In other words, unlike much of Hollywood, animation companies are not technophobic.
Even with computers, however, the process of making an animated movie (or even a cartoon) remains extraordinarily expensive, requiring squadrons of artists, animators, graphic designers, 3D modelers and other craftspeople. Studios have a big incentive to find a more efficient way, and AI can already do many of those things far faster, with far fewer people.
Jeffrey Katzenberg, a former chair of Walt Disney Studios and a co-founder of DreamWorks Animation, has predicted that by next year, it will take only about 50 people to make a major animated movie, down from 500 a decade ago. If he were founding DreamWorks today, Katzenberg said of AI on a recent episode of the podcast “The Speed of Culture,” he would be “jumping into it hook, line and sinker.”
This is exactly what Toonstar has done — and the AI-assisted production process it has adopted offers a glimpse of where the giants might be headed.
What would the Jetsons do? Attanasio, 53, and Toonstar’s other founder, Luisa Huang, 46, started their careers at analog studios. He joined DreamWorks Animation’s marketing team in 2001, the year “Shrek” became a sensation. She got hired in Disney’s strategic planning department around the same time. They later worked at Warner Bros.
Put bluntly, they’re not carpetbaggers.
But they are defectors. “Hollywood avoids change to the point of paralysis,” Attanasio said. “We couldn’t take it anymore.”
Huang and Attanasio, both MBAs, left Warner Bros. in early 2015. “It was the year after Saturday morning cartoons ended — the last broadcast network airing them gave up because kids had started to live on YouTube,”
Huang said. “And we were like, ‘Wait, what’s happening to cartoons? They’re just going to die?’”
That question led to another one: What would Hanna-Barbera, one of the animation greats, look like today? Starting in the late 1950s, William Hanna and Joseph Barbera built a cartoon studio by changing the animation process to fit the transformational technology of the time: broadcast television.
Small screens needed more content than big ones — a string of episodes to fill a weekly time slot. And budgets for early television shows were tiny; the economics of the medium were being worked out on the fly.
“TV needed a much faster production schedule at a fraction of the cost,” Huang said.
Hanna-Barbera figured out how to make do with more rudimentary animation, cutting in half the number of original frames per second. The company also reused art from episode to episode, reducing the number of new drawings needed for a segment to around 2,000 from 14,000.
Perhaps surprisingly, it resulted in a golden age of animation.
Hanna-Barbera hits included “The Flintstones” (1960), “The Jetsons” (1962), “Scooby- Doo” (1969), “Super Friends”
(1973) and “The Smurfs” (1981), all of which continue to make money through reruns and merchandising.
Toonstar, founded in 2015 in a warehouse in downtown Los Angeles, would try to rethink animated content and production for the YouTube age.
“Hollywood mostly sat on its hands while distribution got upended through streaming,” Attanasio said. “We didn’t want to watch the same thing get repeated with emerging content creation technology.”
‘Die A.I. Scum’
Huang and Attanasio made shorts for the messaging app Snapchat, which became a Toonstar investor. There was a detour into blockchain-powered content. They dabbled in creating virtual influencers on Musical.ly, an app for lip-sync videos that later morphed into TikTok, and raised $5 million from venture firms like Greycroft.
By 2021, they had built an animation assembly line that was fast enough to make content for a social media audience. But one major choke point remained: Toonstar artists were having to scramble to create all of the necessary drawings. Put simply, AI still hadn’t advanced far enough to do the heavy lifting of image creation.
A more powerful breed of AI that could generate text and images in response to short prompts — generative AI — arrived in 2022. OpenAI, for instance, rolled out ChatGPT and DALL-E, which let people generate photorealistic images simply by describing what they wanted to see.
Toonstar used open source AI infrastructure to build its own image generator. Attanasio and Huang named it Ink & Pixel, a riff on a Disney department where a battalion of inkers and painters prepared cels, transparent sheets on which images were drawn or painted and then photographed.
“I get emotional because it was like, ‘Wow, it really works,’ ” Attanasio said when asked to recall his reaction when Ink & Pixel began to come together.
Just then, however, a virulent backlash against generative AI swept through Hollywood.
Creative workers viewed the technology as a threat to their very existence. More than 170,000 actors and writers spent much of 2023 on strike, in part to demand AI protections from studios.Attanasio publicly pushed back at the time. Stop worrying so much about AI taking jobs away, he wrote in a Hollywood trade publication, and embrace the ways the technology could “propel moviemaking forward.”
If costs are lower, he said, studios will start taking creative risks again.
“Die A.I. Scum,” one commenter replied.
Getting 70% there.
Toonstar relies on what is known as a “copyright clean” generative AI engine, which means Ink & Pixel has been trained exclusively on commissioned artwork rather than on data from the open internet and elsewhere. To put it another way, Ink & Pixel allows an infinite number of “StEvEn & Parker” images to be generated from a few hundred originals.
“It has to be artists leading the technology instead of the other way around,” Huang said. Toonstar’s watchword is HITL, she added: Humans in the Loop.
The production cycle for a “StEvEn & Parker” episode typically starts with a video call on Thursday. Parker James, a 24-year-old TikTok star and musician who created the series with Toonstar, joins from his home in Texas. A writer and Toonstar’s head of production (who also writes scripts) also attend. Together, they flesh out an episode of several minutes in length that will get uploaded to YouTube by the end of the following week.
“Sometimes my ideas aren’t fully formed,” James said. “Toonstar really sits and takes the time to develop something that everyone’s happy with.”
Scripts are never generated in whole or part with AI, according to Toonstar. But the company does use an AI analytics tool called Spot to inform stories.
Spot combs the viewer data from previous episodes to see what worked and what didn’t: What snippets did people rewatch? If they left before the end, what was happening in the story at the time? It turns out, according to Spot, that “StEvEn & Parker” viewers are meh on Murphy, the family dog. “That surprised us because who doesn’t like a cute dog?”
Huang said.
Scripts are written Friday.
(The one for the three-minuteand-30-second pirate episode took two hours to hammer out.) By Monday, AI-assisted preproduction begins. Toonstar computers analyze the script to see what cartoon images will need to be created — what doesn’t already exist in the Ink & Pixel system from past episodes. (A pirate, for example.) A “generative concept artist” then uses Ink & Pixel to produce the needed images. To create the pirate, prompts like “red headband,” “brown boots” and “scruffy beard” were typed into the system. Instantly, the AI engine kicked out several options, each in the art style of “StEvEn & Parker” and, crucially, with the correct proportions in relation to existing characters.
A designer selects one option and tweaks it by hand. To make the pirate more comical, he was given a peg leg, for example. It took about 20 minutes.
“This part is very subjective, which is why artists and designers matter,” Huang said. “Our goal is for AI to get it 70% there.”
Next, an animator uses proprietary software to build out scenes. Dialogue, recorded starting when scripts are finalized (though sometimes even earlier), is automatically synced to character lips using an AI tool. An AI program is also used to add sound effects, like a skateboard skidding down a curb.
“Traditionally, that person would have to go into a sound library and hunt and peck,” Huang said. Toonstar also uses AI wizardry from a startup called ElevenLabs to instantly dub character voices into Spanish, Portuguese and Japanese to distribute in other countries via YouTube.
By Thursday, the process starts again.
“Because animation has historically been very expensive and taken a long time to do, a lot of talent has been kept out — particularly young talent,” Attanasio said. “We don’t think animation should be a private club.
“And with AI,” he continued, “It’s not going to be.”