
You might not know Viggle AI, but youāve likely seen the viral memes it created. The Canadian AI startup is responsible for dozens of videos remixing the rapper Lil Yachty bouncing onstage at a summer music festival. In one video, Lil Yachty is replaced by Joaquin Phoenixās Joker. In another, Jesus seemed to be hyping the crowd up. Users made countless versions of this video, but one AI startup was fueling the memes. And Viggleās CEO says YouTube videos fuel its AI models.
Viggle trained a 3D-video foundation model, JST-1, to have a āgenuine understanding of physics,ā as the company claims in its press release. Viggle CEO Hang Chu says the key difference between Viggle and other AI video models is that Viggle allows users to specify the motion they want characters to take on. Other AI video models will often create unrealistic character motions that donāt abide by the laws of physics, but Chu claims Viggleās models are different.
āWe are essentially building a new type of graphics engine, but purely with neural networks,ā said Chu in an interview. āThe model itself is quite different from existing video generators, which are mainly pixel based, and donāt really understand structure and properties of physics. Our model is designed to have such understanding, and thatās why itās been significantly better in terms of controllability and efficiency of generation.ā
To create the video of the Joker as Lil Yachty, for instance, just upload the original video (Lil Yachty dancing onstage) and an image of the character (the Joker) to take on that motion. Alternatively, users can upload images of characters alongside text prompts with instructions on how to animate them. As a third option, Viggle allows users to create animated characters from scratch with text prompts alone.
But the memes are only a small percent of Viggleās users; Chu says the model has seen wide adoption as a visualization tool for creatives. The videos are far from perfect ā theyāre shaky and the faces are expressionless ā but Chu says itās proven effective for filmmakers, animators and video game designers to turn their ideas into something visual. Right now, Viggleās models only create characters, but Chu hopes to enable more complex videos later on.
Viggle currently offers a free, limited version of its AI model on Discord and its web app. The company also offers a $9.99 subscription for increased capacity, and gives some creators special access through a creator program. The CEO says Viggle is talking with film and video game studios about licensing the technology, but he also is seeing adoption amongst independent animators and content creators.
On Monday, Viggle announced it had raised a $19 million Series A led by Andreessen Horowitz, with participation from Two Small Fish. The startup says this round will help Viggle scale, accelerate product development and expand its team. Viggle tells Dakidarts that it partners with Google Cloud, among other cloud providers, to train and run its AI models. Those Google Cloud partnerships often include access to GPU and TPU clusters, but typically not YouTube videos to train AI models on.
Training data
āSo far weāve been relying on data that has been publicly available,ā said Chu, relaying a similar line to what OpenAIās CTO Mira Murati answered about Soraās training data.
Asked if Viggleās training dataset included YouTube videos, Chu responded plainly: āYeah.ā
That might be a problem. In April, YouTube CEO Neal Mohan told Bloomberg that using YouTube videos to train an AI text-to-video generator would be a āclear violationā of the platformās terms of service. The comments were in the context of OpenAI potentially having used YouTube videos to train Sora.
Mohan clarified that Google, which owns YouTube, may have contracts with certain creators to use their videos in training datasets for Google DeepMindās Gemini. However, harvesting video from the platform is not allowed, according to Mohan and YouTubeās terms of service, without obtaining permission from the company.
After pointing out that Chuās earlier comments were on the record and asking for a clear statement on the matter, Viggleās spokesperson confirmed in their reply that the AI startup trains on YouTube videos:
Viggle leverages a variety of public sources, including YouTube, to generate AI content. Our training data has been carefully curated and refined, ensuring compliance with all terms of service throughout the process. We prioritize maintaining strong relationships with platforms like YouTube, and we are committed to respecting their terms by avoiding massive amounts of downloads and any other actions that would involve unauthorized video downloads.
We reached out to spokespeople for YouTube and Google, but have yet to hear back.
The startup joins others using YouTube as training data and thus operating in a gray area. Itās been reported that lots of AI model developers ā including Nvidia, Apple and Anthropic ā use YouTube video transcriptions or clips for training. Itās the dirty secret in Silicon Valley thatās not so secret: everybody is likely doing it. Whatās actually rare is saying it out loud.

