Transcript
Logan (0:00)
Foreign.
Podcast Host (0:05)
Welcome to reshaping workflows with Dell Pro Precision and Nvidia, where innovation meets real world impact in high performance computing.
Logan (0:19)
This is Logan live from GTC 2026. This is day three and we're here at LTX at Light Tricks. So I'm with Yaron, also known if you've ever been to Starbucks with him. Known as Dan apparently, is what he said, not me. So first you're on. Really appreciate the time. Tell us, you know, all the listeners out there, kind of what your position is, what you do and then we'll get right into it.
Jeroen (0:40)
Okay. So hi, I'm Jeroen, co founder and CTO at Lytrix. And you know we're a company that exists for 13 years now, did a lot of apps on mobile like Facetune. In the last three years we completely transitioned to become an AI. First company started developing our own video model called LTX. The latest version is LTX2 that was released about two and a half months ago. Amazing video model, ranked number one in the world doing video and audio. So multimodal. And it's also open source which is great because everybody can just download and run it. We already have more than 5 million downloads on Hugging Face. And besides the foundational model itself, we also offer API for inference and also a product layer on top which is a one stop shop for video and image creation and storytelling called LTX Studio.
Logan (1:37)
So as you know. I did not know. I mean I work with Noah a little bit and I know LTX has been around for a little while. Right. And not a knock to your previous product, but I'd use the original version about a year ago and I mean it was good but nothing was great. Right. And I've installed the newer version and it's great. And I'm not a creative by any stretch of the imagination, but it definitely LTX rivals anything. I mean I'm not comparing your competition but like you know, Runway, like all the stuff you kind of expect, you guys are kind of blowing them out of the water. So I guess what really, and you don't have to share any proprietary secrets or anything like that, but really kind of gen over gen from like LTX1 to say LTX2. What has really changed? Like what have you done to really. I mean, because it went from good to freaking fantastic. So what really changed in that to make that jump?
Jeroen (2:20)
So basically everything we scaled everything, whether it's the number of parameters, the data and of course we also added the audio support which is amazing because who nowadays creates a video which is without an audio. And surprisingly so the audio also really affects the quality of the video because the way that we generate video is not just the video, the audio itself. And the audio generation also affects the video generation and vice versa. And so we got a ton of great feedback on features that use the model in maybe surprising ways, like audio to video, where you can maybe you have your own audio track and you want to generate a video for that, like music video for example. And then you can generate characters that sing based on that original soundtrack. And the community says that this is the best model out there for these kind of performances because the audio really drives the expression of the characters and so on. The reason going back to maybe the reason of why this is all open source. So we have various reasons. One is democratization of creativity. We think that right now, as more and more companies go closed source, it's really, really important for us and for the world, I think that to have a model that is completely open, we open source not just the weights, but also the inference code and the training code which allows you to fine tune the model. So if you have your own data, you can just use that to make a better model or a model that's customized for yourself. If you're a studio, and we have many clients that are studios that want to protect their ip, they don't want to send it over the Internet, so they want to do it on Prem. And the ability to run locally on Prem is also something that's very, very important because it means that you can run it in your own network, whether it's on data center gpu, which is worked amazing because our model is extremely efficient and very compressed. But also you can run it on consumer GPUs like we see behind us, we have a dell workstation here, T2 with RTX 6000 and it works amazingly well, which is crazy. So basically all the stack and you can also mix and match the hardware, you can run in low res, generating low resolution on consumer GPUs and then send it to the cloud for upscaling.
