| LTX-2 | |
|---|---|
| Developer | Lightricks |
| Initial release | October 23, 2025 |
| Operating system | Cross-platform |
| Type | Text-to-video model; audio-video generative model |
| Website | https://ltx.video/ |
| Part of a series on |
| Artificial intelligence (AI) |
|---|
LTX-2 is an open-source artificial intelligence video foundation model released by Lightricks in October 2025. It creates videos based on user prompts and was preceded by LTX Video, which was released in 2024 as the company's first text-to-video model.
In November 2024 Lightricks publicly released its first text-to-video model, LTX Video. It was a 2-billion parameter model, available as open source. [1] [2]
In May 2025 Lightricks launched LTXV-13b, a version with 13-billion parameters. [3] [4] Two months later, the model broke the 60 second barrier for generated video. [5] [6]
In October 2025 Lightricks announced its latest model, and renamed it LTX-2. The model was described as capable of generating synchronized audio and video at native 4K resolution and up to 50 frames per second (fps), using a variety of conditions and prompts, including text-to-video and image-to-video. [7] [8]
Google highlighted the fact that LTX-2 was trained on its infrastructure, [9] and saying it was "The first open source AI video generation model, powered by Google Cloud". [10]
Upon its release it was ranked in the top-3 models for image-to-video creation by Artificial Analysis, behind Kling 3.5 by Kling AI and Veo 3.1 by Google. Its text-to-image option was ranked 7th. [11]
LTX-2 builds upon the LTX Video architecture with several major improvements: [8] [12] [13]
Initial reception to LTX-2 was broadly positive, with several technology and media outlets highlighting its open-source approach and multimodal capabilities. [2] Open Source For You described LTX-2 as “one of the first AI video systems to combine 4K output, synchronized audio, and an open model release,” noting that it positioned Lightricks as a significant competitor to proprietary systems such as OpenAI's Sora and Google's Veo. [7]
IEA Green said that the model “could rewrite the AI filmmaking game,” emphasizing that its 50-fps rendering and unified audio-video generation made it suitable for professional studios and independent creators alike. [14]
AI News characterized LTX-2 as a “major step forward in the democratization of cinematic-quality video generation,” praising its consumer-grade hardware efficiency and multi-tier generation modes, while also noting ongoing challenges in long-form temporal stability. [13]
FinancialContent reported strong interest among creative agencies, attributing the attention to Lightricks’ decision to release model weights and datasets, which reviewers said enabled “a level of transparency not typically seen in commercial AI video models.” [15]
Some early reviewers also pointed out quality limitations. The Ray3 technical review noted occasional inconsistencies in lip-sync and motion tracking during long scenes, though it stated these were “in line with the challenges faced by all current AI video diffusion models” and expected to improve with continued iteration. [16]