back to top
HomeTechAI ModelsLTX 2.3 Is Here: The AI Video Generator That Runs on Your...

LTX 2.3 Is Here: The AI Video Generator That Runs on Your PC and Challenges Veo 3.1

- Advertisement -

Two years ago, if you wanted to generate a decent AI video, the only real option was a subscription. Pick a tool, pay monthly, generate on their servers. That was just how AI video worked.

Open source models eventually closed the gap on quality, but running them locally meant terminals, dependency errors, and a lot of patience. Not everyone wanted that headache. Most people didn’t.

But now that just changed. On March 5th, Lightricks dropped two things at once. LTX 2.3, a major upgrade to their open source video model, and LTX Desktop, a proper video editor built entirely on top of it. Its Open Source & you can install it like any other app on your computer.

If you have the hardware, we’re talking 32GB VRAM for the full experience, you genuinely don’t need a subscription for video generation anymore. Just your GPU doing the work.

And if you’re not there on hardware yet, LTX still offers an API. Paid, but flexible. The point is you have options now.

So What Did Lightricks Actually Build?

Lightricks has been building LTX for a while now. LTX 2 was already turning heads in the open source community, and 2.3 is the most complete version they’ve shipped yet. It’s a diffusion-based video model that generates both video and audio together in a single pass. The sound, the motion & the visuals all come from the same model at the same time. That alone separates it from most of what’s out there.

The base model is 22B parameters and uses Google’s Gemma 3 12B as its text encoder, which is the part that reads your prompt and figures out what to generate. It’s a capable brain for a capable model.

Here’s what actually changed in 2.3.

Sharper output: Lightricks rebuilt the VAE, the part responsible for encoding and decoding visuals. Previous versions softened fine details like hair and edges, especially at lower resolutions. Cleaner textures now, less fixing in post.

Better prompt following: Complex prompts with multiple subjects or specific spatial relationships used to drift from what you asked for. If you’ve been dumbing down your prompts to get consistent results, you can stop.

Image to video that actually moves: Previous versions would often produce a slow pan or freeze entirely when animating a still image. They reworked the training specifically to fix this.

Native portrait video: Vertical resolution up to 1080×1920, trained on actual vertical data. First time in LTX. Relevant if you’re making content for TikTok, Reels, or Shorts.

Cleaner audio: New vocoder, cleaner training data. The audio actually aligns with what’s on screen instead of feeling tacked on.

LTX-Desktop: AI Video Without the Setup

ltx desktop AI video editor

This is where it gets better. LTX-Desktop is a proper video editor built on LTX 2.3 and you install it like any other app.

If you have at least 24GB VRAM, everything runs fully local. Your GPU does the work, simply no cost per generation or if you don’t have the hardware? LTX Desktop connects to their API instead but generation just happens on their servers & its paid.

Either way you’re working inside one app. And if one small detail looks off, the Retake tool lets you fix just that area without regenerating the whole clip. On a paid cloud tool you’d re-roll the entire thing and pay for it. Here you just fix it.

LTX 2.3 vs Veo 3.1

Veo 3.1 is probably the most talked about AI video model right now, so that’s the one worth comparing directly.

MetricLTX 2.3 (Local)LTX 2.3 (API)Veo 3.1 (API)
PriceFree$0.04–$0.24/sec$0.15–$0.60/sec
4K pricingFree (your GPU)$0.16–$0.24/sec$0.35–$0.60/sec
AudioNative, includedNative, includedNative, included
Privacy100% localCloudCloud
Retake/EditYes, in Desktop appYesNo
Open SourceYes YesNo
Runs offlineYesNoNo

Pricing is based on official pages as of March 2026 and may change.

Before You Start Using LTX 2.3

Lightricks recommends a Windows machine with a CUDA GPU and at least 32GB VRAM to run this locally. That’s the honest requirement. Not everyone has that sitting on their desk right now, and that’s fine.

If you’re not there on hardware yet, the API route through LTX Desktop is still a solid option. You get the same editor, the same Retake feature, just without the local generation. When you go through the setup you’ll see the pricing plans and you can pick what works for your usage.

The local route is where the real freedom is. But the API keeps it accessible while you get there.

So, Is LTX 2.3 Worth It?

If you have the hardware, honestly yes. A free, open source video model with native audio, portrait support, a proper desktop app, and no subscription attached to it — that’s not a small thing. Six months ago this combination didn’t exist in open source.

If you don’t have the hardware yet, it’s still worth keeping an eye on. The model is only going to get faster, the community is already building on it, and the API option means you can start using it today without waiting on a GPU upgrade.

Paid tools aren’t going away tomorrow. But they’re going to have a harder time justifying their price tags every time something like this drops.

LTX 2.3 is one of those releases that quietly shifts what people expect for free.

YOU MAY ALSO LIKE
OpenMythos

OpenMythos: The Closest Thing to Claude Mythos You Can Run (And It’s Open Source)

0
Anthropic hasn't told anyone how Claude Mythos works. No architecture paper or model card with details. Just a product that keeps surprising people and a company that stays quiet about why. That silence has been driving the research community a little crazy. So one developer Kye Gomez did something about it. He read every public paper he could find on recurrent transformers, looped architectures, and inference-time scaling. He studied the behavioral patterns people were reporting from Mythos. Then he built what he thinks is inside it, published the code under MIT, and made it pip installable. It's called OpenMythos. It is not Claude Mythos. Gomez is explicit about that but the hypothesis behind it is serious, the architecture is real, and the reasoning for why Mythos might work this way is harder to dismiss than you'd expect.
Nucleus-Image AI image MOE model

Nucleus-Image: 17B Open-Source MoE Image Model Delivering GPT-Image Level Performance

0
The mixture-of-experts trick changed how people think about LLMs. Instead of running every parameter on every token, you activate a small fraction of the network per forward pass and somehow the quality stays competitive while the compute drops. It's the reason models like Mixtral punched above their weight. Everyone in the LLM space understood it immediately. Nobody had done it openly for image generation. Until now. Nucleus-Image is a 17B parameter diffusion transformer that activates roughly 2B parameters per forward pass. It beats Imagen4 on OneIG-Bench, sits at number one on DPG-Bench overall, and matches Qwen-Image on GenEval. It's also a base model. No fine-tuning, reinforcement learning or human preference tuning. What you're seeing in those benchmarks is raw pre-training performance. That's either impressive or a caveat depending on what you need it for, probably both.
ERNIE-Image Open-Source 8B Text-to-Image Model for Posters Comics and control

ERNIE-Image: Open-Source 8B Text-to-Image Model for Posters, Comics & Structured Generation

0
Text rendering in open source AI image generation has been broken for a long time. Ask most models to put readable words on a poster, lay out a comic panel, or generate anything where the text actually has to make sense and only few models can do it accurately and from rest you get something that looks like it was written by someone who learned the alphabet from a fever dream. ERNIE-Image is Baidu's answer to that specific problem. It's an 8B open weight text-to-image model built on a Diffusion Transformer and it's genuinely good at dense text, structured layouts, posters, infographics and multi-panel compositions. It can run on a 24GB consumer GPU, it's on Hugging Face right now, and it comes in two versions, a full quality model and a turbo variant that gets there in 8 steps instead of 50.

Don’t miss any Tech Story

Subscribe To Firethering NewsLetter

You Can Unsubscribe Anytime! Read more in our privacy policy