LTX 2.3 adds distilled LoRA v1.1 for better motion-audio sync
Stable Diffusion and VFX creators say LTX 2.3's distilled LoRA v1.1 improves motion and custom-audio sync. Posts show local short-film and flight-shot workflows running through ComfyUI and Resolve on consumer GPUs.

TL;DR
- According to Tokyo_Jab's Reddit post, the updated distilled LoRA v1.1 for LTX 2.3 improves motion and custom-audio sync, with the underlying weights linked on Hugging Face.
- In hiskuriosity's VFX demo, a handheld phone shot transitions into an AI flight sequence built with DaVinci Resolve and LTX Video 2.3, showing the model in a compositing workflow rather than a text-to-video one.
- Previous_Night_7154's Suno crossover post maps a fully local pipeline: key images from Nano Banana2, video rendering in ComfyUI with LTX 2.3, and a consumer 4070 Ti Super finishing the piece in about two days.
- manmaynakhashi's audio experiment suggests the LTX 2.3 stack is already being pushed beyond video, with prompt examples for villain, noir detective, talk-show host, and action-hero voice performances.
You can browse the official LTX-2.3 page, watch a mobile-shot flight transition, and check a music-video workflow that stayed local. The interesting part is how quickly creators moved from model-drop chatter to specific pipelines: custom-audio sync in Tokyo_Jab's example, Resolve compositing in hiskuriosity's post, and ComfyUI first-frame-last-frame runs in Previous_Night_7154's breakdown.
Distilled LoRA v1.1
The clearest claim in the evidence pool is simple: the new distilled LoRA v1.1 looks better in motion. In Tokyo_Jab's post, the creator says it "vastly improve[s] the output," specifically calling out better motion and sync when pairing custom audio with an input image.
TWEEDLES - Example 2
2 comments
That matters because it points to a narrower upgrade than a whole new model family. The public anchor is still the same Lightricks LTX-2.3 model page, but the creator reaction here is focused on a distilled add-on that changes output quality inside an existing workflow.
Resolve and ComfyUI pipelines
The creator examples are more useful than generic launch framing because they show where LTX 2.3 is actually landing. In hiskuriosity's VFX demo, a real handheld mobile shot is composited in DaVinci Resolve, then extended into a fully AI-generated flight sequence with LTX Video 2.3.
My 5-year-old told me he could fly. Here's what it took to prove him right. 🚀
0 comments
A second post turns that into a reproducible indie stack. Previous_Night_7154's workflow post says the key images came from Nano Banana2 on Flow, while the video segments were rendered locally in ComfyUI using LTX 2.3 image-to-video and first-frame-last-frame on a 4070 Ti Super with 16 GB VRAM.
[Happy hardcore ] "Drifting past the sun" by Jo Spamiti the second
0 comments
That same post adds two concrete production details: the piece took about two days from first character-sheet image to YouTube upload, and the soundtrack came from Suno after many iterations on style and lyrics. For AI filmmakers, that is the interesting combo, short-form narrative video, music generation, and local rendering on a prosumer card.
Audio prompt tests
The oddest spillover is audio. manmaynakhashi's experiment shares LTX-2.3-based audio outputs driven by long performance prompts, including a theatrical villain laugh, a gravelly noir detective, a talk-show host breaking into laughter, and a panting action hero.
LTX-2.3 based audio model outputs
1 comments
The included Reddit reply adds the next obvious question. In the comment thread under that post, one commenter asks for workflow details, maximum clip length, and whether the setup can hold a consistent voice across generations, which is a sharper signal of what creators still want from this stack than any benchmark table would be.