GGUF

ltx2.3-gguf

screenshot

see picture above for setup (inside sub-graph)

  • pull ltx2.3-22b-dev-iq4_xs.gguf to diffusion_models
  • pull ltx2.3_text_encoder_projection-q4_0.gguf and gemma3-12b-it to text_encoders

btw, replace the ltx2.3-22b-dev or -distill with this universal checkpoint clip (see below)

  • pull ltx2.3-22b-checkpoint_fp8_e4m3fn.safetensors to checkpoints

note: you don't need any extra node for running ltx2.3 under this setting; probably run it with beginner level gpu without problem, just wait slightly longer; you might need protobuf for rebuilding the tokenizer if you opt gemma3 gguf, simply install it with:

.\python_embeded\python.exe -s -m pip install protobuf

Downloads last month
892
GGUF
Model size
1B params
Architecture
pig
Hardware compatibility
Log In to add your hardware

4-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for gguf-org/ltx2.3-gguf

Quantized
(15)
this model