u/Iory1998

▲ 2 r/StableDiffusion+1 crossposts

Which Version of LTX2.3 are You Using?

Hi,

I'd like to use LTX2.3, But I am not sure which models do I use. I'd prefer to use a base LTX2 model + LTX2.3 LoRA as that gives me more flexibility to control LoRA strength, but I am not sure if that's possible.

What are your recommendations? Any tips? Could you please provide the links to the models you are actually using?

Thanks.

reddit.com
u/Iory1998 — 9 hours ago
🔥 Hot ▲ 179 r/LocalLLaMA

My biggest Issue with the Gemma-4 Models is the Massive KV Cache!!

I mean, I have 40GB of Vram and I still cannot fit the entire Unsloth Gemma-4-31B-it-UD-Q8 (35GB) even at 2K context size unless I quantize KV to Q4 with 2K context size? WTF? For comparison, I can fit the entire UD-Q8 Qwen3.5-27B at full context without KV quantization!

If I have to run a Q4 Gemma-4-31B-it-UD with a Q8 KV cache, then I am better off just using Qwen3.5-27B. After all, the latter beats the former in basically all benchmarks.

What's your experience with the Gemma-4 models so far?

reddit.com
u/Iory1998 — 10 hours ago