r/StableDiffusion
15 subscribers
29.5K photos
2.44K videos
1 file
13.9K links
Download Telegram
Preview with Flux Klein models in ComfyUI?

I tried to search for it, but haven't really found much info. Does anyone know if there's a way to make preview in ComfyUI work properly with Klein models? Using taesd method, the preview always lags a step behind, including showing the image from the previous generation after the first step, and the image it does show looks like it's not decoded properly, kind of noisy, and the colors are off. Like so:

https://preview.redd.it/rd28puh7y0sg1.png?width=1000&format=png&auto=webp&s=6ccd0141d7c0afcd2fe525afa146c9253f3de0f2

latent2rgb looks basically the same. Is there any way to get a normal preview?

https://redd.it/1s72dtm
@stablediffusion_r
Whats the verdict on Sage Attention 3 now? or stick with Sage 2.2?

I use Image Z Turbo, Wan 2.2 and LTX 2.3

I noticed that Sage Attention 3 altered the dress in a video of a dancing woman to a trousers when using LTX 2.3, I switched to Sage 2.2 and also tried disabling it and the issue was fixed

I actually thought it was the GGUF text encoder that causes the dress to turn into a pants but to my surprise it was Sage 3 that was causing it.

I went back to 2.2 only lost a few seconds speed by the quality was like if it' was disabled very good.

https://redd.it/1s73r4e
@stablediffusion_r
This media is not supported in your browser
VIEW IN TELEGRAM
I went from being a total dummy at ComfyUi to generating this I2V using LTX 2.3, I feel so proud of myself.

https://redd.it/1s76eod
@stablediffusion_r
What can you do if your hardware can generate 15,000 token/s?

[https://taalas.com/](https://taalas.com/)

Demo:

[https://chatjimmy.ai/](https://chatjimmy.ai/)

Saw this posted from r/Qwen_AI and r/LocalLLM today. I also remember seeing this from a few years ago when they first published their studies, but completely forgot about it.

Basically instead of inference on a graphics card where models are loaded onto memory, we burn the model into hardware. Remember CDs? It is cheap to build this compare to GPUs, they are using 6nm chips instead of the latest tech, no memories needed! The biggest downside is you can't swap models, there is no flexibility.

Thoughts? Would this making live streaming AI movies, games possible? You can have a MMO where every single npc have their own unique dialog with no delay for thousands of players.

What a crazy world we live in.

https://redd.it/1s77t1e
@stablediffusion_r
I see many people praising Klein, Zimage (turbo, base), and other models. But few examples. Please post here what you consider to represent the pinnacle of each model. Especially for photorealism.
https://redd.it/1s7ahcc
@stablediffusion_r
This media is not supported in your browser
VIEW IN TELEGRAM
For the many of you who claim to be getting very poor results/eyes/faces with LTX 2.3 ITV: do you have your distillation set too high? (First video, 0.6. Second video, 1.0)

https://redd.it/1s77hzb
@stablediffusion_r