[Release] Latent Model Organizer v1.0.0 - A free, open-source tool to automatically sort models by architecture and fetch CivitAI previews
https://redd.it/1ryr350
@stablediffusion_r
https://redd.it/1ryr350
@stablediffusion_r
Inpainting in 3 commands: remove objects or add accessories with any base model, no dedicated inpaint model needed
https://redd.it/1ryvv5p
@stablediffusion_r
https://redd.it/1ryvv5p
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: Inpainting in 3 commands: remove objects or add accessories with any base model,…
Explore this post and more from the StableDiffusion community
Ubisoft Chord PBR Material Estimation
I hadn't seen this mentioned anywhere, but Ubisoft has an open source model to make a PBR material from any image. It seems pretty amazing and already integrated into comfyui!
I found it by having this video come up on my youtube feed
https://www.youtube.com/watch?v=rE1M8_FaXtk
It seems pretty amazing: https://github.com/ubisoft/ubisoft-laforge-chord
https://github.com/ubisoft/ComfyUI-Chord?tab=readme-ov-file
https://redd.it/1ryvqpj
@stablediffusion_r
I hadn't seen this mentioned anywhere, but Ubisoft has an open source model to make a PBR material from any image. It seems pretty amazing and already integrated into comfyui!
I found it by having this video come up on my youtube feed
https://www.youtube.com/watch?v=rE1M8_FaXtk
It seems pretty amazing: https://github.com/ubisoft/ubisoft-laforge-chord
https://github.com/ubisoft/ComfyUI-Chord?tab=readme-ov-file
https://redd.it/1ryvqpj
@stablediffusion_r
YouTube
Can Ubisoft’s CHORD Model Replace Substance Designer? I Tested It
Learn how to generate AAA-quality PBR materials using Ubisoft’s CHORD model inside @comfyorg
In this tutorial, I walk through the full workflow — from setup to output — showing how CHORD can create stunning textures ready for Unreal Engine and other real…
In this tutorial, I walk through the full workflow — from setup to output — showing how CHORD can create stunning textures ready for Unreal Engine and other real…
I created a few helpful nodes for ComfyUI. I think "JLC Padded Image" is particularly useful for inpaint/outpaint workflows.
https://redd.it/1rylry3
@stablediffusion_r
https://redd.it/1rylry3
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: I created a few helpful nodes for ComfyUI. I think "JLC Padded Image" is particularly…
Explore this post and more from the StableDiffusion community
Nvidia SANA Video 2B
https://www.youtube.com/watch?list=TLGG-iNIhzqJ0OgyMDAzMjAyNg&v=7eNfDzA4yBs
Efficient-Large-Model/SANA-Video\_2B\_720p · Hugging Face
SANA-Video is a small, ultra-efficient diffusion model designed for rapid generation of high-quality, minute-long videos at resolutions up to 720×1280.
Key innovations and efficiency drivers include:
(1) Linear DiT: Leverages linear attention as the core operation, offering significantly more efficiency than vanilla attention when processing the massive number of tokens required for video generation.
(2) Constant-Memory KV Cache for Block Linear Attention: Implements a block-wise autoregressive approach that uses the cumulative properties of linear attention to maintain global context at a fixed memory cost, eliminating the traditional KV cache bottleneck and enabling efficient, minute-long video synthesis.
SANA-Video achieves exceptional efficiency and cost savings: its training cost is only 1% of MovieGen's (12 days on 64 H100 GPUs). Compared to modern state-of-the-art small diffusion models (e.g., Wan 2.1 and SkyReel-V2), SANA-Video maintains competitive performance while being 16× faster in measured latency. SANA-Video is deployable on RTX 5090 GPUs, accelerating the inference speed for a 5-second 720p video from 71s down to 29s (2.4× speedup), setting a new standard for low-cost, high-quality video generation.
More comparison samples here: SANA Video
https://redd.it/1rz153l
@stablediffusion_r
https://www.youtube.com/watch?list=TLGG-iNIhzqJ0OgyMDAzMjAyNg&v=7eNfDzA4yBs
Efficient-Large-Model/SANA-Video\_2B\_720p · Hugging Face
SANA-Video is a small, ultra-efficient diffusion model designed for rapid generation of high-quality, minute-long videos at resolutions up to 720×1280.
Key innovations and efficiency drivers include:
(1) Linear DiT: Leverages linear attention as the core operation, offering significantly more efficiency than vanilla attention when processing the massive number of tokens required for video generation.
(2) Constant-Memory KV Cache for Block Linear Attention: Implements a block-wise autoregressive approach that uses the cumulative properties of linear attention to maintain global context at a fixed memory cost, eliminating the traditional KV cache bottleneck and enabling efficient, minute-long video synthesis.
SANA-Video achieves exceptional efficiency and cost savings: its training cost is only 1% of MovieGen's (12 days on 64 H100 GPUs). Compared to modern state-of-the-art small diffusion models (e.g., Wan 2.1 and SkyReel-V2), SANA-Video maintains competitive performance while being 16× faster in measured latency. SANA-Video is deployable on RTX 5090 GPUs, accelerating the inference speed for a 5-second 720p video from 71s down to 29s (2.4× speedup), setting a new standard for low-cost, high-quality video generation.
More comparison samples here: SANA Video
https://redd.it/1rz153l
@stablediffusion_r
YouTube
SANA-Video Quality Comparison
This video shows how good can SANA-Video a pure linear transformer diffuison model can do.
PSA: Use the official LTX 2.3 workflow, not the ComfyUI included one. It's significantly better.
Most of the time I rely on the default ComfyUI workflows. They're producing results just as good as 90% of the overly-complicated workflows I see floating around online. So I was fighting with the default Comfy LTX 2.3 template for a while, just not getting anything good. Saw someone mention the official LTX workflows and figured I'd give it a try.
Yeah, huge difference. Easily makes LTX blow past WAN 2.2 into SOTA territory for me. So something's up with the Comfy default workflow.
If you're having issues with weird LTX 2 or LTX 2.3 generations, use the official workflow instead:
https://github.com/Lightricks/ComfyUI-LTXVideo/blob/master/example\_workflows/2.3/LTX-2.3\_T2V\_I2V\_Single\_Stage\_Distilled\_Full.json
This runs the distilled and non-distilled at the same time. I find they pretty evenly trade blows to give me what I'm looking for, so I just left it as generating both.
https://redd.it/1rz1u3j
@stablediffusion_r
Most of the time I rely on the default ComfyUI workflows. They're producing results just as good as 90% of the overly-complicated workflows I see floating around online. So I was fighting with the default Comfy LTX 2.3 template for a while, just not getting anything good. Saw someone mention the official LTX workflows and figured I'd give it a try.
Yeah, huge difference. Easily makes LTX blow past WAN 2.2 into SOTA territory for me. So something's up with the Comfy default workflow.
If you're having issues with weird LTX 2 or LTX 2.3 generations, use the official workflow instead:
https://github.com/Lightricks/ComfyUI-LTXVideo/blob/master/example\_workflows/2.3/LTX-2.3\_T2V\_I2V\_Single\_Stage\_Distilled\_Full.json
This runs the distilled and non-distilled at the same time. I find they pretty evenly trade blows to give me what I'm looking for, so I just left it as generating both.
https://redd.it/1rz1u3j
@stablediffusion_r
GitHub
ComfyUI-LTXVideo/example_workflows/2.3/LTX-2.3_T2V_I2V_Single_Stage_Distilled_Full.json at master · Lightricks/ComfyUI-LTXVideo
LTX-Video Support for ComfyUI. Contribute to Lightricks/ComfyUI-LTXVideo development by creating an account on GitHub.
This media is not supported in your browser
VIEW IN TELEGRAM
ComfyUI Nodes for Filmmaking (LTX 2.3 Shot Sequencing, Keyframing, First Frame/Last Frame)
https://redd.it/1rz355d
@stablediffusion_r
https://redd.it/1rz355d
@stablediffusion_r
Interesting. Images generated with low resolution + latent upscale. Qwen 2512.
https://redd.it/1rz6bz4
@stablediffusion_r
https://redd.it/1rz6bz4
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: Interesting. Images generated with low resolution + latent upscale. Qwen 2512.
Explore this post and more from the StableDiffusion community
Where do people train LoRA for ZIT?
Hey guys, I’ve been trying to figure out how people are training LoRA for ZIT but I honestly can’t find any clear info anywhere, I searched around Reddit, Civitai and other places but there’s barely anything detailed and most posts just mention it without explaining how to actually do it, I’m not sure what tools or workflow people are using for ZIT LoRA specifically or if it’s different from the usual setups, if anyone knows where to train it or has a guide/workflow that actually works I’d really appreciate it if you can share, thanks 🙏
https://redd.it/1rz8d5x
@stablediffusion_r
Hey guys, I’ve been trying to figure out how people are training LoRA for ZIT but I honestly can’t find any clear info anywhere, I searched around Reddit, Civitai and other places but there’s barely anything detailed and most posts just mention it without explaining how to actually do it, I’m not sure what tools or workflow people are using for ZIT LoRA specifically or if it’s different from the usual setups, if anyone knows where to train it or has a guide/workflow that actually works I’d really appreciate it if you can share, thanks 🙏
https://redd.it/1rz8d5x
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
SAMA 14b - Video Editing Model based off Wan 2.1 (Apache 2.0)
https://github.com/Cynthiazxy123/SAMA
https://huggingface.co/syxbb/SAMA-14B
https://redd.it/1rzauw4
@stablediffusion_r
https://github.com/Cynthiazxy123/SAMA
https://huggingface.co/syxbb/SAMA-14B
https://redd.it/1rzauw4
@stablediffusion_r
GitHub
GitHub - Cynthiazxy123/SAMA: Official inference code for SAMA: Factorized Semantic Anchoring and Motion Alignment for Instruction…
Official inference code for SAMA: Factorized Semantic Anchoring and Motion Alignment for Instruction-Guided Video Editing. - Cynthiazxy123/SAMA