Ubisoft Chord PBR Material Estimation
I hadn't seen this mentioned anywhere, but Ubisoft has an open source model to make a PBR material from any image. It seems pretty amazing and already integrated into comfyui!
I found it by having this video come up on my youtube feed
https://www.youtube.com/watch?v=rE1M8_FaXtk
It seems pretty amazing: https://github.com/ubisoft/ubisoft-laforge-chord
https://github.com/ubisoft/ComfyUI-Chord?tab=readme-ov-file
https://redd.it/1ryvqpj
@stablediffusion_r
I hadn't seen this mentioned anywhere, but Ubisoft has an open source model to make a PBR material from any image. It seems pretty amazing and already integrated into comfyui!
I found it by having this video come up on my youtube feed
https://www.youtube.com/watch?v=rE1M8_FaXtk
It seems pretty amazing: https://github.com/ubisoft/ubisoft-laforge-chord
https://github.com/ubisoft/ComfyUI-Chord?tab=readme-ov-file
https://redd.it/1ryvqpj
@stablediffusion_r
YouTube
Can Ubisoft’s CHORD Model Replace Substance Designer? I Tested It
Learn how to generate AAA-quality PBR materials using Ubisoft’s CHORD model inside @comfyorg
In this tutorial, I walk through the full workflow — from setup to output — showing how CHORD can create stunning textures ready for Unreal Engine and other real…
In this tutorial, I walk through the full workflow — from setup to output — showing how CHORD can create stunning textures ready for Unreal Engine and other real…
I created a few helpful nodes for ComfyUI. I think "JLC Padded Image" is particularly useful for inpaint/outpaint workflows.
https://redd.it/1rylry3
@stablediffusion_r
https://redd.it/1rylry3
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: I created a few helpful nodes for ComfyUI. I think "JLC Padded Image" is particularly…
Explore this post and more from the StableDiffusion community
Nvidia SANA Video 2B
https://www.youtube.com/watch?list=TLGG-iNIhzqJ0OgyMDAzMjAyNg&v=7eNfDzA4yBs
Efficient-Large-Model/SANA-Video\_2B\_720p · Hugging Face
SANA-Video is a small, ultra-efficient diffusion model designed for rapid generation of high-quality, minute-long videos at resolutions up to 720×1280.
Key innovations and efficiency drivers include:
(1) Linear DiT: Leverages linear attention as the core operation, offering significantly more efficiency than vanilla attention when processing the massive number of tokens required for video generation.
(2) Constant-Memory KV Cache for Block Linear Attention: Implements a block-wise autoregressive approach that uses the cumulative properties of linear attention to maintain global context at a fixed memory cost, eliminating the traditional KV cache bottleneck and enabling efficient, minute-long video synthesis.
SANA-Video achieves exceptional efficiency and cost savings: its training cost is only 1% of MovieGen's (12 days on 64 H100 GPUs). Compared to modern state-of-the-art small diffusion models (e.g., Wan 2.1 and SkyReel-V2), SANA-Video maintains competitive performance while being 16× faster in measured latency. SANA-Video is deployable on RTX 5090 GPUs, accelerating the inference speed for a 5-second 720p video from 71s down to 29s (2.4× speedup), setting a new standard for low-cost, high-quality video generation.
More comparison samples here: SANA Video
https://redd.it/1rz153l
@stablediffusion_r
https://www.youtube.com/watch?list=TLGG-iNIhzqJ0OgyMDAzMjAyNg&v=7eNfDzA4yBs
Efficient-Large-Model/SANA-Video\_2B\_720p · Hugging Face
SANA-Video is a small, ultra-efficient diffusion model designed for rapid generation of high-quality, minute-long videos at resolutions up to 720×1280.
Key innovations and efficiency drivers include:
(1) Linear DiT: Leverages linear attention as the core operation, offering significantly more efficiency than vanilla attention when processing the massive number of tokens required for video generation.
(2) Constant-Memory KV Cache for Block Linear Attention: Implements a block-wise autoregressive approach that uses the cumulative properties of linear attention to maintain global context at a fixed memory cost, eliminating the traditional KV cache bottleneck and enabling efficient, minute-long video synthesis.
SANA-Video achieves exceptional efficiency and cost savings: its training cost is only 1% of MovieGen's (12 days on 64 H100 GPUs). Compared to modern state-of-the-art small diffusion models (e.g., Wan 2.1 and SkyReel-V2), SANA-Video maintains competitive performance while being 16× faster in measured latency. SANA-Video is deployable on RTX 5090 GPUs, accelerating the inference speed for a 5-second 720p video from 71s down to 29s (2.4× speedup), setting a new standard for low-cost, high-quality video generation.
More comparison samples here: SANA Video
https://redd.it/1rz153l
@stablediffusion_r
YouTube
SANA-Video Quality Comparison
This video shows how good can SANA-Video a pure linear transformer diffuison model can do.
PSA: Use the official LTX 2.3 workflow, not the ComfyUI included one. It's significantly better.
Most of the time I rely on the default ComfyUI workflows. They're producing results just as good as 90% of the overly-complicated workflows I see floating around online. So I was fighting with the default Comfy LTX 2.3 template for a while, just not getting anything good. Saw someone mention the official LTX workflows and figured I'd give it a try.
Yeah, huge difference. Easily makes LTX blow past WAN 2.2 into SOTA territory for me. So something's up with the Comfy default workflow.
If you're having issues with weird LTX 2 or LTX 2.3 generations, use the official workflow instead:
https://github.com/Lightricks/ComfyUI-LTXVideo/blob/master/example\_workflows/2.3/LTX-2.3\_T2V\_I2V\_Single\_Stage\_Distilled\_Full.json
This runs the distilled and non-distilled at the same time. I find they pretty evenly trade blows to give me what I'm looking for, so I just left it as generating both.
https://redd.it/1rz1u3j
@stablediffusion_r
Most of the time I rely on the default ComfyUI workflows. They're producing results just as good as 90% of the overly-complicated workflows I see floating around online. So I was fighting with the default Comfy LTX 2.3 template for a while, just not getting anything good. Saw someone mention the official LTX workflows and figured I'd give it a try.
Yeah, huge difference. Easily makes LTX blow past WAN 2.2 into SOTA territory for me. So something's up with the Comfy default workflow.
If you're having issues with weird LTX 2 or LTX 2.3 generations, use the official workflow instead:
https://github.com/Lightricks/ComfyUI-LTXVideo/blob/master/example\_workflows/2.3/LTX-2.3\_T2V\_I2V\_Single\_Stage\_Distilled\_Full.json
This runs the distilled and non-distilled at the same time. I find they pretty evenly trade blows to give me what I'm looking for, so I just left it as generating both.
https://redd.it/1rz1u3j
@stablediffusion_r
GitHub
ComfyUI-LTXVideo/example_workflows/2.3/LTX-2.3_T2V_I2V_Single_Stage_Distilled_Full.json at master · Lightricks/ComfyUI-LTXVideo
LTX-Video Support for ComfyUI. Contribute to Lightricks/ComfyUI-LTXVideo development by creating an account on GitHub.
This media is not supported in your browser
VIEW IN TELEGRAM
ComfyUI Nodes for Filmmaking (LTX 2.3 Shot Sequencing, Keyframing, First Frame/Last Frame)
https://redd.it/1rz355d
@stablediffusion_r
https://redd.it/1rz355d
@stablediffusion_r
Interesting. Images generated with low resolution + latent upscale. Qwen 2512.
https://redd.it/1rz6bz4
@stablediffusion_r
https://redd.it/1rz6bz4
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: Interesting. Images generated with low resolution + latent upscale. Qwen 2512.
Explore this post and more from the StableDiffusion community
Where do people train LoRA for ZIT?
Hey guys, I’ve been trying to figure out how people are training LoRA for ZIT but I honestly can’t find any clear info anywhere, I searched around Reddit, Civitai and other places but there’s barely anything detailed and most posts just mention it without explaining how to actually do it, I’m not sure what tools or workflow people are using for ZIT LoRA specifically or if it’s different from the usual setups, if anyone knows where to train it or has a guide/workflow that actually works I’d really appreciate it if you can share, thanks 🙏
https://redd.it/1rz8d5x
@stablediffusion_r
Hey guys, I’ve been trying to figure out how people are training LoRA for ZIT but I honestly can’t find any clear info anywhere, I searched around Reddit, Civitai and other places but there’s barely anything detailed and most posts just mention it without explaining how to actually do it, I’m not sure what tools or workflow people are using for ZIT LoRA specifically or if it’s different from the usual setups, if anyone knows where to train it or has a guide/workflow that actually works I’d really appreciate it if you can share, thanks 🙏
https://redd.it/1rz8d5x
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
SAMA 14b - Video Editing Model based off Wan 2.1 (Apache 2.0)
https://github.com/Cynthiazxy123/SAMA
https://huggingface.co/syxbb/SAMA-14B
https://redd.it/1rzauw4
@stablediffusion_r
https://github.com/Cynthiazxy123/SAMA
https://huggingface.co/syxbb/SAMA-14B
https://redd.it/1rzauw4
@stablediffusion_r
GitHub
GitHub - Cynthiazxy123/SAMA: Official inference code for SAMA: Factorized Semantic Anchoring and Motion Alignment for Instruction…
Official inference code for SAMA: Factorized Semantic Anchoring and Motion Alignment for Instruction-Guided Video Editing. - Cynthiazxy123/SAMA
Have you tried fish audio S2Pro?
What is your experience with it? Do you think it can compete with Elevenlabs?
I have tried it and it is 80% as good as Elevenlabs.
https://redd.it/1rz7wjh
@stablediffusion_r
What is your experience with it? Do you think it can compete with Elevenlabs?
I have tried it and it is 80% as good as Elevenlabs.
https://redd.it/1rz7wjh
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
What's the best pipeline to uniformize and upscale a large collection of old book cover scans?
https://redd.it/1rzbpeg
@stablediffusion_r
https://redd.it/1rzbpeg
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit: What's the best pipeline to uniformize and upscale a large collection of old book…
Explore this post and more from the StableDiffusion community
GPU Temps for Local Gen
What sort of temps are acceptable for local image generation? I generate images at 832x1216 and upscale by 1.5x and i'm seeing hot spot temps on my RTX 4080 peak out at 103c
is it time for me to replace the thermal paste on my GPU or is this expected temps? Worried that these temps will cause damage and be a costly replacement.
https://redd.it/1rz9je1
@stablediffusion_r
What sort of temps are acceptable for local image generation? I generate images at 832x1216 and upscale by 1.5x and i'm seeing hot spot temps on my RTX 4080 peak out at 103c
is it time for me to replace the thermal paste on my GPU or is this expected temps? Worried that these temps will cause damage and be a costly replacement.
https://redd.it/1rz9je1
@stablediffusion_r
Reddit
From the StableDiffusion community on Reddit
Explore this post and more from the StableDiffusion community
This media is not supported in your browser
VIEW IN TELEGRAM
How is this done? Are we going to live in a world of catfishing?
https://redd.it/1rzicfw
@stablediffusion_r
https://redd.it/1rzicfw
@stablediffusion_r