![Latent Vision](/img/default-banner.jpg)
- 28
- 825 518
Latent Vision
Italy
Приєднався 30 вер 2023
Dissecting SD3
How does SD3 work? Is it any good? No drama, no politics, only the technical side of things.
The SD3 Negative node is part of the Comfy Essentials: github.com/cubiq/ComfyUI_essentials
Free SD3 generations at OpenArt: openart.ai/create?ai_model=stable-diffusion-3-sd3
Discord server: discord.com/invite/W2DhHkcjgn
Github sponsorship: github.com/sponsors/cubiq
Support with paypal: www.paypal.me/matt3o
Twitter: cubiq
00:00 Intro
00:35 Default workflow
05:46 Testing the negatives
08:44 Lying in the grass
12:40 Prompt adhesion
15:19 Noise
16:30 High resolutions
17:28 Control Nets
18:41 License
The SD3 Negative node is part of the Comfy Essentials: github.com/cubiq/ComfyUI_essentials
Free SD3 generations at OpenArt: openart.ai/create?ai_model=stable-diffusion-3-sd3
Discord server: discord.com/invite/W2DhHkcjgn
Github sponsorship: github.com/sponsors/cubiq
Support with paypal: www.paypal.me/matt3o
Twitter: cubiq
00:00 Intro
00:35 Default workflow
05:46 Testing the negatives
08:44 Lying in the grass
12:40 Prompt adhesion
15:19 Noise
16:30 High resolutions
17:28 Control Nets
18:41 License
Переглядів: 12 339
Відео
Higher quality images by prompting individual UNet blocks
Переглядів 13 тис.14 днів тому
This time we are going to do some R&D and I will need your help to reverse engineer the UNet. Basically prompting each block of the UNet separately with a dedicated prompt we are able to get higher quality generations. Extension repository: github.com/cubiq/prompt_injection Discord server: discord.com/invite/W2DhHkcjgn Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypa...
About AI, Art, Ethics and the environment
Переглядів 7 тис.21 день тому
Not a tutorial but this is something I wanted to talk about since a while. Ethics of using AI, the environmental costs and Is AI art? The subtitles are hand edited and corrected. What do you think? Discord server: discord.com/invite/W2DhHkcjgn Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypal.me/matt3o Twitter: cubiq TED Talk by Sasha Luccioni: www.ted.co...
How to use Face Analysis to improve your workflows
Переглядів 11 тис.28 днів тому
I often use Face Analysis in my workflows but we never actually talked about how it actually works. Here all you need to know. Remember to upgrade the extensions, these are all new features! Check my Discord for the workflows, they are all free for everybody to use. Discord server: discord.com/invite/W2DhHkcjgn Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypal.me/mat...
How to use PuLID in ComfyUI
Переглядів 24 тис.Місяць тому
In this video I'm going through some basic PuLID usage and also comparing it to other face models. If you have it already installed remember to upgrade the extension! PuLID ComfyUI extention: github.com/cubiq/PuLID_ComfyUI Face Analysis node: github.com/cubiq/ComfyUI_FaceAnalysis Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypal.me/matt3o Twitter: cubiq M...
Animation with weight scheduling and IPAdapter
Переглядів 27 тис.Місяць тому
About time we talked about animations again! I just released new nodes IPAdapter and the Essential that make scheduling IPAdapter, Prompt and controlnet very easy and efficient. Workflows: f.latent.vision/download/scheduled_weights.zip Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypal.me/matt3o Twitter: cubiq My Discord server: discord.com/invite/W2DhHkcj...
All new Attention Masking nodes
Переглядів 21 тис.2 місяці тому
I just pushed an update to simplify attention masking and regional prompting with IPAdapter. Be sure to upgrade the IPAdapter and the ComfyUI Essential to get access to all the new features. The Essentials can be found here: github.com/cubiq/ComfyUI_essentials Download the workflow: f.latent.vision/download/new_attention_masking.zip Github sponsorship: github.com/sponsors/cubiq Support with pay...
Become a Style Transfer Master with ComfyUI and IPAdapter
Переглядів 23 тис.2 місяці тому
This time we are going to: - Play with coloring books - Turn a tiger into ice - Apply a different style to an existing image Github sponsorship: github.com/sponsors/cubiq Support with paypal: www.paypal.me/matt3o Discord server: discord.com/invite/W2DhHkcjgn All the workflows can be downloaded here no strings attached: f.latent.vision/download/style_transfer.zip The SDXL lineart controlnet: hug...
Style and Composition with IPAdapter and ComfyUI
Переглядів 26 тис.2 місяці тому
IPAdapter Extension: github.com/cubiq/ComfyUI_IPAdapter_plus Github sponsorship: github.com/sponsors/cubiq Paypal: www.paypal.me/matt3o Discord server: discord.com/invite/W2DhHkcjgn 00:00 Intro 00:26 Style Transfer 03:05 Composition Transfer 04:56 Style and Composition 07:42 Improve the composition 08:40 Outro
IPAdapter v2: all the new features!
Переглядів 66 тис.2 місяці тому
I updated the IPAdapter extension for ComfyUI. It's a complete code rewrite so unfortunately the old workflows are not compatible anymore and need to be rebuilt. Sorry about that but I don't have time to maintain old code. IPAdapter Extension: github.com/cubiq/ComfyUI_IPAdapter_plus Sponsor the development of my extensions: www.paypal.me/matt3o Discord server: discord.com/invite/W2DhHkcjgn 00:0...
Build Your Own ComfyUI APP!
Переглядів 18 тис.3 місяці тому
This time we are getting our hands dirty into code! I wanted to show you how easy it is to build custom web applications with ComfyUI and absolutely no knowledge of python. Let me know if you'd like more of this kind of content! Comfy Dungeon: github.com/cubiq/Comfy_Dungeon The FastGen extension can be downloaded from here: f.latent.vision/download/fastgen.zip Discord server: discord.com/invite...
Variations with noise injection KSampler (in pills)
Переглядів 8 тис.3 місяці тому
This is a kind of experiment I'm doing... I try to pack my videos with a lot of information and sometimes it might feel overwhelming. I was thinking maybe you'd also appreciate shorter videos dedicated to just one concept or even just one node. Please let me know what you think and if you'd like more of this kind of videos. Discord server: discord.com/invite/W2DhHkcjgn Complete video about imag...
InstantID: Everything you need to know
Переглядів 52 тис.3 місяці тому
InstantID is a style transfer tool targeted to portraits. It's incredibly easy to create a composition in specific style. In this video I'm showing you how to improve the likeliness, how to make a scene with multiple people and much more! InstantID Extension: github.com/cubiq/ComfyUI_InstantID Face Analysis Extension: github.com/cubiq/ComfyUI_FaceAnalysis Generic Workflows: github.com/cubiq/Com...
ComfyUI: Advanced understanding Part 2
Переглядів 28 тис.4 місяці тому
This is Part 2 of my basics series. Last time we learned how to set the conditioning to the whole scene, time to see how to make localized changes. I'm also talking about LCM, Math Nodes and other big and small tricks! As always do let me know what you think and if I should keep releasing "basics" tutorials or you prefer more advanced stuff. Discord server: discord.com/invite/W2DhHkcjgn 00:00 I...
Making Trading Cards with ComfyUI
Переглядів 11 тис.4 місяці тому
This time we are playing some card games! This tutorial will show you how to generate very different elements and mix them together into a nice composition. Workflows: f.latent.vision/download/trading_card.zip Discord server: discord.com/invite/W2DhHkcjgn 00:00 Intro 00:53 The Frame 06:22 The Main Character 07:45 The Text Area 11:07 The Symbol 14:36 The Final Composition 16:40 Writing Text 18:4...
Throwing data to your face (models)!
Переглядів 17 тис.4 місяці тому
Throwing data to your face (models)!
ComfyUI: Advanced Understanding (Part 1)
Переглядів 66 тис.5 місяців тому
ComfyUI: Advanced Understanding (Part 1)
FaceID Take 2! Even more face models! (IPAdapter+ComfyUI)
Переглядів 36 тис.5 місяців тому
FaceID Take 2! Even more face models! (IPAdapter ComfyUI)
Jellyfish Ballerina Animation with AnimateDiff
Переглядів 15 тис.6 місяців тому
Jellyfish Ballerina Animation with AnimateDiff
Image stability and repeatability (ComfyUI + IPAdapter)
Переглядів 56 тис.6 місяців тому
Image stability and repeatability (ComfyUI IPAdapter)
Animations with IPAdapter and ComfyUI
Переглядів 33 тис.6 місяців тому
Animations with IPAdapter and ComfyUI
Attention Masking with IPAdapter and ComfyUI
Переглядів 41 тис.7 місяців тому
Attention Masking with IPAdapter and ComfyUI
From real to anime (with IPAdapter and ComfyUI)
Переглядів 15 тис.7 місяців тому
From real to anime (with IPAdapter and ComfyUI)
ComfyUI IPAdapter Advanced Features
Переглядів 31 тис.8 місяців тому
ComfyUI IPAdapter Advanced Features
How to use IPAdapter models in ComfyUI
Переглядів 92 тис.8 місяців тому
How to use IPAdapter models in ComfyUI
I've almost binged your entire playlist in the last few days. Have learnt a lot and I thank you for sharing your knowledge. Had a query and the comparison b/w the available tools shown near the end of this video probably makes this the best place to ask: Q) In terms of character consistency and likeness, how does PulID/ FaceID/ InstantID compare against results obtained through a well-baked character LoRA model?
Bro check please FasterViT
This was the video I was looking for!
"This is not magic." But it sure helluva feels like it, boss!
Hey your discord link is outdated
don't know why that sometimes happens, but it's actually valid
@@latentvision I still can't use it :( had this happen also with another Discord also for SD related content...
We got to see Peggy Carter's beautiful skull as a bonus.
thanks!!
Error occurred when executing IPAdapterUnifiedLoader: ClipVision model not found. File "/teamspace/studios/this_studio/mohit/ComfyUI/execution.py", line 151, in recursive_execute output_data, output_ui = get_output_data(obj, input_data_all) File "/teamspace/studios/this_studio/mohit/ComfyUI/execution.py", line 81, in get_output_data return_values = map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True) File "/teamspace/studios/this_studio/mohit/ComfyUI/execution.py", line 74, in map_node_over_list results.append(getattr(obj, func)(**slice_dict(input_data_all, i))) File "/teamspace/studios/this_studio/mohit/ComfyUI/custom_nodes/ComfyUI_IPAdapter_plus/IPAdapterPlus.py", line 506, in load_models raise Exception("ClipVision model not found.") I am facing this error
In general, watching similar YT tutorials by other people feels like mugging up each setting. Example: "Increase this, decrease that, keep this 0 or 1 and you'll do just fine." But that is not the case with you! I like the way you naturally interpret how the proportions of different settings affects the image and know the best blend to mix them in to obtain the desired results. Not only do you understand the 'how', but you are also able to 'communicate' it to your audience. I can watch this video four times, and still learn something new from it.
I really like these videos. Densely packed with information, that is all very useful. They also teach some good core principles. So please keep making more.
This is somehow (maybe mistakenly) Video 2 on the IPAdapter playlist, but is actually Video 1 of the series.
Thank you for explaning things really well without getting sucked into the dumb reddit drama. Looking forward to see how this model evolves, although i've heard rumors it may be tricky to finetune.
Really? I've heard the opposite, the problem is that there's very little official documentation. Controlnets for example were developed in record time and they work pretty well... so maybe it won't be that bad
Anyone knows how you execute only a portion of the woflow so only the preview of the crop will be executed and not the whole workflow as he is doing at ua-cam.com/video/4jq6VQHyXjg/v-deo.html
disable the preview or the save node. The workflow won't execute all the way down but only up to the last enabled prreview
@@latentvision thanks!!
Perfect thanks
I tried your workflow but keeps saying missing node "IP Adapter Apply"
replace the node with ipadapter advanced. things change fast in this world
The word ART is an identity enhancer. What is ART? it's something you can say that you do to feel better about other people. When anyone can do ART then it loses its value as an identity enhancer
11:00 hahaha excellent
😛
Why'd you stop? I want more advanced understanding in my life!!!
I'm prepping another but I feel it's a bit "weak" compared to the first two. I need to re-do something
@@latentvision Just the fact that you're working on it gives me comfort. I'll check back. Thanks! 👍
I really enjoyed this video! Your in-depth look at ComfyUI and Stable Diffusion was both informative and engaging. The way you explained each component, especially the KSampler and texture inversion, made complex concepts much easier to understand. It's clear that you put a lot of effort into this tutorial. Thank you for creating such useful content-looking forward to more videos like this!
After watching this video, I know how Buddha must have felt after reaching enlightenment! :D
lol thanks for the laugh
As a front-end engineer, I completely understand your sentiment and agree with everything you said in this video. You are an incredible person for doing what you do and bringing light to these topics and issues in such a great and non-abrasive way. There are so many times that I wish i could do more to help but I don't know much about Python or the science behind AI that with my toolset I don't know how to help, but honestly if I did know of a way to help or if someone could nudge me in the right way that I could help. i would contribute in a heart beat! Maybe it's time for me to pick up a new skill and learn something new haha, but Like everyone else has already said - we appreciate you and your work tremendously! It really is an inspiration
Great tutorial with some accessible explanations and quite a few tricks. Thanks
13:56❤
This is gold, we need more of this please!!
I've watched a few SD3 videos so far and this has been the best one by quite some distance. Very informative and balanced, thank you :)
Thank you for the laugh. Like always, tellin it like it is… As for the anatomy problem. I think it is just poorly trained. Asking for a person swimming will result in horrors. Why would it try to censor swimming people. I think the training just did not allow for uncommon poses. Perhaps all the synthetic training data was in normal portrait, etc poses. ??
this version of SD3 is a demo, let's keep it at that. Hopefully we'll get a 3.1 soon
@@latentvision I am more excited about the CN working in Comfy and an AD model for what we have... =] I do not normally generate women in the grass any how... =p
I wonder what a textual inversion would do on this, like the character turn around, in some cases the details of the character can be lost. This makes me think that you could use charturner on just one of these inputs, ipadapter for a character reference, and the prompts to help guide it a bit.
What happens when you use a controlnet on the various inputs?
that's technically feasible, haven't tested it yet
It would stand to reason that words like blurry and artifacts wouldn't work as well, because these are features that only become visible at later steps when we're zeroing out the negative embedding. Concepts like blonde, asian, etc. are things that would be visible at the first steps and are thus affected.
"Since at Stability AI they only do the missionary" lol
😉
Your are the best. Very well explained. I see you using that RESCALE CFG node, but i cannot figure what it is doing. If it's a multiplier, why not just put the result in the CFG of the ksampler?
rescaling is gradual based on the sigma, not fixed
I have a question Matteo, please. Using insightface requires a licence. So if I share the outputs that I get from a workflow ,having used insightface.. If I share the outputs on f.x Instagram and get money because I have followers.. is that illegal, is it commercial use,do I need a licence?
technically it's only for educational purposes
"Nobody's gonna know." "They're gonna know." "How are they gonna know?"
@@divye.ruhela haha
How to copy nodes with links??
BROOOO, I CRACKED SO HARD at 13:55, thanks for that xDD
Hi, @Latent Vision How do you use multiple attention mask when you have multiple reference photos ?
there's a video about attention masking!
@@latentvision my bad i thought the mask was for what is considered on the input! But it is for what is influence on the output ! Sorry
I did a git pull to update and also used the comfy manager to update the ComfyUI_essentials nodes but SD3 negative conditioning node is not available?
mh try to delete the directory and reinstall
از آموزش های بسیار جذاب شما بسیار سپاسگذارم شما باعث ارتقا سطح روابط میان کاربران و مهندسان نرم افزار هستید و برخورد شما خوب و منطقی بود ممنون
Give it a couple months so we get custom models and other workarounds rolling. Sdxl was the same, and I only started using it when lightning came out then hyper is my go to choice. I appreciate all the people jumping in to test it and it looks very promising
Nvidia could buy Stabilty Ai, Autodesk and The Foundry to create a new free media development platform, to compliment the massive hardware profits, that harnesses pre-exiting CG development paradigm of points, models, particles textures, compositing etc ..... as an interface to AI rendering through Comfyui etc.... ? Image Control = Essential to Output Professional clients usually operate by specifics. 😊👍
nvidia would never make it open 😅
Hu-po analyses the SD3 technical paper: it's really advanced..... in his opinion. The developers abandoned it without full completion ?
Thumbnail = Trident ie Neptune Pitchfork = two thin prongs ? 😊😅
give me some slack I was in a hurry 😛
@@latentvision It's great ! 😁👍
I'd love to see how IPadapter plays with SD3.
wouldn't we all?
@11:01 I had to rewind this to hear it multiple times to make sure I was hearing correctly. Damn, that's funny stuff!
😛
the best youtube video about SD3 is from someone that dont consider himself a youtuber, thanks for the info
great video as usual!)
I appreciate the hacks you've found and all the work you do, but SD3 is just a garbage model. Not because of its architecture, as you mentioned, but the absolute trash training SAI did for it and how they just dumped it into our laps in this completely broken state. It's not worth the effort to hack just to get it to maybe work. The original employees at SAI maybe would have deserved the benefit of the doubt, but this new leadership group doesn't deserve any of our good will. I would love to see you give your thoughts on some of the alternative models, ie. Pixart Sigma and Lumina Next.
Lumina is my next project
I love this guy!
Very rational and informative video as usual. Thank you!
"Since at Stability AI they only do the missionary"...damn bro, fatality! 😆😆😆
Want to see some analysis of the text encoder difference between fp8 and fp16 version
I can give you the TL;DR: use 16fp