Tell us what is missing in ControlNet Integrated #932
Replies: 85 comments 124 replies
-
There are 3 new controlnet-based architectures to try: ctrlora only SD1.5 for now. One input with multiple preprocessor inputs Modified variants of some blocks that seem to perform well on SVD and SDXL. |
Beta Was this translation helpful? Give feedback.
-
instant id+inpaint PLZ |
Beta Was this translation helpful? Give feedback.
-
First of all I want to say - thanks for the Great tool. That the previous version was very good, that the new version looks very promising (judging by commits). Regarding controlnet.
|
Beta Was this translation helpful? Give feedback.
-
A feature request: segmentation tagging control net (not yet existing) Inputs: character in default frontal pose (photo), a segment map where segment color defines clothing region in frontal pose, a segment map with same colors but in target pose. Desired output: ControlNet should map clothing features from default pose into target pose using color-coded segments as strict guidance. That means - without adding features not present in segment in default pose, without symmetry breaks and so on. Probably specific color channel should hint about SYMMETRY for robustness. Target pose should not be restricted to "usual poses", theoretically upside-down poses, partially occluded poses, self occlusions can be resolved just fine by such 1-1 segmentation mapping |
Beta Was this translation helpful? Give feedback.
-
The outpaint of xinsir union-controlnet-promax ... if i could outpaint with SDXL it would be wonderful. Kepp up the good works ... |
Beta Was this translation helpful? Give feedback.
-
I hopefully update controlnet will work with 2 major extension : btw ur works is awsome |
Beta Was this translation helpful? Give feedback.
-
The biggest need here might simply be a document outlining how Forge's built-in CN differs in options and behavior. It's not much, but the will-not-adds and the workarounds to some of those would be useful for newcomers and converts from A1111. |
Beta Was this translation helpful? Give feedback.
-
First of all, Thank you for your work lllyasviel. |
Beta Was this translation helpful? Give feedback.
-
Well, I can't get it to work in the Photoshop Plugin/Extention... that's missing for me, but probably not something that can be helped since it's pretty much not getting updates on their end. I'd probably rather have it work faster in the UI than have a poor implementation in photoshop anyway at this point. Looking forward to seeing "The Chosen One" for character consistency soon as a CN model or extension of some kind. |
Beta Was this translation helpful? Give feedback.
-
Fooocus sdxl inpaint. Inpaint_v26.fooocus.patch never worked as well on Forge for me as it did on Foocus. |
Beta Was this translation helpful? Give feedback.
-
For hands you also have: And this is looking interesting as well: I've seen people add manual images to OpenPose, but ClickDiff does it automatically, and the other do too. Would be great if we can finally fix hands and feet with ControlNet.
|
Beta Was this translation helpful? Give feedback.
-
1.) SDXL / Pony inpainting would be amazing. 2.) A custom ControlNet which utilizes either photos or prompts to transfer outfits/clothing onto a generation. This could maybe be enhanced by utilizing layerdiffuse to overlay it correctly upon the subject through multiple passes. (Idea, not sure if could be implemented) 3.) I don't know if this is possible either, but instead of something like regional prompter or Forge Couple, creating a controlnet that instead recognizes separate individual characters interacting. I know there are things similar to this that exist, but they aren't quite what I'm looking for to generate multiple characters interacting easily while maintaining a level of versatility and not being constrained to one specific pose via OpenPose. If this idea is at all possible, a controlnet model could utilize a reference image to apply an interaction/action between multiple characters. Or instead of a reference image, it could utilize a syntax similar to Forge Couple's "NEWLINE" , but instead of generating in a new space of the 2d plane, applies character adherence to subject(s) in the scene. Then, an action could be specified, detailing how these characters are interacting. Basically, a prompt adherence helper for models that struggle with prompt adherence, especially as it relates to characteristics of specific individuals in a scene. Example: Again, not sure if this is even possible without a multimodal model, but I think it would be more useful to have something that helps ensure characters look a certain way than it is to make sure they are in a specific place/pose (most of the time.) |
Beta Was this translation helpful? Give feedback.
-
For me, what's missing is
2a. Dsine on independent controlnet for A1111 has a bug that if you use a different resolution/ratio than the image, the preprocessor will generate zoomed into the image rather than crop or resize. Its been a bug since the release of Dsine as a preprocessor. This happens regardless of settings. You essentially have to format images to be the same exact resolution and ratio as the preprocessor settings, otherwise it zooms in, which can be a pain if you're working with images of mixed resolutions. 2b. Propose a feature based on the last problem: A new feature that allows people to use a draggable and resizable selector box over the controlnet image to determine the focus area of the preprocessor. I'd imagine this would also be useful for other controlnet preprocessors too, as you can then take any part of any image and make it the focus of the preprocessor on the fly. So for example, you drag a 1:1 ratio box overlay on the image, have it be resizable, and then it'll use the inputted resolution, even if the image itself is smaller; So you could zoom into a face on a 512x512 image and make just the face area preprocess at 1024x1024. Simple and versatile way of handling controlnet inputs of varying resolutions and aspect ratios and fills the gap that "just resize" and "crop and resize" doesn't cover.
|
Beta Was this translation helpful? Give feedback.
-
Illyas, I'm very new to this and probably the most ignorant of all those involved here, but that doesn't stop me from telling you that I admire not only your great work in Forge, which by the way has been very useful for me as a graphic designer, but, above all, your selflessness. I'm sure that if there were more people like you in this world, things would be different, as we say in Chile. Thank you very much!! A big hug. |
Beta Was this translation helpful? Give feedback.
-
Ah! y lo nuevo de Forge funcionando muy bien en mi 3060ti, algo lento aún pero con muy buena calidad, Flux y SDXL |
Beta Was this translation helpful? Give feedback.
-
You should volunteer to build it if you want to whine about it. This is
FREE software. Oh, you can't code the python? Then get over it.
…On Thu, Nov 21, 2024, 1:53 PM CCpt5 ***@***.***> wrote:
Where is this sourced from? That's extremely disappointing if so. Is there
not another developer who can assist? 3months is an eternity in terms of
generative media.
No disrespect of course to lllyasviel - he's a genius - clearly in high
demand if that feb date is the next possible update date.
—
Reply to this email directly, view it on GitHub
<#932 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BFC6YP4Q5XTWUBT2WF5HDUT2BY23JAVCNFSM6AAAAABMAFM6XCVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTCMZUGA3TCNI>
.
You are receiving this because you commented.Message ID:
<lllyasviel/stable-diffusion-webui-forge/repo-discussions/932/comments/11340715
@github.com>
|
Beta Was this translation helpful? Give feedback.
-
Yeah, actually, you were rude. If you're not volunteering stop complaining.
…On Thu, Nov 21, 2024, 2:07 PM CCpt5 ***@***.***> wrote:
I wasn't rude at all, so stop misreading and acting all tough.
—
Reply to this email directly, view it on GitHub
<#932 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/BFC6YP5AWMX3ZBOFZVHGBBT2BY4QHAVCNFSM6AAAAABMAFM6XCVHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTCMZUGA4DIOA>
.
You are receiving this because you commented.Message ID:
<lllyasviel/stable-diffusion-webui-forge/repo-discussions/932/comments/11340848
@github.com>
|
Beta Was this translation helpful? Give feedback.
-
The delayed rewrite coincides with the release of new tools I see. Perhaps there were insights provided by BFL in private when the delays were decided on. Take your time to do it right guys. All the time you need. Delays are a good thing. Do em more. |
Beta Was this translation helpful? Give feedback.
-
Thing is, someone is waiting for this update (me included) and I am still waiting.. If I knew at beginning it would be so massively delayed, I would just switch my backend to comfy and vioala Forge.. But he delays it like +20 days, +15 days etc, so I am always like.. argh, I will wait it because I have all backend on Forge, so I am not that happy to rewrite it to comfy, but those continous delays already added pretty big pile of days where if I knew it on beginning it would take so long I would rewrite it and went comfy way. But I know @lllyasviel is doing this in his free time and he is legend I appreciate his work very much and he is in no way obligated to deliver something. Just those continous delays are making situation bitter... moreover when he wrote somewhere that implementation of CN Union is like 1-2 hours for him :D so I am like, come on.. I am waiting for it 2 months and still not delivered.. whatever :) I will give it a one more week and then really switching to comfy as it looks like they are shipping new things light fast |
Beta Was this translation helpful? Give feedback.
-
Are we there yet? 🙃 |
Beta Was this translation helpful? Give feedback.
-
Oh my God! What has this thread become?! 😢 |
Beta Was this translation helpful? Give feedback.
-
Maybe he has become fed up,? A lot of self imposed work for little or no recognition, know I would: #2151 It's not like there aren't alternatives, people just don't like to change, and to quote a game developer: " many when given a pile of gold all they can do is complain is too heavy" EDIT: And its not like he is sitted on his hands, recognize the app background? should be be coming to Forge: https://github.com/lllyasviel/IC-Light |
Beta Was this translation helpful? Give feedback.
-
XYZ plot items for ControlNet values |
Beta Was this translation helpful? Give feedback.
-
Let him or her cook. also I would love support for Xinsir SDXL models :) |
Beta Was this translation helpful? Give feedback.
-
Invoke 5.5 just introduced support for Flux Controlnet Canny and Depth. Maybe collaboration with them can help to integrate Flux Controlnet to Forge quicker? |
Beta Was this translation helpful? Give feedback.
-
The integrated ControlNet is not updated for a while, and we are going to make it a bit more up-to-date.
However that will happen after some other newer experiments. Old features are relatively lower priority - so that we can experiment with more new ideas first.
I personally know some models like uni-controlnet (or called promax by some non-research people) and some preprocessors like DSINE/DepthAnything/etc. Those are very easy to add – after I get other things done, those only need me 1 or 2 hours to add them all.
However I will not add the below things:
By the way, several experiments will be implemented in a “Forge Space” (that I will add later, similar to some local HuggingFace Spaces) rather than that ControlNet extension. So many functionalities will happen elsewhere.
This post is only for controlnets. Experiments with other newer diffusion models will happen before controlnets.
Please tell us what is missing in ControlNet Integrated. Please do not talk about things that are already mentioned in this post – I will take a look several days later when I finish some other things.
Update Sep 1:
The rewrite of ControlNet Intergrated will
start at about Sep 29(delayed)start at about Oct 15(delayed)start at about Oct 30(delayed) start at about Nov 20. (When this note is announced, the main targets include some diffusers formatted Flux ControlNets and some community implementation of Union ControlNets. However, this may be extended if stronger models come out after this note.)About IPAdapter:
Note that the CLIP Vision names of
sd-webui-controlnet
are non-standard. Forge uses correct names:Beta Was this translation helpful? Give feedback.
All reactions