GithubHelp home page GithubHelp logo

mikubill / sd-webui-controlnet Goto Github PK

View Code? Open in Web Editor NEW
16.5K 16.5K 1.9K 18 MB

WebUI extension for ControlNet

License: GNU General Public License v3.0

Python 95.06% Jupyter Notebook 0.02% C++ 1.46% Cuda 2.69% Shell 0.07% Dockerfile 0.02% CMake 0.11% JavaScript 0.52% CSS 0.06%

sd-webui-controlnet's Introduction

ControlNet for Stable Diffusion WebUI

The WebUI extension for ControlNet and other injection-based SD controls. image

This extension is for AUTOMATIC1111's Stable Diffusion web UI, allows the Web UI to add ControlNet to the original Stable Diffusion model to generate images. The addition is on-the-fly, the merging is not required.

News

  • [2024-07-09] ๐Ÿ”ฅ[v1.1.454] ControlNet union model support [Discussion thread: #2989]
  • [2024-07-01] ๐Ÿ”ฅ[v1.1.452] Depth Anything V2 - UDAV2 depth Preprocessor [Pull thread: #2969]
  • [2024-05-19] ๐Ÿ”ฅ[v1.1.449] Anyline Preprocessor & MistoLine SDXL model [Discussion thread: #2907]
  • [2024-05-04] ๐Ÿ”ฅ[v1.1.447] PuLID [Discussion thread: #2841]
  • [2024-04-30] ๐Ÿ”ฅ[v1.1.446] Effective region mask supported for ControlNet/IPAdapter [Discussion thread: #2831]
  • [2024-04-27] ๐Ÿ”ฅControlNet-lllite Normal Dsine released [Discussion thread: #2813]
  • [2024-04-19] ๐Ÿ”ฅ[v1.1.445] IPAdapter advanced weight [Instant Style] [Discussion thread: #2770]
  • [2024-04-17] ๐Ÿ”ฅ[v1.1.444] Marigold depth preprocessor [Discussion thread: #2760]
  • [2024-04-15] ๐Ÿ”ฅControlNet++ models released [Discussion thread: #2778]
  • [2024-04-13] ๐Ÿ”ฅTTPLanet_SDXL_Controlnet_Tile_Realistic v2 released [Civitai Page]
  • [2024-03-31] ๐Ÿ”ฅ[v1.1.443] IP-Adapter CLIP mask and ip-adapter-auto preprocessor [Discussion thread: #2723]
  • [2024-03-20] ๐Ÿ”ฅIPAdapter Composition [Discussion thread: #2781]

Installation

  1. Open "Extensions" tab.
  2. Open "Install from URL" tab in the tab.
  3. Enter https://github.com/Mikubill/sd-webui-controlnet.git to "URL for extension's git repository".
  4. Press "Install" button.
  5. Wait for 5 seconds, and you will see the message "Installed into stable-diffusion-webui\extensions\sd-webui-controlnet. Use Installed tab to restart".
  6. Go to "Installed" tab, click "Check for updates", and then click "Apply and restart UI". (The next time you can also use these buttons to update ControlNet.)
  7. Completely restart A1111 webui including your terminal. (If you do not know what is a "terminal", you can reboot your computer to achieve the same effect.)
  8. Download models (see below).
  9. After you put models in the correct folder, you may need to refresh to see the models. The refresh button is right to your "Model" dropdown.

Download Models

You can find all download links here: https://github.com/Mikubill/sd-webui-controlnet/wiki/Model-download.

Features in ControlNet 1.1

Perfect Support for All ControlNet 1.0/1.1 and T2I Adapter Models.

Now we have perfect support all available models and preprocessors, including perfect support for T2I style adapter and ControlNet 1.1 Shuffle. (Make sure that your YAML file names and model file names are same, see also YAML files in "stable-diffusion-webui\extensions\sd-webui-controlnet\models".)

Perfect Support for A1111 High-Res. Fix

Now if you turn on High-Res Fix in A1111, each controlnet will output two different control images: a small one and a large one. The small one is for your basic generating, and the big one is for your High-Res Fix generating. The two control images are computed by a smart algorithm called "super high-quality control image resampling". This is turned on by default, and you do not need to change any setting.

Perfect Support for All A1111 Img2Img or Inpaint Settings and All Mask Types

Now ControlNet is extensively tested with A1111's different types of masks, including "Inpaint masked"/"Inpaint not masked", and "Whole picture"/"Only masked", and "Only masked padding"&"Mask blur". The resizing perfectly matches A1111's "Just resize"/"Crop and resize"/"Resize and fill". This means you can use ControlNet in nearly everywhere in your A1111 UI without difficulty!

The New "Pixel-Perfect" Mode

Now if you turn on pixel-perfect mode, you do not need to set preprocessor (annotator) resolutions manually. The ControlNet will automatically compute the best annotator resolution for you so that each pixel perfectly matches Stable Diffusion.

User-Friendly GUI and Preprocessor Preview

We reorganized some previously confusing UI like "canvas width/height for new canvas" and it is in the ๐Ÿ“ button now. Now the preview GUI is controlled by the "allow preview" option and the trigger button ๐Ÿ’ฅ. The preview image size is better than before, and you do not need to scroll up and down - your a1111 GUI will not be messed up anymore!

Support for Almost All Upscaling Scripts

Now ControlNet 1.1 can support almost all Upscaling/Tile methods. ControlNet 1.1 support the script "Ultimate SD upscale" and almost all other tile-based extensions. Please do not confuse "Ultimate SD upscale" with "SD upscale" - they are different scripts. Note that the most recommended upscaling method is "Tiled VAE/Diffusion" but we test as many methods/extensions as possible. Note that "SD upscale" is supported since 1.1.117, and if you use it, you need to leave all ControlNet images as blank (We do not recommend "SD upscale" since it is somewhat buggy and cannot be maintained - use the "Ultimate SD upscale" instead).

More Control Modes (previously called Guess Mode)

We have fixed many bugs in previous 1.0โ€™s Guess Mode and now it is called Control Mode

image

Now you can control which aspect is more important (your prompt or your ControlNet)๏ผš

  • "Balanced": ControlNet on both sides of CFG scale, same as turning off "Guess Mode" in ControlNet 1.0

  • "My prompt is more important": ControlNet on both sides of CFG scale, with progressively reduced SD U-Net injections (layer_weight*=0.825**I, where 0<=I <13, and the 13 means ControlNet injected SD 13 times). In this way, you can make sure that your prompts are perfectly displayed in your generated images.

  • "ControlNet is more important": ControlNet only on the Conditional Side of CFG scale (the cond in A1111's batch-cond-uncond). This means the ControlNet will be X times stronger if your cfg-scale is X. For example, if your cfg-scale is 7, then ControlNet is 7 times stronger. Note that here the X times stronger is different from "Control Weights" since your weights are not modified. This "stronger" effect usually has less artifact and give ControlNet more room to guess what is missing from your prompts (and in the previous 1.0, it is called "Guess Mode").

Input (depth+canny+hed) "Balanced" "My prompt is more important" "ControlNet is more important"

Reference-Only Control

Now we have a reference-only preprocessor that does not require any control models. It can guide the diffusion directly using images as references.

(Prompt "a dog running on grassland, best quality, ...")

image

This method is similar to inpaint-based reference but it does not make your image disordered.

Many professional A1111 users know a trick to diffuse image with references by inpaint. For example, if you have a 512x512 image of a dog, and want to generate another 512x512 image with the same dog, some users will connect the 512x512 dog image and a 512x512 blank image into a 1024x512 image, send to inpaint, and mask out the blank 512x512 part to diffuse a dog with similar appearance. However, that method is usually not very satisfying since images are connected and many distortions will appear.

This reference-only ControlNet can directly link the attention layers of your SD to any independent images, so that your SD will read arbitrary images for reference. You need at least ControlNet 1.1.153 to use it.

To use, just select reference-only as preprocessor and put an image. Your SD will just use the image as reference.

Note that this method is as "non-opinioned" as possible. It only contains very basic connection codes, without any personal preferences, to connect the attention layers with your reference images. However, even if we tried best to not include any opinioned codes, we still need to write some subjective implementations to deal with weighting, cfg-scale, etc - tech report is on the way.

More examples here.

Technical Documents

See also the documents of ControlNet 1.1:

https://github.com/lllyasviel/ControlNet-v1-1-nightly#model-specification

Default Setting

This is my setting. If you run into any problem, you can use this setting as a sanity check

image

Use Previous Models

Use ControlNet 1.0 Models

https://huggingface.co/lllyasviel/ControlNet/tree/main/models

You can still use all previous models in the previous ControlNet 1.0. Now, the previous "depth" is now called "depth_midas", the previous "normal" is called "normal_midas", the previous "hed" is called "softedge_hed". And starting from 1.1, all line maps, edge maps, lineart maps, boundary maps will have black background and white lines.

Use T2I-Adapter Models

(From TencentARC/T2I-Adapter)

To use T2I-Adapter models:

  1. Download files from https://huggingface.co/TencentARC/T2I-Adapter/tree/main/models
  2. Put them in "stable-diffusion-webui\extensions\sd-webui-controlnet\models".
  3. Make sure that the file names of pth files and yaml files are consistent.

Note that "CoAdapter" is not implemented yet.

Gallery

The below results are from ControlNet 1.0.

Source Input Output
(no preprocessor)
(no preprocessor)

The below examples are from T2I-Adapter.

From t2iadapter_color_sd14v1.pth :

Source Input Output

From t2iadapter_style_sd14v1.pth :

Source Input Output
(clip, non-image)

Minimum Requirements

  • (Windows) (NVIDIA: Ampere) 4gb - with --xformers enabled, and Low VRAM mode ticked in the UI, goes up to 768x832

Multi-ControlNet

This option allows multiple ControlNet inputs for a single generation. To enable this option, change Multi ControlNet: Max models amount (requires restart) in the settings. Note that you will need to restart the WebUI for changes to take effect.

Source A Source B Output

Control Weight/Start/End

Weight is the weight of the controlnet "influence". It's analogous to prompt attention/emphasis. E.g. (myprompt: 1.2). Technically, it's the factor by which to multiply the ControlNet outputs before merging them with original SD Unet.

Guidance Start/End is the percentage of total steps the controlnet applies (guidance strength = guidance end). It's analogous to prompt editing/shifting. E.g. [myprompt::0.8] (It applies from the beginning until 80% of total steps)

Batch Mode

Put any unit into batch mode to activate batch mode for all units. Specify a batch directory for each unit, or use the new textbox in the img2img batch tab as a fallback. Although the textbox is located in the img2img batch tab, you can use it to generate images in the txt2img tab as well.

Note that this feature is only available in the gradio user interface. Call the APIs as many times as you want for custom batch scheduling.

API and Script Access

This extension can accept txt2img or img2img tasks via API or external extension call. Note that you may need to enable Allow other scripts to control this extension in settings for external calls.

To use the API: start WebUI with argument --api and go to http://webui-address/docs for documents or checkout examples.

To use external call: Checkout Wiki

Command Line Arguments

This extension adds these command line arguments to the webui:

    --controlnet-dir <path to directory with controlnet models>                                ADD a controlnet models directory
    --controlnet-annotator-models-path <path to directory with annotator model directories>    SET the directory for annotator models
    --no-half-controlnet                                                                       load controlnet models in full precision
    --controlnet-preprocessor-cache-size                                                       Cache size for controlnet preprocessor results
    --controlnet-loglevel                                                                      Log level for the controlnet extension
    --controlnet-tracemalloc                                                                   Enable malloc memory tracing

MacOS Support

Tested with pytorch nightly: #143 (comment)

To use this extension with mps and normal pytorch, currently you may need to start WebUI with --no-half.

Archive of Deprecated Versions

The previous version (sd-webui-controlnet 1.0) is archived in

https://github.com/lllyasviel/webui-controlnet-v1-archived

Using this version is not a temporary stop of updates. You will stop all updates forever.

Please consider this version if you work with professional studios that requires 100% reproducing of all previous results pixel by pixel.

Thanks

This implementation is inspired by kohya-ss/sd-webui-additional-networks

sd-webui-controlnet's People

Contributors

105gun avatar aiton-sd avatar brkirch avatar catboxanon avatar ccrcmcpe avatar cmeka avatar danielkauss avatar ddpn08 avatar etdofresh avatar fishslot avatar gitadmin0608 avatar guillaume-fgt avatar hithereai avatar huchenlei avatar josephcatrambone-crucible avatar kft334 avatar ljleb avatar lllyasviel avatar mattyamonaca avatar mikubill avatar mishafarms avatar missionfloyd avatar phoenixcreation avatar sangww avatar sdbds avatar space-nuko avatar timmahw avatar vespinian avatar w-e-w avatar zombieyang avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

sd-webui-controlnet's Issues

How to use Normal Map?

p18

I set the preprocessor - midas / model - control_sd15_normal, but it doesn't work.

How can I use Normal map in this extension.

Add option for secondary input resize modes

Currently, the preprocessed output is resized in a fixed way.

control = Resize(h if h>w else w, interpolation=InterpolationMode.BICUBIC)(control)
control = CenterCrop((h, w))(control)

I propose adding two more options, on a dropdown / radio button:

  • Scale to Fit
control = Resize(h if h<w else w, interpolation=InterpolationMode.BICUBIC)(control)
control = CenterCrop((h, w))(control)
  • Just Resize
control = Resize((h,w), interpolation=InterpolationMode.BICUBIC)(control)

This should solve some cropping issues with non 1:1 aspect ratio inputs

Img2img support?

I don't see a ControlNet section on the img2img tab, but as far as I know, it should be something the model could support. I think it would allow some extremely fine control over the output.

Error when using Img2Img Batch

Hello,

I get this error when I try to process multiple images with the img2img batch function:

Traceback (most recent call last):
  File "C:\StableDiffusion2023\stable-diffusion-webui-master(3)\stable-diffusion-webui-master\modules\call_queue.py", line 56, in f
   res = list(func(*args, **kwargs))
  File "C:\StableDiffusion2023\stable-diffusion-webui-master(3)\stable-diffusion-webui-master\modules\call_queue.py", line 37, in f
    res = func(*args, **kwargs)
  File "C:\StableDiffusion2023\stable-diffusion-webui-master(3)\stable-diffusion-webui-master\modules\img2img.py", line 163, in img2img
    process_batch(p, img2img_batch_input_dir, img2img_batch_output_dir, img2img_batch_inpaint_mask_dir, args)
  File "C:\StableDiffusion2023\stable-diffusion-webui-master(3)\stable-diffusion-webui-master\modules\img2img.py", line 76, in process_batch
    processed_image.save(os.path.join(output_dir, filename))
AttributeError: 'numpy.ndarray' object has no attribute 'save'

I tried to use different preprocessors and models but everytime I get the same error. When I just use img2img with just a single image there is no error. I using Windows 10 whith a NVIDIA RTX 3090 Ti. I set the Input directory and the Output directory correctly.

[Feature Request] move models and midas to stable-diffusion-webui\models\

Examples:
https://github.com/thygate/stable-diffusion-webui-depthmap-script - uses MiDaS and LeReS directly from webui\models\ so its not needed to download it again, can be shared by multiple extensions
https://github.com/Extraltodeus/depthmap2mask - for midas as well

https://github.com/arenatemp/stable-diffusion-webui-model-toolkit - dont remember if it downloads something to models, but it extracts components from checkpoints to a folder there
https://github.com/Klace/stable-diffusion-webui-instruct-pix2pix - was already doing that before being integrated into main

there are some other extensions that either share or create their own folders in models

also ControlNet folder in default path IMO would be better suited

anyway, outstanding job, really well done, thanks

EDIT:
dpt_hybrid-midas-501f0c75 - specificly i have it already in stable-diffusion-webui\models\midas, for depth extensions, aslo:
dpt_beit_large_384.pt
dpt_beit_large_512.pt
dpt_large-midas-2f21e586.pt
midas_v21_small-70d6b9c8.pt
midas_v21-f6b98070.pt

Weight?

what does Weight for? seem like adjust it doesn't change anything?

Feature Request: Directly feed openpose data

It would be amazing if I could take existing openpose pose data and feed it directly into the txt2img or img2img process rather than it trying to first generate the pose estimation from a given image (which it may or may not get right).

ControlNet not working or activating

Hello,

Just got back from work and been hear the craze over this. I installed the extension, updated my WebUI, got everything set up, appiled the highres fix, but whenever I generate an image with ControlNet enabled, I get hit with this error.

Error running process: C:\Users\Joseph\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py
Traceback (most recent call last):
File "C:\Users\Joseph\stable-diffusion-webui\modules\scripts.py", line 386, in process
script.process(p, *script_args)
File "C:\Users\Joseph\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py", line 231, in process
restore_networks()
File "C:\Users\Joseph\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py", line 217, in restore_networks
self.latest_network.restore(unet)
File "C:\Users\Joseph\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\cldm.py", line 111, in restore
model.forward = model._original_forward
File "C:\Users\Joseph\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1269, in getattr
raise AttributeError("'{}' object has no attribute '{}'".format(
AttributeError: 'UNetModel' object has no attribute '_original_forward'

Tried to figure out what's going on, but this is something far beyond my knowledge. Any help would be appreciated. Thanks.

ไป€ไนˆๆ—ถๅ€™ๅฏไปฅๆ”ฏๆŒhighfixๅ’Œi2i

็›ฎๅ‰ๆ— ๆณ•็”Ÿๅญ˜้ซ˜ๅˆ†่พจ็Ž‡ๅ›พๅพˆๅฏๆƒœๅ•Š๏ผŒ็”จๅทฒ็ป็”Ÿๆˆๅฅฝ็š„ๅ›พๅŽปi2iๆˆ–่€…highres fix๏ผŒๅŠจไฝœๅพˆๅฎนๆ˜“ๅฐฑๅ˜ๅฝขๆˆ–่€…ๅ˜ๅพ—ๅฅ‡ๆ€ชไบ†

Option to add good upscaling for the mask images

When I use higher resolutions the black and white mask image gets blurry and the details wont come out as good.

Can it be upscaled in the background with SwinIR or Esrgan, when choosing a resolution above 512x512? The details would then stay reasonably crispy

Adding depth or similar function to Pose preprocessor

This tool is a blast! Especially the pose one! Absolutely brilliant. The only thing it lacks is somekind detect the depth (like the script does) of characters, determine which are closer, which are farther, and what limbs or body parts are behind the body or objects. For example if the model put her hands behind back, it thinks just that they are not visible and make non visible parts random but in front, not behind. And same for characters that are behind other characters, it tries to draw them not behind. I can't fully suggest how to achieve it but it would be awesome!

P.S. I tried to use both Controlnet + depth script, and it somewhy drags images from original image even when denoise is 1.

Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0!

Errors will be reported when the new version is used:

Error completing request
Arguments: ('task(88y7i4r4jcugm8w)', '1girl', 'lowres, bad anatomy, bad hands, (text:1.6), error, missing fingers, clothe pull, extra digit, fewer digits, (cropped:1.2), (censored:1.2), (low quality, worst quality:1.4), fat, (dress, ribbon:1.2), pubic hair, jpeg artifacts, (signature, watermark, username:1.3), (blurry:1.2), mutated, mutation, out of focus, mutated, extra limb, poorly drawn hands and fingers, missing limb, floating limbs, disconnected limbs, malformed hands and fingers, (motion lines:1.2)', [], 30, 15, False, False, 1, 1, 9, -1.0, -1.0, 0, 0, 0, False, 672, 480, False, 0.35, 2, '4x_foolhardy_Remacri', 20, 0, 0, [], 0, False, False, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'Refresh models', True, 'openpose', 'control_any3_openpose(95070f75)', 1, {'image': array([[[238, 247, 255],
[236, 245, 254],
[233, 242, 251],
...,
[192, 203, 225],
[187, 198, 220],
[183, 194, 216]],

   [[236, 245, 254],
    [235, 244, 253],
    [233, 242, 251],
    ...,
    [183, 194, 216],
    [178, 189, 211],
    [175, 186, 208]],

   [[233, 242, 251],
    [233, 242, 251],
    [232, 241, 250],
    ...,
    [184, 195, 215],
    [179, 190, 210],
    [176, 187, 207]],

   ...,

   [[181, 181, 189],
    [181, 181, 189],
    [180, 180, 188],
    ...,
    [225, 225, 233],
    [210, 210, 218],
    [201, 201, 209]],

   [[181, 181, 189],
    [181, 181, 189],
    [181, 181, 189],
    ...,
    [224, 224, 232],
    [220, 220, 228],
    [212, 212, 220]],

   [[181, 181, 189],
    [181, 181, 189],
    [181, 181, 189],
    ...,
    [218, 218, 226],
    [225, 225, 233],
    [218, 218, 226]]], dtype=uint8), 'mask': array([[[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]],

   [[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]],

   [[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]],

   ...,

   [[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]],

   [[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]],

   [[  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    ...,
    [  0,   0,   0, 255],
    [  0,   0,   0, 255],
    [  0,   0,   0, 255]]], dtype=uint8)}, False, 'Scale to Fit (Inner Fit)', True, False, False, 'positive', 'comma', 0, False, False, '', 1, '', 0, '', 0, '', True, False, False, False, 0) {}

Traceback (most recent call last):
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\call_queue.py", line 56, in f
res = list(func(*args, **kwargs))
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\call_queue.py", line 37, in f
res = func(*args, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\txt2img.py", line 56, in txt2img
processed = process_images(p)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\processing.py", line 486, in process_images
res = process_images_inner(p)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\processing.py", line 628, in process_images_inner
samples_ddim = p.sample(conditioning=c, unconditional_conditioning=uc, seeds=seeds, subseeds=subseeds, subseed_strength=p.subseed_strength, prompts=prompts)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\processing.py", line 828, in sample
samples = self.sampler.sample(self, x, conditioning, unconditional_conditioning, image_conditioning=self.txt2img_image_conditioning(x))
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 323, in sample
samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 221, in launch_sampling
return func()
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 323, in
samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\sampling.py", line 594, in sample_dpmpp_2m
denoised = model(x, sigmas[i] * s_in, **extra_args)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 135, in forward
x_out[a:b] = self.inner_model(x_in[a:b], sigma_in[a:b], cond={"c_crossattn": c_crossattn, "c_concat": [image_cond_in[a:b]]})
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 112, in forward
eps = self.get_eps(input * c_in, self.sigma_to_t(sigma), **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 138, in get_eps
return self.inner_model.apply_model(*args, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_hijack_utils.py", line 17, in
setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs))
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\modules\sd_hijack_utils.py", line 28, in call
return self.__orig_func(*args, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 858, in apply_model
x_recon = self.model(x_noisy, t, **cond)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1212, in _call_impl
result = forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 1329, in forward
out = self.diffusion_model(x, t, context=cc)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\cldm.py", line 97, in forward2
return forward(*args, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\cldm.py", line 70, in forward
emb = self.time_embed(t_emb)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\container.py", line 204, in forward
input = module(input)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\extensions-builtin\Lora\lora.py", line 178, in lora_Linear_forward
return lora_forward(self, input, torch.nn.Linear_forward_before_lora(self, input))
File "C:\Users\60552\Documents\AI folder\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\linear.py", line 114, in forward
return F.linear(input, self.weight, self.bias)
RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cpu and cuda:0! (when checking argument for argument mat1 in method wrapper_addmm)

Change "midas" preprocessor text to "depth"

This is a minor request, but changing the preprocessor text of the midas depth model from "midas" to "depth" would allow it to match the model's actual name, like with the other ControlNet preprocessors. This could help prevent some potential confusion for users who download the depth model but don't immediately see a "depth" option. Personally, even though I'm aware that midas = depth, it's still an extra reminder I have to give myself to select the right preprocessor each time. As a bonus, the sorting for the preprocessors and the models will also match better.

[Feature Request] Upload our own DepthMap/NormalMap/Pose/etc...

This would be a useful feature for example when starting from a 3D Model you rendered, which means you can also render better maps than the auto-generated ones, and they could be fed to the process for more accurate results.
Pretty sure there is an extension for the 2.0 depth model exactly for this purpose.

Thanks for listening.

Is there any parameter can adjust it's controlling strength?

Hi, greate tool, but there is one problem.

When source image is a 2D image and use a realistic model to generate target image, it copied everything from that 2D image, which makes output image not realistic anymore. More like a 3D rendering result.

Even I choose openpose as control net's model, still has this issue. Character's eyes are too big, mouth size is unrealistic, hair is super big. All those 2D image's feature, comes to the result, even with openpose.

So, is there a way to add a parameter to control its strength?

Thanks

Feature request: cache preprocessor outputs

Right now if you're trying a bunch of prompts on the same starting image you have to wait for the preprocessor each time, or manually upload the preprocessed image. This could be cached and reused.

Requesting Crop and Resize mode

bbbb

Here is the apple image (835 x 1000 / 1:1.2 ratio), But canvas size is 512 x 768 (1:1.5 ratio)

In this case, current resize mode work like this.

06027-1523115-masterpiece, best quality, apple

Could you add Crop and Resize mode? Crop as canvas ratio first and resize.

If you look at WEBUI img2img, there is a "crop and resize" function, which is the same as the method I suggested.

RuntimeError: Sizes of tensors must match except in dimension 1. When using openpose model.

If i change width or heigth to something other than 512 i get:

RuntimeError: Sizes of tensors must match except in dimension 1.

Also, canvas width and height are currently reversed in your script. Increasing canvas width actually increases height of the canvas.

using mask as input
  0%|                                                                                           | 0/20 [00:00<?, ?it/s]
Error completing request
Arguments: ('task(p823x0yzj0q7ebe)', '1girl black hair, ', '(worst quality:1.2), (low quality:1.2) , (monochrome:0.7)', [], 20, 0, False, False, 1, 1, 7, -1.0, -1.0, 0, 0, 0, False, 456, 344, False, 0.7, 2, 'Latent', 0, 0, 0, [], 0, False, False, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'LoRA', 'None', 1, 1, 'Refresh models', True, 'openpose', 'out(b46e25f5)', 1, {'image': array([[[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]],

       [[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]],

       [[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]],

       ...,

       [[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]],

       [[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]],

       [[255, 255, 255],
        [255, 255, 255],
        [255, 255, 255],
        ...,
        [255, 255, 255],
        [255, 255, 255],
        [255, 255, 255]]], dtype=uint8), 'mask': array([[[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]],

       [[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]],

       [[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]],

       ...,

       [[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]],

       [[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]],

       [[  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        ...,
        [  0,   0,   0, 255],
        [  0,   0,   0, 255],
        [  0,   0,   0, 255]]], dtype=uint8)}, False, False, False, 'positive', 'comma', 0, False, False, '', 1, '', 0, '', 0, '', True, False, False, False, 0) {}
Traceback (most recent call last):
  File "B:\AIimages\stable-diffusion-webui\modules\call_queue.py", line 56, in f
    res = list(func(*args, **kwargs))
  File "B:\AIimages\stable-diffusion-webui\modules\call_queue.py", line 37, in f
    res = func(*args, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\modules\txt2img.py", line 56, in txt2img
    processed = process_images(p)
  File "B:\AIimages\stable-diffusion-webui\modules\processing.py", line 486, in process_images
    res = process_images_inner(p)
  File "B:\AIimages\stable-diffusion-webui\modules\processing.py", line 628, in process_images_inner
    samples_ddim = p.sample(conditioning=c, unconditional_conditioning=uc, seeds=seeds, subseeds=subseeds, subseed_strength=p.subseed_strength, prompts=prompts)
  File "B:\AIimages\stable-diffusion-webui\modules\processing.py", line 828, in sample
    samples = self.sampler.sample(self, x, conditioning, unconditional_conditioning, image_conditioning=self.txt2img_image_conditioning(x))
  File "B:\AIimages\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 327, in sample
    samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
  File "B:\AIimages\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 225, in launch_sampling
    return func()
  File "B:\AIimages\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 327, in <lambda>
    samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
  File "B:\AIimages\stable-diffusion-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
    return func(*args, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\sampling.py", line 145, in sample_euler_ancestral
    denoised = model(x, sigmas[i] * s_in, **extra_args)
  File "B:\AIimages\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
    return forward_call(*input, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 123, in forward
    x_out[a:b] = self.inner_model(x_in[a:b], sigma_in[a:b], cond={"c_crossattn": [cond_in[a:b]], "c_concat": [image_cond_in[a:b]]})
  File "B:\AIimages\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
    return forward_call(*input, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 112, in forward
    eps = self.get_eps(input * c_in, self.sigma_to_t(sigma), **kwargs)
  File "B:\AIimages\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 138, in get_eps
    return self.inner_model.apply_model(*args, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\modules\sd_hijack_utils.py", line 17, in <lambda>
    setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs))
  File "B:\AIimages\stable-diffusion-webui\modules\sd_hijack_utils.py", line 28, in __call__
    return self.__orig_func(*args, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 858, in apply_model
    x_recon = self.model(x_noisy, t, **cond)
  File "B:\AIimages\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1212, in _call_impl
    result = forward_call(*input, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 1329, in forward
    out = self.diffusion_model(x, t, context=cc)
  File "B:\AIimages\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
    return forward_call(*input, **kwargs)
  File "B:\AIimages\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\cldm.py", line 77, in forward
    h = torch.cat(
RuntimeError: Sizes of tensors must match except in dimension 1. Expected size 16 but got size 15 for tensor number 1 in the list.

Midas loads repeatedly without unloading?

So one thing I'm noticing is that, when I change away from a depth-based model (or just disable the ControlNet stuff for a few gens), and then switch back, it tries to load Midas again, and fails with a CUDA memory error. I think it tries to reload the model without having first properly unloaded it. I have to restart the entire program each time this happens, with 16 gb ram and 10 gb vram.

Error when running in google colab

i got this error, how to fix?
Loading preprocessor: canny, model: body_pose_model(f14788ab) Loaded state_dict from [/content/stable-diffusion-webui/extensions/sd-webui-controlnet/models/body_pose_model.pth] Error running process: /content/stable-diffusion-webui/extensions/sd-webui-controlnet/scripts/controlnet.py Traceback (most recent call last): File "/content/stable-diffusion-webui/modules/scripts.py", line 386, in process script.process(p, *script_args) File "/content/stable-diffusion-webui/extensions/sd-webui-controlnet/scripts/controlnet.py", line 252, in process network = PlugableControlModel(model_path, os.path.join(cn_models_dir, "cldm_v15.yaml"), weight, lowvram=lowvram) File "/content/stable-diffusion-webui/extensions/sd-webui-controlnet/scripts/cldm.py", line 57, in __init__ self.control_model.load_state_dict(state_dict) File "/usr/local/lib/python3.8/dist-packages/torch/nn/modules/module.py", line 1671, in load_state_dict raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format( RuntimeError: Error(s) in loading state_dict for ControlNet: Missing key(s) in state_dict: "time_embed.0.weight", "time_embed.0.bias", "time_embed.2.weight", "time_embed.2.bias", "input_blocks.0.0.weight", "input_blocks.0.0.bias", "input_blocks.1.0.in_layers.0.weight", "input_blocks.1.0.in_layers.0.bias", "input_blocks.1.0.in_layers.2.weight", "input_blocks.1.0.in_layers.2.bias", "input_blocks.1.0.emb_layers.1.weight", "input_blocks.1.0.emb_layers.1.bias", "input_blocks.1.0.out_layers.0.weight", "input_blocks.1.0.out_layers.0.bias", "input_blocks.1.0.out_layers.3.weight", "input_blocks.1.0.out_layers.3.bias", "input_blocks.1.1.norm.weight", "input_blocks.1.1.norm.bias", "input_blocks.1.1.proj_in.weight", "input_blocks.1.1.proj_in.bias", "input_blocks.1.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.1.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.1.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.1.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.1.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.1.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.1.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.1.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.1.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.1.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.1.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.1.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.1.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.1.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.1.1.transformer_blocks.0.norm1.weight", "input_blocks.1.1.transformer_blocks.0.norm1.bias", "input_blocks.1.1.transformer_blocks.0.norm2.weight", "input_blocks.1.1.transformer_blocks.0.norm2.bias", "input_blocks.1.1.transformer_blocks.0.norm3.weight", "input_blocks.1.1.transformer_blocks.0.norm3.bias", "input_blocks.1.1.proj_out.weight", "input_blocks.1.1.proj_out.bias", "input_blocks.2.0.in_layers.0.weight", "input_blocks.2.0.in_layers.0.bias", "input_blocks.2.0.in_layers.2.weight", "input_blocks.2.0.in_layers.2.bias", "input_blocks.2.0.emb_layers.1.weight", "input_blocks.2.0.emb_layers.1.bias", "input_blocks.2.0.out_layers.0.weight", "input_blocks.2.0.out_layers.0.bias", "input_blocks.2.0.out_layers.3.weight", "input_blocks.2.0.out_layers.3.bias", "input_blocks.2.1.norm.weight", "input_blocks.2.1.norm.bias", "input_blocks.2.1.proj_in.weight", "input_blocks.2.1.proj_in.bias", "input_blocks.2.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.2.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.2.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.2.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.2.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.2.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.2.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.2.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.2.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.2.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.2.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.2.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.2.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.2.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.2.1.transformer_blocks.0.norm1.weight", "input_blocks.2.1.transformer_blocks.0.norm1.bias", "input_blocks.2.1.transformer_blocks.0.norm2.weight", "input_blocks.2.1.transformer_blocks.0.norm2.bias", "input_blocks.2.1.transformer_blocks.0.norm3.weight", "input_blocks.2.1.transformer_blocks.0.norm3.bias", "input_blocks.2.1.proj_out.weight", "input_blocks.2.1.proj_out.bias", "input_blocks.3.0.op.weight", "input_blocks.3.0.op.bias", "input_blocks.4.0.in_layers.0.weight", "input_blocks.4.0.in_layers.0.bias", "input_blocks.4.0.in_layers.2.weight", "input_blocks.4.0.in_layers.2.bias", "input_blocks.4.0.emb_layers.1.weight", "input_blocks.4.0.emb_layers.1.bias", "input_blocks.4.0.out_layers.0.weight", "input_blocks.4.0.out_layers.0.bias", "input_blocks.4.0.out_layers.3.weight", "input_blocks.4.0.out_layers.3.bias", "input_blocks.4.0.skip_connection.weight", "input_blocks.4.0.skip_connection.bias", "input_blocks.4.1.norm.weight", "input_blocks.4.1.norm.bias", "input_blocks.4.1.proj_in.weight", "input_blocks.4.1.proj_in.bias", "input_blocks.4.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.4.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.4.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.4.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.4.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.4.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.4.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.4.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.4.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.4.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.4.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.4.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.4.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.4.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.4.1.transformer_blocks.0.norm1.weight", "input_blocks.4.1.transformer_blocks.0.norm1.bias", "input_blocks.4.1.transformer_blocks.0.norm2.weight", "input_blocks.4.1.transformer_blocks.0.norm2.bias", "input_blocks.4.1.transformer_blocks.0.norm3.weight", "input_blocks.4.1.transformer_blocks.0.norm3.bias", "input_blocks.4.1.proj_out.weight", "input_blocks.4.1.proj_out.bias", "input_blocks.5.0.in_layers.0.weight", "input_blocks.5.0.in_layers.0.bias", "input_blocks.5.0.in_layers.2.weight", "input_blocks.5.0.in_layers.2.bias", "input_blocks.5.0.emb_layers.1.weight", "input_blocks.5.0.emb_layers.1.bias", "input_blocks.5.0.out_layers.0.weight", "input_blocks.5.0.out_layers.0.bias", "input_blocks.5.0.out_layers.3.weight", "input_blocks.5.0.out_layers.3.bias", "input_blocks.5.1.norm.weight", "input_blocks.5.1.norm.bias", "input_blocks.5.1.proj_in.weight", "input_blocks.5.1.proj_in.bias", "input_blocks.5.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.5.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.5.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.5.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.5.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.5.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.5.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.5.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.5.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.5.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.5.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.5.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.5.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.5.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.5.1.transformer_blocks.0.norm1.weight", "input_blocks.5.1.transformer_blocks.0.norm1.bias", "input_blocks.5.1.transformer_blocks.0.norm2.weight", "input_blocks.5.1.transformer_blocks.0.norm2.bias", "input_blocks.5.1.transformer_blocks.0.norm3.weight", "input_blocks.5.1.transformer_blocks.0.norm3.bias", "input_blocks.5.1.proj_out.weight", "input_blocks.5.1.proj_out.bias", "input_blocks.6.0.op.weight", "input_blocks.6.0.op.bias", "input_blocks.7.0.in_layers.0.weight", "input_blocks.7.0.in_layers.0.bias", "input_blocks.7.0.in_layers.2.weight", "input_blocks.7.0.in_layers.2.bias", "input_blocks.7.0.emb_layers.1.weight", "input_blocks.7.0.emb_layers.1.bias", "input_blocks.7.0.out_layers.0.weight", "input_blocks.7.0.out_layers.0.bias", "input_blocks.7.0.out_layers.3.weight", "input_blocks.7.0.out_layers.3.bias", "input_blocks.7.0.skip_connection.weight", "input_blocks.7.0.skip_connection.bias", "input_blocks.7.1.norm.weight", "input_blocks.7.1.norm.bias", "input_blocks.7.1.proj_in.weight", "input_blocks.7.1.proj_in.bias", "input_blocks.7.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.7.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.7.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.7.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.7.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.7.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.7.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.7.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.7.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.7.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.7.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.7.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.7.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.7.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.7.1.transformer_blocks.0.norm1.weight", "input_blocks.7.1.transformer_blocks.0.norm1.bias", "input_blocks.7.1.transformer_blocks.0.norm2.weight", "input_blocks.7.1.transformer_blocks.0.norm2.bias", "input_blocks.7.1.transformer_blocks.0.norm3.weight", "input_blocks.7.1.transformer_blocks.0.norm3.bias", "input_blocks.7.1.proj_out.weight", "input_blocks.7.1.proj_out.bias", "input_blocks.8.0.in_layers.0.weight", "input_blocks.8.0.in_layers.0.bias", "input_blocks.8.0.in_layers.2.weight", "input_blocks.8.0.in_layers.2.bias", "input_blocks.8.0.emb_layers.1.weight", "input_blocks.8.0.emb_layers.1.bias", "input_blocks.8.0.out_layers.0.weight", "input_blocks.8.0.out_layers.0.bias", "input_blocks.8.0.out_layers.3.weight", "input_blocks.8.0.out_layers.3.bias", "input_blocks.8.1.norm.weight", "input_blocks.8.1.norm.bias", "input_blocks.8.1.proj_in.weight", "input_blocks.8.1.proj_in.bias", "input_blocks.8.1.transformer_blocks.0.attn1.to_q.weight", "input_blocks.8.1.transformer_blocks.0.attn1.to_k.weight", "input_blocks.8.1.transformer_blocks.0.attn1.to_v.weight", "input_blocks.8.1.transformer_blocks.0.attn1.to_out.0.weight", "input_blocks.8.1.transformer_blocks.0.attn1.to_out.0.bias", "input_blocks.8.1.transformer_blocks.0.ff.net.0.proj.weight", "input_blocks.8.1.transformer_blocks.0.ff.net.0.proj.bias", "input_blocks.8.1.transformer_blocks.0.ff.net.2.weight", "input_blocks.8.1.transformer_blocks.0.ff.net.2.bias", "input_blocks.8.1.transformer_blocks.0.attn2.to_q.weight", "input_blocks.8.1.transformer_blocks.0.attn2.to_k.weight", "input_blocks.8.1.transformer_blocks.0.attn2.to_v.weight", "input_blocks.8.1.transformer_blocks.0.attn2.to_out.0.weight", "input_blocks.8.1.transformer_blocks.0.attn2.to_out.0.bias", "input_blocks.8.1.transformer_blocks.0.norm1.weight", "input_blocks.8.1.transformer_blocks.0.norm1.bias", "input_blocks.8.1.transformer_blocks.0.norm2.weight", "input_blocks.8.1.transformer_blocks.0.norm2.bias", "input_blocks.8.1.transformer_blocks.0.norm3.weight", "input_blocks.8.1.transformer_blocks.0.norm3.bias", "input_blocks.8.1.proj_out.weight", "input_blocks.8.1.proj_out.bias", "input_blocks.9.0.op.weight", "input_blocks.9.0.op.bias", "input_blocks.10.0.in_layers.0.weight", "input_blocks.10.0.in_layers.0.bias", "input_blocks.10.0.in_layers.2.weight", "input_blocks.10.0.in_layers.2.bias", "input_blocks.10.0.emb_layers.1.weight", "input_blocks.10.0.emb_layers.1.bias", "input_blocks.10.0.out_layers.0.weight", "input_blocks.10.0.out_layers.0.bias", "input_blocks.10.0.out_layers.3.weight", "input_blocks.10.0.out_layers.3.bias", "input_blocks.11.0.in_layers.0.weight", "input_blocks.11.0.in_layers.0.bias", "input_blocks.11.0.in_layers.2.weight", "input_blocks.11.0.in_layers.2.bias", "input_blocks.11.0.emb_layers.1.weight", "input_blocks.11.0.emb_layers.1.bias", "input_blocks.11.0.out_layers.0.weight", "input_blocks.11.0.out_layers.0.bias", "input_blocks.11.0.out_layers.3.weight", "input_blocks.11.0.out_layers.3.bias", "zero_convs.0.0.weight", "zero_convs.0.0.bias", "zero_convs.1.0.weight", "zero_convs.1.0.bias", "zero_convs.2.0.weight", "zero_convs.2.0.bias", "zero_convs.3.0.weight", "zero_convs.3.0.bias", "zero_convs.4.0.weight", "zero_convs.4.0.bias", "zero_convs.5.0.weight", "zero_convs.5.0.bias", "zero_convs.6.0.weight", "zero_convs.6.0.bias", "zero_convs.7.0.weight", "zero_convs.7.0.bias", "zero_convs.8.0.weight", "zero_convs.8.0.bias", "zero_convs.9.0.weight", "zero_convs.9.0.bias", "zero_convs.10.0.weight", "zero_convs.10.0.bias", "zero_convs.11.0.weight", "zero_convs.11.0.bias", "input_hint_block.0.weight", "input_hint_block.0.bias", "input_hint_block.2.weight", "input_hint_block.2.bias", "input_hint_block.4.weight", "input_hint_block.4.bias", "input_hint_block.6.weight", "input_hint_block.6.bias", "input_hint_block.8.weight", "input_hint_block.8.bias", "input_hint_block.10.weight", "input_hint_block.10.bias", "input_hint_block.12.weight", "input_hint_block.12.bias", "input_hint_block.14.weight", "input_hint_block.14.bias", "middle_block.0.in_layers.0.weight", "middle_block.0.in_layers.0.bias", "middle_block.0.in_layers.2.weight", "middle_block.0.in_layers.2.bias", "middle_block.0.emb_layers.1.weight", "middle_block.0.emb_layers.1.bias", "middle_block.0.out_layers.0.weight", "middle_block.0.out_layers.0.bias", "middle_block.0.out_layers.3.weight", "middle_block.0.out_layers.3.bias", "middle_block.1.norm.weight", "middle_block.1.norm.bias", "middle_block.1.proj_in.weight", "middle_block.1.proj_in.bias", "middle_block.1.transformer_blocks.0.attn1.to_q.weight", "middle_block.1.transformer_blocks.0.attn1.to_k.weight", "middle_block.1.transformer_blocks.0.attn1.to_v.weight", "middle_block.1.transformer_blocks.0.attn1.to_out.0.weight", "middle_block.1.transformer_blocks.0.attn1.to_out.0.bias", "middle_block.1.transformer_blocks.0.ff.net.0.proj.weight", "middle_block.1.transformer_blocks.0.ff.net.0.proj.bias", "middle_block.1.transformer_blocks.0.ff.net.2.weight", "middle_block.1.transformer_blocks.0.ff.net.2.bias", "middle_block.1.transformer_blocks.0.attn2.to_q.weight", "middle_block.1.transformer_blocks.0.attn2.to_k.weight", "middle_block.1.transformer_blocks.0.attn2.to_v.weight", "middle_block.1.transformer_blocks.0.attn2.to_out.0.weight", "middle_block.1.transformer_blocks.0.attn2.to_out.0.bias", "middle_block.1.transformer_blocks.0.norm1.weight", "middle_block.1.transformer_blocks.0.norm1.bias", "middle_block.1.transformer_blocks.0.norm2.weight", "middle_block.1.transformer_blocks.0.norm2.bias", "middle_block.1.transformer_blocks.0.norm3.weight", "middle_block.1.transformer_blocks.0.norm3.bias", "middle_block.1.proj_out.weight", "middle_block.1.proj_out.bias", "middle_block.2.in_layers.0.weight", "middle_block.2.in_layers.0.bias", "middle_block.2.in_layers.2.weight", "middle_block.2.in_layers.2.bias", "middle_block.2.emb_layers.1.weight", "middle_block.2.emb_layers.1.bias", "middle_block.2.out_layers.0.weight", "middle_block.2.out_layers.0.bias", "middle_block.2.out_layers.3.weight", "middle_block.2.out_layers.3.bias", "middle_block_out.0.weight", "middle_block_out.0.bias". Unexpected key(s) in state_dict: "conv3_3.bias", "Mconv5_stage2_L2.bias", "conv5_3_CPM_L2.bias", "conv5_1_CPM_L2.bias", "Mconv3_stage3_L1.bias", "Mconv1_stage2_L1.weight", "Mconv2_stage5_L1.weight", "Mconv7_stage2_L1.bias", "Mconv3_stage2_L2.weight", "Mconv6_stage5_L1.weight", "conv3_1.weight", "Mconv1_stage2_L2.bias", "Mconv3_stage5_L1.weight", "Mconv1_stage3_L1.weight", "Mconv7_stage2_L2.weight", "Mconv5_stage3_L1.bias", "Mconv6_stage3_L2.bias", "Mconv4_stage6_L1.bias", "conv3_3.weight", "Mconv1_stage5_L2.weight", "Mconv2_stage4_L2.bias", "Mconv6_stage2_L2.weight", "Mconv2_stage2_L1.bias", "Mconv5_stage4_L1.bias", "Mconv5_stage2_L1.weight", "conv3_2.weight", "Mconv6_stage3_L1.bias", "conv5_2_CPM_L2.bias", "conv5_1_CPM_L2.weight", "Mconv1_stage5_L1.bias", "Mconv5_stage6_L2.bias", "Mconv2_stage4_L1.bias", "Mconv5_stage3_L2.bias", "conv5_5_CPM_L1.weight", "Mconv4_stage4_L1.weight", "Mconv5_stage4_L2.bias", "Mconv4_stage5_L1.bias", "Mconv3_stage5_L1.bias", "Mconv4_stage2_L1.bias", "Mconv1_stage5_L2.bias", "Mconv6_stage6_L2.bias", "Mconv5_stage6_L1.bias", "Mconv6_stage6_L1.weight", "Mconv7_stage3_L1.bias", "Mconv7_stage6_L1.bias", "Mconv6_stage6_L2.weight", "Mconv7_stage2_L1.weight", "Mconv6_stage3_L1.weight", "Mconv6_stage2_L1.bias", "Mconv6_stage2_L1.weight", "conv5_1_CPM_L1.weight", "Mconv5_stage6_L1.weight", "Mconv4_stage3_L1.bias", "conv5_2_CPM_L1.weight", "Mconv1_stage4_L2.weight", "Mconv2_stage2_L1.weight", "Mconv4_stage3_L2.weight", "conv4_2.weight", "conv2_2.bias", "Mconv6_stage3_L2.weight", "Mconv2_stage6_L1.bias", "conv1_2.bias", "Mconv3_stage2_L2.bias", "Mconv3_stage5_L2.weight", "Mconv7_stage5_L1.weight", "Mconv1_stage4_L2.bias", "Mconv3_stage3_L1.weight", "conv1_1.bias", "Mconv1_stage5_L1.weight", "Mconv4_stage2_L1.weight", "conv4_3_CPM.weight", "conv4_1.bias", "Mconv1_stage2_L1.bias", "conv2_2.weight", "conv4_3_CPM.bias", "Mconv2_stage3_L1.bias", "conv5_2_CPM_L2.weight", "conv5_5_CPM_L2.weight", "Mconv7_stage5_L2.bias", "Mconv3_stage3_L2.weight", "Mconv5_stage3_L1.weight", "Mconv2_stage5_L1.bias", "Mconv3_stage6_L2.bias", "Mconv1_stage3_L2.weight", "conv4_2.bias", "conv5_1_CPM_L1.bias", "Mconv6_stage4_L2.bias", "conv5_5_CPM_L1.bias", "Mconv5_stage4_L1.weight", "conv5_4_CPM_L2.bias", "Mconv6_stage2_L2.bias", "Mconv2_stage3_L1.weight", "Mconv6_stage4_L1.weight", "Mconv5_stage6_L2.weight", "Mconv3_stage4_L2.weight", "Mconv3_stage4_L2.bias", "Mconv3_stage6_L1.bias", "conv5_5_CPM_L2.bias", "Mconv7_stage6_L2.weight", "Mconv7_stage3_L1.weight", "Mconv6_stage5_L2.weight", "Mconv4_stage6_L2.bias", "Mconv7_stage5_L1.bias", "Mconv3_stage6_L2.weight", "Mconv1_stage6_L1.weight", "Mconv4_stage6_L2.weight", "Mconv5_stage2_L1.bias", "Mconv3_stage2_L1.weight", "Mconv3_stage3_L2.bias", "Mconv2_stage4_L1.weight", "Mconv6_stage6_L1.bias", "Mconv5_stage2_L2.weight", "Mconv4_stage3_L1.weight", "Mconv7_stage4_L2.weight", "Mconv4_stage4_L1.bias", "Mconv4_stage5_L2.bias", "conv4_4_CPM.weight", "Mconv2_stage4_L2.weight", "Mconv2_stage5_L2.weight", "Mconv7_stage6_L1.weight", "conv4_1.weight", "Mconv2_stage6_L2.weight", "conv2_1.bias", "Mconv6_stage5_L2.bias", "Mconv4_stage5_L1.weight", "Mconv2_stage3_L2.weight", "conv3_2.bias", "conv4_4_CPM.bias", "Mconv5_stage3_L2.weight", "Mconv3_stage4_L1.bias", "conv5_3_CPM_L1.bias", "Mconv5_stage5_L1.weight", "conv1_2.weight", "conv5_3_CPM_L1.weight", "Mconv4_stage4_L2.weight", "Mconv3_stage2_L1.bias", "Mconv3_stage4_L1.weight", "Mconv3_stage6_L1.weight", "conv5_2_CPM_L1.bias", "Mconv1_stage6_L1.bias", "Mconv2_stage3_L2.bias", "Mconv2_stage6_L1.weight", "Mconv7_stage4_L2.bias", "Mconv4_stage3_L2.bias", "conv3_1.bias", "Mconv2_stage2_L2.bias", "Mconv3_stage5_L2.bias", "Mconv4_stage2_L2.bias", "Mconv1_stage4_L1.bias", "Mconv4_stage6_L1.weight", "Mconv5_stage5_L2.weight", "Mconv6_stage5_L1.bias", "Mconv2_stage2_L2.weight", "Mconv4_stage2_L2.weight", "Mconv7_stage6_L2.bias", "Mconv1_stage6_L2.weight", "Mconv1_stage2_L2.weight", "Mconv1_stage4_L1.weight", "Mconv1_stage3_L1.bias", "conv5_4_CPM_L1.weight", "Mconv7_stage4_L1.bias", "Mconv6_stage4_L1.bias", "Mconv2_stage5_L2.bias", "conv3_4.weight", "conv3_4.bias", "Mconv5_stage5_L1.bias", "Mconv7_stage3_L2.weight", "Mconv1_stage6_L2.bias", "conv5_3_CPM_L2.weight", "Mconv5_stage4_L2.weight", "Mconv4_stage4_L2.bias", "Mconv7_stage4_L1.weight", "conv5_4_CPM_L1.bias", "conv5_4_CPM_L2.weight", "conv1_1.weight", "Mconv7_stage2_L2.bias", "Mconv7_stage3_L2.bias", "conv2_1.weight", "Mconv1_stage3_L2.bias", "Mconv2_stage6_L2.bias", "Mconv4_stage5_L2.weight", "Mconv5_stage5_L2.bias", "Mconv7_stage5_L2.weight", "Mconv6_stage4_L2.weight".

SDv2 and (V-Parameterization) support

Currently when using an SD2 model an error is received. An error is received on a normal working SD V2 model, and a V-Parameterization model (SD v2.1 768)

RuntimeError: mat1 and mat2 shapes cannot be multiplied (154x1024 and 768x320)

Full error log attached.
error.log

traceback error on startup

โ”€ Traceback (most recent call last) โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฎ
โ”‚ D:\stable-diffusion-webui\launch.py:361 in โ”‚
โ”‚ โ”‚
โ”‚ 358 โ”‚
โ”‚ 359 if name == "main": โ”‚
โ”‚ 360 โ”‚ prepare_environment() โ”‚
โ”‚ โฑ 361 โ”‚ start() โ”‚
โ”‚ 362 โ”‚
โ”‚ โ”‚
โ”‚ D:\stable-diffusion-webui\launch.py:356 in start โ”‚
โ”‚ โ”‚
โ”‚ 353 โ”‚ if '--nowebui' in sys.argv: โ”‚
โ”‚ 354 โ”‚ โ”‚ webui.api_only() โ”‚
โ”‚ 355 โ”‚ else: โ”‚
โ”‚ โฑ 356 โ”‚ โ”‚ webui.webui() โ”‚
โ”‚ 357 โ”‚
โ”‚ 358 โ”‚
โ”‚ 359 if name == "main": โ”‚
โ”‚ โ”‚
โ”‚ D:\stable-diffusion-webui\webui.py:205 in webui โ”‚
โ”‚ โ”‚
โ”‚ 202 โ”‚ โ”‚ โ”‚
โ”‚ 203 โ”‚ โ”‚ modules.script_callbacks.before_ui_callback() โ”‚
โ”‚ 204 โ”‚ โ”‚ โ”‚
โ”‚ โฑ 205 โ”‚ โ”‚ shared.demo = modules.ui.create_ui() โ”‚
โ”‚ 206 โ”‚ โ”‚ โ”‚
โ”‚ 207 โ”‚ โ”‚ if cmd_opts.gradio_queue: โ”‚
โ”‚ 208 โ”‚ โ”‚ โ”‚ shared.demo.queue(64) โ”‚
โ”‚ โ”‚
โ”‚ D:\stable-diffusion-webui\modules\ui.py:458 in create_ui โ”‚
โ”‚ โ”‚
โ”‚ 455 โ”‚ parameters_copypaste.reset() โ”‚
โ”‚ 456 โ”‚ โ”‚
โ”‚ 457 โ”‚ modules.scripts.scripts_current = modules.scripts.scripts_txt2img โ”‚
โ”‚ โฑ 458 โ”‚ modules.scripts.scripts_txt2img.initialize_scripts(is_img2img=False) โ”‚
โ”‚ 459 โ”‚ โ”‚
โ”‚ 460 โ”‚ with gr.Blocks(analytics_enabled=False) as txt2img_interface: โ”‚
โ”‚ 461 โ”‚ โ”‚ txt2img_prompt, txt2img_prompt_styles, txt2img_negative_prompt, submit, _, , tx โ”‚
โ”‚ โ”‚
โ”‚ D:\stable-diffusion-webui\modules\scripts.py:270 in initialize_scripts โ”‚
โ”‚ โ”‚
โ”‚ 267 โ”‚ โ”‚ auto_processing_scripts = scripts_auto_postprocessing.create_auto_preprocessing
โ”‚
โ”‚ 268 โ”‚ โ”‚ โ”‚
โ”‚ 269 โ”‚ โ”‚ for script_class, path, basedir, script_module in auto_processing_scripts + scri โ”‚
โ”‚ โฑ 270 โ”‚ โ”‚ โ”‚ script = script_class() โ”‚
โ”‚ 271 โ”‚ โ”‚ โ”‚ script.filename = path โ”‚
โ”‚ 272 โ”‚ โ”‚ โ”‚ script.is_txt2img = not is_img2img โ”‚
โ”‚ 273 โ”‚ โ”‚ โ”‚ script.is_img2img = is_img2img โ”‚
โ”‚ โ”‚
โ”‚ D:\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py:115 in init โ”‚
โ”‚ โ”‚
โ”‚ 112 โ”‚ โ”‚ โ”‚ "depth": midas, โ”‚
โ”‚ 113 โ”‚ โ”‚ โ”‚ "hed": hed, โ”‚
โ”‚ 114 โ”‚ โ”‚ โ”‚ "mlsd": mlsd, โ”‚
โ”‚ โฑ 115 โ”‚ โ”‚ โ”‚ "normal_map": midas_normal, โ”‚
โ”‚ 116 โ”‚ โ”‚ โ”‚ "openpose": openpose, โ”‚
โ”‚ 117 โ”‚ โ”‚ โ”‚ "openpose_hand": openpose_hand, โ”‚
โ”‚ 118 โ”‚ โ”‚ โ”‚ "fake_scribble": fake_scribble, โ”‚
โ•ฐโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ•ฏ
NameError: name 'midas_normal' is not defined

The error occurs when Low VRAM (8GB or below) is enabled.

The error occurs when Low VRAM (8GB or below) is enabled. However, the image creation itself is normally performed.

image

`ERROR: Exception in ASGI application | 2/20 [00:00<00:05, 3.38it/s]
Traceback (most recent call last):
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio\streams\memory.py", line 94, in receive
return self.receive_nowait()
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio\streams\memory.py", line 89, in receive_nowait
raise WouldBlock
anyio.WouldBlock

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\base.py", line 77, in call_next
message = await recv_stream.receive()
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio\streams\memory.py", line 114, in receive
raise EndOfStream
anyio.EndOfStream

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "C:\stable-diffusion-webui\venv\lib\site-packages\uvicorn\protocols\http\h11_impl.py", line 407, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "C:\stable-diffusion-webui\venv\lib\site-packages\uvicorn\middleware\proxy_headers.py", line 78, in call
return await self.app(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\fastapi\applications.py", line 271, in call
await super().call(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\applications.py", line 125, in call
await self.middleware_stack(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\errors.py", line 184, in call
raise exc
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\errors.py", line 162, in call
await self.app(scope, receive, _send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\base.py", line 104, in call
response = await self.dispatch_func(request, call_next)
File "C:\stable-diffusion-webui\modules\api\api.py", line 96, in log_and_time
res: Response = await call_next(req)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\base.py", line 80, in call_next
raise app_exc
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\base.py", line 69, in coro
await self.app(scope, receive_or_disconnect, send_no_error)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\gzip.py", line 24, in call
await responder(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\gzip.py", line 44, in call
await self.app(scope, receive, self.send_with_gzip)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\exceptions.py", line 79, in call
raise exc
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\middleware\exceptions.py", line 68, in call
await self.app(scope, receive, sender)
File "C:\stable-diffusion-webui\venv\lib\site-packages\fastapi\middleware\asyncexitstack.py", line 21, in call
raise e
File "C:\stable-diffusion-webui\venv\lib\site-packages\fastapi\middleware\asyncexitstack.py", line 18, in call
await self.app(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\routing.py", line 706, in call
await route.handle(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\routing.py", line 276, in handle
await self.app(scope, receive, send)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\routing.py", line 66, in app
response = await func(request)
File "C:\stable-diffusion-webui\venv\lib\site-packages\fastapi\routing.py", line 237, in app
raw_response = await run_endpoint_function(
File "C:\stable-diffusion-webui\venv\lib\site-packages\fastapi\routing.py", line 165, in run_endpoint_function
return await run_in_threadpool(dependant.call, **values)
File "C:\stable-diffusion-webui\venv\lib\site-packages\starlette\concurrency.py", line 41, in run_in_threadpool
return await anyio.to_thread.run_sync(func, *args)
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio_backends_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "C:\stable-diffusion-webui\venv\lib\site-packages\anyio_backends_asyncio.py", line 867, in run
result = context.run(func, *args)
File "C:\stable-diffusion-webui\modules\progress.py", line 85, in progressapi
shared.state.set_current_image()
File "C:\stable-diffusion-webui\modules\shared.py", line 243, in set_current_image
self.do_set_current_image()
File "C:\stable-diffusion-webui\modules\shared.py", line 251, in do_set_current_image
self.assign_current_image(modules.sd_samplers.samples_to_image_grid(self.current_latent))
File "C:\stable-diffusion-webui\modules\sd_samplers_common.py", line 50, in samples_to_image_grid
return images.image_grid([single_sample_to_image(sample, approximation) for sample in samples])
File "C:\stable-diffusion-webui\modules\sd_samplers_common.py", line 50, in
return images.image_grid([single_sample_to_image(sample, approximation) for sample in samples])
File "C:\stable-diffusion-webui\modules\sd_samplers_common.py", line 37, in single_sample_to_image
x_sample = processing.decode_first_stage(shared.sd_model, sample.unsqueeze(0))[0]
File "C:\stable-diffusion-webui\modules\processing.py", line 423, in decode_first_stage
x = model.decode_first_stage(x)
File "C:\stable-diffusion-webui\modules\sd_hijack_utils.py", line 17, in
setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs))
File "C:\stable-diffusion-webui\modules\sd_hijack_utils.py", line 28, in call
return self.__orig_func(*args, **kwargs)
File "C:\stable-diffusion-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "C:\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 826, in decode_first_stage
return self.first_stage_model.decode(z)
File "C:\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\autoencoder.py", line 89, in decode
z = self.post_quant_conv(z)
File "C:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "C:\stable-diffusion-webui\extensions-builtin\Lora\lora.py", line 182, in lora_Conv2d_forward
return lora_forward(self, input, torch.nn.Conv2d_forward_before_lora(self, input))
File "C:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\conv.py", line 463, in forward
return self._conv_forward(input, self.weight, self.bias)
File "C:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\conv.py", line 459, in _conv_forward
return F.conv2d(input, weight, bias, self.stride,
RuntimeError: Input type (torch.cuda.FloatTensor) and weight type (torch.FloatTensor) should be the same`

Scribble Mode (Reverse color) should be automatic

As the title described. If the user used open canvas and paint, inverse color should be automatically enabled to avoid bad generation quality and subsequent confusions.

Example:
image
^Enabled

image
^Disabled

RuntimeError: Error(s) in loading state_dict for ControlNet

Help! what the wrong it is...

Loaded state_dict from [E:\AI\stable-diffusion-webui\extensions\sd-webui-controlnet\models\Anything3.ckpt]
Error running process: E:\AI\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py
Traceback (most recent call last):
File "E:\AI\stable-diffusion-webui\modules\scripts.py", line 386, in process
script.process(p, *script_args)
File "E:\AI\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py", line 247, in process
network = PlugableControlModel(model_path, os.path.join(cn_models_dir, "cldm_v15.yaml"), weight, lowvram=lowvram)
File "E:\AI\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\cldm.py", line 49, in init
self.control_model.load_state_dict(state_dict)
File "E:\AI\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1671, in load_state_dict
raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format(
RuntimeError: Error(s) in loading state_dict for ControlNet:

combine control net with ultimate sd upscale

would it be possible to combine controlnet with this extention?
https://github.com/Coyote-A/ultimate-upscale-for-automatic1111 (or would that developer have to integrate the functionality?)

currently it copies the feature-image for every tile that is used in the upscale and thus copies the image over itself. The improved detail coherency from canary or hde might work wonders to allow regaining detailes (especially thinking about hands etc.)

EDIT:
just found the thread talking about general upscaling the mask images. I do think the combination with ultimate upscale would be preferable to general upscaling though, because the individual tile-sizes stay in 512-768px range and thus should work better with the trained models from controlnet, hopefully XP

An error occurs in img2img when using the Controlnet extension.

An error occurs in img2img when using the Controlnet extension. There is no problem if remove the Controlnet extension.

Error running process: C:\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py
Traceback (most recent call last):
File "C:\stable-diffusion-webui\modules\scripts.py", line 386, in process
script.process(p, *script_args)
File "C:\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py", line 237, in process
raise RuntimeError(f"model not found: {model}")
RuntimeError: model not found: 0.5

~

Error running process: C:\stable-diffusion-webui\extensions\stable-diffusion-webui-daam\scripts\daam_script.py
Traceback (most recent call last):
File "C:\stable-diffusion-webui\modules\scripts.py", line 386, in process
script.process(p, *script_args)
File "C:\stable-diffusion-webui\extensions\stable-diffusion-webui-daam\scripts\daam_script.py", line 99, in process
self.attentions = [s.strip() for s in attention_texts.split(",") if s.strip()]
AttributeError: 'bool' object has no attribute 'split'

0%| | 0/16 [00:00<?, ?it/s]ssii_intermediate_type, ssii_every_n, ssii_start_at_n, ssii_stop_at_n, ssii_video, ssii_video_format, ssii_mp4_parms, ssii_video_fps, ssii_add_first_frames, ssii_add_last_frames, ssii_smooth, ssii_seconds, ssii_lores, ssii_hires, ssii_debug:0.0, 0.0, False, 0.0, True, True, False, , False, False, False, False, Auto, 0.5, 1
Step, abs_step, hr, hr_active: 0, 0, False, False
0%| | 0/16 [00:00<?, ?it/s]
Error completing request
Arguments: ('task(dew37470n9yuaad)', 0, 'masterpiece, best quality, 1girl, solo, white swimsut, beach, blonde hair, looking at viewer, jumping, blue sky, white cloud, sun, lensflare, bubbles', '(worst quality, low quality:1.4), cap, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, mutation, missing arms, missing legs, extra arms, extra legs, blurry, mutated hands, fused fingers, too many fingers, extra fingers, extra others, futanari, fused penis, missing penis, extra penis, mutated penis, username, mask, furry, odd eyes, artist name', [], <PIL.Image.Image image mode=RGBA size=512x768 at 0x17C388D59F0>, None, None, None, None, None, None, 20, 15, 4, 0, 0, False, False, 1, 1, 8.5, 1.5, 0.75, 3991550646.0, -1.0, 0, 0, 0, False, 768, 512, 0, 0, 32, 0, '', '', '', [], 0, 0, 0, 0, 0, 0.25, False, 7, 100, 'Constant', 0, 'Constant', 0, 4, False, False, 'LoRA', 'None', 0, 0, 'LoRA', 'None', 0, 0, 'LoRA', 'None', 0, 0, 'LoRA', 'None', 0, 0, 'LoRA', 'None', 0, 0, 'Refresh models', 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, 0.5, [], False, True, 'Denoised', 1.0, 0.0, 0.0, True, 'gif', 'h264', 2.0, 0.0, 0.0, False, 0.0, True, True, False, '', False, False, False, False, 'Auto', 0.5, 1, False, False, 1, False, '

    \n
  • CFG Scale should be 2 or lower.
  • \n
\n', True, True, '', '', True, 50, True, 1, 0, False, 1, 0, '#000000', True, False, 0, 256, 0, None, '', 0.2, 0.1, 1, 1, False, True, True, False, False, False, False, 4, 1, 4, 0.09, True, 1, 0, 7, False, False, 'Show/Hide AlphaCanvas', 384, 'Update Outpainting Size', 8, '

Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8

', 128, 8, ['left', 'right', 'up', 'down'], 1, 0.05, 128, 4, 0, ['left', 'right', 'up', 'down'], False, False, 'positive', 'comma', 0, False, False, '', '

Will upscale the image by the selected scale factor; use width and height sliders to set tile size

', 64, 0, 2, 'Positive', 0, ', ', True, 32, 1, '', 0, '', 0, '', True, False, False, False, 0, False, None, True, None, None, False, True, True, True, 0, 0, 384, 384, False, False, True, True, True, False, True, 1, False, False, 2.5, 4, 0, False, 0, 1, False, False, 'u2net', False, False, False, False, 0, 1, 384, 384, True, False, True, True, True, False, 1, True, 3, False, 3, False, 3, 1, '

Will upscale the image depending on the selected target size type

', 512, 0, 8, 32, 64, 0.35, 32, 0, True, 2, False, 8, 0, 0, 2048, 2048, 2) {}
Traceback (most recent call last):
File "C:\stable-diffusion-webui\modules\call_queue.py", line 56, in f
res = list(func(*args, **kwargs))
File "C:\stable-diffusion-webui\modules\call_queue.py", line 37, in f
res = func(*args, **kwargs)
File "C:\stable-diffusion-webui\modules\img2img.py", line 169, in img2img
processed = process_images(p)
File "C:\stable-diffusion-webui\modules\processing.py", line 486, in process_images
res = process_images_inner(p)
File "C:\stable-diffusion-webui\modules\processing.py", line 628, in process_images_inner
samples_ddim = p.sample(conditioning=c, unconditional_conditioning=uc, seeds=seeds, subseeds=subseeds, subseed_strength=p.subseed_strength, prompts=prompts)
File "C:\stable-diffusion-webui\modules\processing.py", line 1044, in sample
samples = self.sampler.sample_img2img(self, self.init_latent, x, conditioning, unconditional_conditioning, image_conditioning=self.image_conditioning)
File "C:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 302, in sample_img2img
samples = self.launch_sampling(t_enc + 1, lambda: self.func(self.model_wrap_cfg, xi, extra_args=extra_args, disable=False, callback=self.callback_state, **extra_params_kwargs))
File "C:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 221, in launch_sampling
return func()
File "C:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 302, in
samples = self.launch_sampling(t_enc + 1, lambda: self.func(self.model_wrap_cfg, xi, extra_args=extra_args, disable=False, callback=self.callback_state, **extra_params_kwargs))
File "C:\stable-diffusion-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "C:\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\sampling.py", line 596, in sample_dpmpp_2m
callback({'x': x, 'i': i, 'sigma': sigmas[i], 'sigma_hat': sigmas[i], 'denoised': denoised})
File "C:\stable-diffusion-webui\extensions\sd_save_intermediate_images\scripts\sd_save_intermediate_images.py", line 499, in callback_state
if ssii_start_at_n % ssii_every_n == 0:
ZeroDivisionError: float modulo

[Feature Request] Possibility of more than one ControlNet input?

This one may be iffy, but based on what I read, it seems like it might be possible to stack more than one ControlNet onto Stable Diffusion, at the same time. If that's possible, it would be really interesting to use that. Being able to define both depth and normals when generating images of a building, or depth + pose for characters, would allow a lot of control.

RuntimeError: Input type (torch.cuda.HalfTensor) and weight type (torch.HalfTensor) should be the same

ERROR: Exception in ASGI applicationโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–Œ | 15/30 [00:13<00:14, 1.02it/s]
Traceback (most recent call last):
File "G:\stable-webui\venv\lib\site-packages\anyio\streams\memory.py", line 94, in receive
return self.receive_nowait()
File "G:\stable-webui\venv\lib\site-packages\anyio\streams\memory.py", line 89, in receive_nowait
raise WouldBlock
anyio.WouldBlock

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\base.py", line 77, in call_next
message = await recv_stream.receive()
File "G:\stable-webui\venv\lib\site-packages\anyio\streams\memory.py", line 114, in receive
raise EndOfStream
anyio.EndOfStream

During handling of the above exception, another exception occurred:

Traceback (most recent call last):
File "G:\stable-webui\venv\lib\site-packages\uvicorn\protocols\http\h11_impl.py", line 407, in run_asgi
result = await app( # type: ignore[func-returns-value]
File "G:\stable-webui\venv\lib\site-packages\uvicorn\middleware\proxy_headers.py", line 78, in call
return await self.app(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\fastapi\applications.py", line 271, in call
await super().call(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\applications.py", line 125, in call
await self.middleware_stack(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\errors.py", line 184, in call
raise exc
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\errors.py", line 162, in call
await self.app(scope, receive, _send)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\base.py", line 106, in call
response = await self.dispatch_func(request, call_next)
File "G:\stable-webui\modules\api\api.py", line 96, in log_and_time
res: Response = await call_next(req)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\base.py", line 80, in call_next
raise app_exc
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\base.py", line 69, in coro
await self.app(scope, receive_or_disconnect, send_no_error)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\gzip.py", line 24, in call
await responder(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\gzip.py", line 44, in call
await self.app(scope, receive, self.send_with_gzip)
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\exceptions.py", line 79, in call
raise exc
File "G:\stable-webui\venv\lib\site-packages\starlette\middleware\exceptions.py", line 68, in call
await self.app(scope, receive, sender)
File "G:\stable-webui\venv\lib\site-packages\fastapi\middleware\asyncexitstack.py", line 21, in call
raise e
File "G:\stable-webui\venv\lib\site-packages\fastapi\middleware\asyncexitstack.py", line 18, in call
await self.app(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\routing.py", line 706, in call
await route.handle(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\routing.py", line 276, in handle
await self.app(scope, receive, send)
File "G:\stable-webui\venv\lib\site-packages\starlette\routing.py", line 66, in app
response = await func(request)
File "G:\stable-webui\venv\lib\site-packages\fastapi\routing.py", line 237, in app
raw_response = await run_endpoint_function(
File "G:\stable-webui\venv\lib\site-packages\fastapi\routing.py", line 165, in run_endpoint_function
return await run_in_threadpool(dependant.call, **values)
File "G:\stable-webui\venv\lib\site-packages\starlette\concurrency.py", line 41, in run_in_threadpool
return await anyio.to_thread.run_sync(func, *args)
File "G:\stable-webui\venv\lib\site-packages\anyio\to_thread.py", line 31, in run_sync
return await get_asynclib().run_sync_in_worker_thread(
File "G:\stable-webui\venv\lib\site-packages\anyio_backends_asyncio.py", line 937, in run_sync_in_worker_thread
return await future
File "G:\stable-webui\venv\lib\site-packages\anyio_backends_asyncio.py", line 867, in run
result = context.run(func, *args)
File "G:\stable-webui\modules\progress.py", line 85, in progressapi
shared.state.set_current_image()
File "G:\stable-webui\modules\shared.py", line 243, in set_current_image
self.do_set_current_image()
File "G:\stable-webui\modules\shared.py", line 251, in do_set_current_image
self.assign_current_image(modules.sd_samplers.samples_to_image_grid(self.current_latent))
File "G:\stable-webui\modules\sd_samplers_common.py", line 50, in samples_to_image_grid
return images.image_grid([single_sample_to_image(sample, approximation) for sample in samples])
File "G:\stable-webui\modules\sd_samplers_common.py", line 50, in
return images.image_grid([single_sample_to_image(sample, approximation) for sample in samples])
File "G:\stable-webui\modules\sd_samplers_common.py", line 37, in single_sample_to_image
x_sample = processing.decode_first_stage(shared.sd_model, sample.unsqueeze(0))[0]
File "G:\stable-webui\modules\processing.py", line 423, in decode_first_stage
x = model.decode_first_stage(x)
File "G:\stable-webui\modules\sd_hijack_utils.py", line 17, in
setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs))
File "G:\stable-webui\modules\sd_hijack_utils.py", line 28, in call
return self.__orig_func(*args, **kwargs)
File "G:\stable-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "G:\stable-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 826, in decode_first_stage
return self.first_stage_model.decode(z)
File "G:\stable-webui\repositories\stable-diffusion-stability-ai\ldm\models\autoencoder.py", line 89, in decode
z = self.post_quant_conv(z)
File "G:\stable-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1194, in _call_impl
return forward_call(*input, **kwargs)
File "G:\stable-webui\extensions-builtin\Lora\lora.py", line 182, in lora_Conv2d_forward
return lora_forward(self, input, torch.nn.Conv2d_forward_before_lora(self, input))
File "G:\stable-webui\venv\lib\site-packages\torch\nn\modules\conv.py", line 463, in forward
return self._conv_forward(input, self.weight, self.bias)
File "G:\stable-webui\venv\lib\site-packages\torch\nn\modules\conv.py", line 459, in _conv_forward
return F.conv2d(input, weight, bias, self.stride,
RuntimeError: Input type (torch.cuda.HalfTensor) and weight type (torch.HalfTensor) should be the same

[Request]Add a checkbox for not using facial bones of openpose

There is an issue for open pose. When source image and target image are not the same style, for example. source image is a 2D cartoon one, target image is a realistic photo, facial bones of openpose gonna always make the face terrible, even after adjust the weight option.

So, the best solution when source image and target image are different styles, is just offer a choice to not use the facial bones of openpose.

They are: eye, nose and ear bones.

Sliders limited to 1024.

I attempted to directly resolve this myself by modifying the slider values in the py from 1024 to 4096, however it is still limited to 1024.
Almost every image I work with has a side well beyond 1024, normally more than 3x that. However even when the max for the sliders is raised to 4096, the sliders themselves are locked at 1024 max values.

Segmentation preprocessor is missing prettytable package

When trying the segmentation preprocessor, I had a Python error trying to import from the prettytable package.

I don't know if this error was specific to my system, but it was easily corrected with :

venv\Script\activate.bat
pip install prettytable

To be safe, the package should be added to the requirements for auto-install.

Error when trying to update via a1

When I try to update the repo, it gives out this:

File "D:\AI_WORKPLACE\AUTOMATIC1111\current\stable-diffusion-webui\modules\scripts.py", line 270, in initialize_scripts script = script_class() File "D:\AI_WORKPLACE\AUTOMATIC1111\current\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\controlnet.py", line 118, in __init__ "normal_map": midas_normal, NameError: name 'midas_normal' is not defined

RuntimeError: mat1 and mat2 shapes cannot be multiplied (616x1024 and 768x320)

Thanks for this, I hope I can get it to work soon!

I tried the Scribble Model, but I got this error once I ran it:
RuntimeError: mat1 and mat2 shapes cannot be multiplied (616x1024 and 768x320)

I resized the input image to 512 x 704 pixels and set the normal Width & Height accordingly. I tried running it without any Preprocessor or with the Fake_Scribble processor, but got the same error message both times. Weight left at 1, Scribble Mode On & Off tried.

4GB gpu works?

Hey mikubill, I'm trying to understand why a1111 requires so little gpu. The model itself is 5.7G and a simple read is already OOM. How is 4GB optimized?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.