Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature Request] Manually choosing the inpainting unet #23

Open
MoonMoon82 opened this issue Sep 18, 2024 · 4 comments
Open

[Feature Request] Manually choosing the inpainting unet #23

MoonMoon82 opened this issue Sep 18, 2024 · 4 comments

Comments

@MoonMoon82
Copy link

Hi! At first, this is an awesome custom node!! Love it!

Is there a way or could you implement an option to manually choose a different inpainting unet file?
If i just replace the diffusion_pytorch_model.safetensors with my own inpainting SD1.5 unet this error shows up:

Traceback (most recent call last):
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
    process_inputs(input_dict, i)
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_CatVTON_Wrapper\py\cat_vton.py", line 42, in catvton
    pipeline = CatVTONPipeline(
               ^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI_CatVTON_Wrapper\py\catvton\pipeline.py", line 43, in __init__
    self.unet = UNet2DConditionModel.from_pretrained(base_ckpt, subfolder="unet").to(device, dtype=weight_dtype)
                ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\python_embeded\Lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "E:\StableDiffusion\ComfyUI_windows_portable\python_embeded\Lib\site-packages\diffusers\models\modeling_utils.py", line 750, in from_pretrained
    raise ValueError(
ValueError: Cannot load <class 'diffusers.models.unets.unet_2d_condition.UNet2DConditionModel'> from E:\StableDiffusion\ComfyUI_windows_portable\ComfyUI\models\CatVTON\stable-diffusion-inpainting because the following keys are missing:
 up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.3.resnets.0.norm2.bias, up_blocks.1.resnets.1.conv1.weight, mid_block.resnets.0.time_emb_proj.bias, up_blocks.1.attentions.0.proj_out.bias, up_blocks.0.resnets.2.norm2.weight, mid_block.resnets.0.conv2.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm2.bias, down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.2.attentions.0.proj_out.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias, up_blocks.0.resnets.2.conv_shortcut.bias, down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight, mid_block.attentions.0.norm.weight, up_blocks.2.resnets.0.norm2.bias, up_blocks.2.attentions.2.norm.weight, down_blocks.2.attentions.1.proj_out.weight, up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj.weight, conv_out.weight, down_blocks.2.attentions.1.transformer_blocks.0.norm3.weight, mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.3.resnets.0.norm1.bias, up_blocks.3.attentions.1.transformer_blocks.0.norm3.weight, up_blocks.1.attentions.1.proj_in.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_k.weight, down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm2.weight, up_blocks.0.resnets.0.norm2.bias, up_blocks.3.resnets.2.conv2.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.bias, down_blocks.2.resnets.1.norm2.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.3.resnets.0.conv1.bias, up_blocks.3.resnets.1.conv2.weight, up_blocks.0.resnets.0.conv_shortcut.bias, down_blocks.3.resnets.0.time_emb_proj.bias, down_blocks.1.attentions.1.proj_out.weight, down_blocks.2.resnets.1.conv1.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm1.bias, down_blocks.3.resnets.0.conv2.bias, down_blocks.0.resnets.1.time_emb_proj.bias, down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.0.resnets.0.norm1.weight, down_blocks.3.resnets.1.norm2.bias, down_blocks.0.resnets.0.norm2.weight, down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.bias, up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight, down_blocks.3.resnets.1.time_emb_proj.bias, up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight, down_blocks.2.resnets.1.conv2.bias, up_blocks.3.attentions.0.transformer_blocks.0.norm2.bias, up_blocks.3.attentions.2.transformer_blocks.0.norm1.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm3.weight, down_blocks.3.resnets.1.time_emb_proj.weight, up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_v.weight, down_blocks.2.attentions.1.transformer_blocks.0.norm3.bias, time_embedding.linear_1.weight, down_blocks.2.attentions.0.norm.bias, down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight, down_blocks.0.attentions.0.proj_out.bias, mid_block.resnets.1.conv1.bias, up_blocks.1.attentions.1.proj_out.bias, up_blocks.0.resnets.1.norm1.weight, down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.weight, up_blocks.0.upsamplers.0.conv.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, mid_block.attentions.0.proj_out.weight, up_blocks.2.resnets.0.norm1.bias, up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight, up_blocks.1.attentions.1.transformer_blocks.0.norm2.bias, up_blocks.3.resnets.1.norm1.bias, up_blocks.3.resnets.0.conv2.weight, up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.0.downsamplers.0.conv.bias, down_blocks.0.resnets.1.conv1.bias, up_blocks.2.attentions.0.transformer_blocks.0.norm3.weight, down_blocks.0.resnets.0.conv1.bias, down_blocks.1.attentions.0.transformer_blocks.0.norm2.weight, mid_block.attentions.0.proj_in.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_v.weight, down_blocks.1.resnets.1.time_emb_proj.bias, up_blocks.1.resnets.1.conv_shortcut.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm2.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_v.weight, up_blocks.3.resnets.2.conv_shortcut.bias, up_blocks.0.resnets.1.conv_shortcut.bias, up_blocks.3.resnets.0.conv1.weight, up_blocks.1.resnets.0.conv_shortcut.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight, mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.2.resnets.2.time_emb_proj.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, mid_block.attentions.0.transformer_blocks.0.attn1.to_q.weight, mid_block.resnets.0.conv1.weight, up_blocks.3.resnets.2.conv2.bias, up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.0.resnets.2.norm1.bias, up_blocks.3.attentions.0.transformer_blocks.0.norm1.weight, down_blocks.2.downsamplers.0.conv.weight, down_blocks.2.resnets.0.conv_shortcut.weight, up_blocks.2.resnets.1.conv1.weight, up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.resnets.1.time_emb_proj.weight, down_blocks.0.attentions.0.proj_in.bias, down_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.2.attentions.0.transformer_blocks.0.norm1.bias, up_blocks.3.resnets.2.norm1.bias, up_blocks.1.resnets.0.conv2.weight, up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight, mid_block.attentions.0.proj_out.bias, up_blocks.3.attentions.2.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.1.attentions.1.transformer_blocks.0.norm3.bias, down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_q.weight, down_blocks.2.attentions.1.proj_out.bias, up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.bias, down_blocks.2.resnets.0.time_emb_proj.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_v.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight, up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2.bias, up_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.weight, up_blocks.3.attentions.2.norm.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_q.weight, down_blocks.2.resnets.0.conv1.bias, down_blocks.3.resnets.1.conv1.weight, up_blocks.2.attentions.2.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.2.resnets.0.conv1.bias, up_blocks.2.attentions.2.proj_in.bias, down_blocks.1.resnets.1.norm2.weight, up_blocks.2.resnets.1.conv2.weight, up_blocks.3.attentions.2.proj_out.bias, up_blocks.1.resnets.2.norm2.weight, up_blocks.3.attentions.0.transformer_blocks.0.ff.net.2.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, down_blocks.1.resnets.0.norm2.bias, up_blocks.1.resnets.2.time_emb_proj.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, down_blocks.2.attentions.0.proj_in.weight, up_blocks.2.attentions.2.transformer_blocks.0.norm1.weight, up_blocks.2.attentions.2.transformer_blocks.0.norm2.weight, down_blocks.1.resnets.0.conv1.weight, down_blocks.1.attentions.1.norm.weight, up_blocks.1.attentions.2.norm.bias, up_blocks.1.upsamplers.0.conv.weight, up_blocks.1.resnets.1.conv2.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, mid_block.resnets.1.norm2.weight, down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.1.attentions.0.transformer_blocks.0.norm2.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_q.weight, up_blocks.2.resnets.0.time_emb_proj.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, mid_block.resnets.1.norm1.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight, up_blocks.3.attentions.2.transformer_blocks.0.norm1.weight, up_blocks.1.attentions.2.proj_out.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_k.weight, down_blocks.1.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.1.resnets.1.norm2.bias, up_blocks.2.attentions.0.proj_in.weight, down_blocks.3.resnets.1.conv2.weight, up_blocks.0.resnets.2.norm1.weight, up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.3.resnets.1.norm2.weight, time_embedding.linear_2.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.1.attentions.2.proj_in.weight, up_blocks.3.resnets.1.time_emb_proj.weight, down_blocks.0.attentions.1.proj_out.bias, down_blocks.2.resnets.1.norm2.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_q.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight, up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_q.weight, up_blocks.3.attentions.1.transformer_blocks.0.norm1.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight, down_blocks.2.resnets.0.conv2.weight, up_blocks.2.resnets.1.norm2.weight, down_blocks.0.resnets.0.conv1.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight, up_blocks.1.resnets.1.norm1.weight, up_blocks.2.resnets.0.conv_shortcut.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight, up_blocks.3.resnets.0.conv1.bias, up_blocks.0.resnets.1.time_emb_proj.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, down_blocks.1.attentions.1.proj_out.bias, up_blocks.0.resnets.0.conv_shortcut.weight, mid_block.resnets.1.norm1.weight, down_blocks.0.attentions.1.proj_out.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.1.attentions.1.norm.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm3.weight, down_blocks.1.attentions.0.transformer_blocks.0.norm3.bias, up_blocks.3.attentions.0.proj_in.weight, up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.3.attentions.2.transformer_blocks.0.norm2.bias, up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.bias, mid_block.attentions.0.transformer_blocks.0.norm1.bias, down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.bias, down_blocks.2.resnets.0.conv1.weight, down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.1.resnets.0.norm2.weight, up_blocks.2.resnets.0.conv2.weight, up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_q.weight, up_blocks.2.resnets.0.time_emb_proj.bias, up_blocks.1.attentions.0.proj_in.weight, up_blocks.1.resnets.1.conv_shortcut.weight, up_blocks.3.resnets.1.norm1.weight, up_blocks.3.resnets.1.conv_shortcut.weight, up_blocks.2.resnets.2.conv1.bias, down_blocks.0.resnets.1.conv2.bias, down_blocks.2.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.2.attentions.1.transformer_blocks.0.norm2.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight, up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.3.attentions.1.proj_out.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm1.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_k.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.attentions.0.transformer_blocks.0.norm1.weight, up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_q.weight, down_blocks.3.resnets.0.conv2.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight, up_blocks.2.resnets.2.norm2.bias, up_blocks.3.resnets.0.norm1.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.resnets.2.conv2.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_k.weight, mid_block.resnets.0.norm2.bias, up_blocks.2.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.3.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.2.resnets.1.norm1.weight, down_blocks.1.resnets.0.conv2.bias, down_blocks.1.attentions.1.proj_in.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.2.attentions.1.norm.weight, down_blocks.1.resnets.0.norm1.weight, up_blocks.1.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.0.resnets.2.time_emb_proj.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight, up_blocks.2.resnets.2.norm1.bias, up_blocks.0.resnets.2.time_emb_proj.bias, down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.bias, down_blocks.1.attentions.0.proj_in.weight, down_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.1.attentions.0.transformer_blocks.0.norm3.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_q.weight, up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2.bias, up_blocks.3.attentions.0.transformer_blocks.0.norm2.weight, down_blocks.1.resnets.0.time_emb_proj.weight, down_blocks.0.attentions.1.transformer_blocks.0.ff.net.2.bias, up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight, down_blocks.0.resnets.0.time_emb_proj.weight, up_blocks.1.resnets.1.time_emb_proj.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.0.resnets.0.conv2.weight, up_blocks.0.resnets.1.norm2.weight, up_blocks.2.resnets.1.conv2.bias, up_blocks.1.resnets.0.time_emb_proj.bias, up_blocks.3.attentions.0.norm.weight, mid_block.resnets.1.conv1.weight, down_blocks.2.attentions.1.transformer_blocks.0.norm2.bias, down_blocks.0.attentions.1.proj_in.weight, up_blocks.1.resnets.1.conv1.bias, up_blocks.3.resnets.1.time_emb_proj.bias, down_blocks.0.attentions.0.transformer_blocks.0.norm2.weight, up_blocks.1.attentions.0.transformer_blocks.0.norm2.bias, up_blocks.2.attentions.0.proj_out.bias, up_blocks.1.attentions.0.norm.bias, up_blocks.3.attentions.2.transformer_blocks.0.norm2.weight, up_blocks.2.resnets.2.conv1.weight, up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.bias, up_blocks.2.attentions.2.proj_in.weight, down_blocks.2.attentions.1.norm.weight, down_blocks.2.attentions.0.proj_out.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.1.resnets.0.norm1.bias, up_blocks.1.upsamplers.0.conv.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight, up_blocks.1.attentions.2.proj_in.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight, down_blocks.0.attentions.0.proj_out.weight, up_blocks.1.resnets.1.conv2.weight, up_blocks.1.resnets.2.conv1.weight, up_blocks.1.attentions.0.proj_out.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, down_blocks.3.resnets.0.conv1.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm1.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_v.weight, up_blocks.0.resnets.2.conv_shortcut.weight, conv_in.bias, mid_block.attentions.0.transformer_blocks.0.norm2.bias, mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.1.downsamplers.0.conv.bias, down_blocks.2.attentions.1.transformer_blocks.0.norm1.bias, down_blocks.2.attentions.1.proj_in.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_q.weight, up_blocks.0.resnets.1.conv2.bias, up_blocks.2.resnets.1.conv1.bias, down_blocks.1.attentions.1.transformer_blocks.0.norm2.weight, up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight, mid_block.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.3.attentions.0.norm.bias, down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.3.resnets.1.conv2.bias, down_blocks.0.resnets.1.conv1.weight, down_blocks.2.resnets.1.conv2.weight, up_blocks.1.resnets.2.time_emb_proj.bias, up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.2.attentions.2.transformer_blocks.0.norm2.bias, up_blocks.3.attentions.1.norm.weight, up_blocks.1.resnets.2.norm2.bias, down_blocks.1.resnets.1.conv1.weight, up_blocks.2.resnets.0.conv_shortcut.bias, up_blocks.2.resnets.1.conv_shortcut.bias, up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_q.weight, up_blocks.3.resnets.0.time_emb_proj.bias, up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, down_blocks.1.resnets.1.conv2.weight, conv_in.weight, up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_q.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn1.to_v.weight, up_blocks.2.attentions.0.transformer_blocks.0.norm2.bias, down_blocks.0.resnets.0.conv2.bias, down_blocks.2.attentions.0.proj_out.bias, up_blocks.0.resnets.2.conv1.weight, up_blocks.3.attentions.0.transformer_blocks.0.norm3.bias, up_blocks.2.resnets.2.norm2.weight, down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight, up_blocks.1.resnets.0.time_emb_proj.weight, up_blocks.1.attentions.0.proj_in.bias, down_blocks.1.resnets.1.norm1.weight, down_blocks.0.resnets.0.norm1.weight, mid_block.attentions.0.transformer_blocks.0.norm2.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.3.resnets.2.norm2.bias, down_blocks.1.resnets.1.time_emb_proj.weight, up_blocks.1.resnets.1.norm2.weight, down_blocks.0.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.1.resnets.2.conv1.bias, down_blocks.0.resnets.1.conv2.weight, down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.weight, up_blocks.2.attentions.0.proj_in.bias, up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight, up_blocks.0.resnets.2.norm2.bias, up_blocks.3.resnets.2.time_emb_proj.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm3.bias, down_blocks.2.attentions.1.transformer_blocks.0.norm2.weight, down_blocks.2.attentions.0.transformer_blocks.0.norm3.bias, down_blocks.2.resnets.1.conv1.bias, down_blocks.1.resnets.0.conv1.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_out.0.bias, down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_q.weight, down_blocks.2.attentions.0.transformer_blocks.0.norm2.bias, up_blocks.1.attentions.2.proj_out.weight, mid_block.resnets.1.time_emb_proj.weight, up_blocks.1.resnets.2.norm1.bias, up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.2.resnets.0.norm1.weight, up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.3.resnets.1.conv2.bias, mid_block.resnets.0.conv2.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight, down_blocks.1.attentions.0.proj_out.bias, up_blocks.1.attentions.0.transformer_blocks.0.norm1.weight, up_blocks.3.attentions.2.proj_in.bias, mid_block.resnets.0.conv1.bias, up_blocks.0.resnets.2.conv2.bias, up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.1.resnets.0.norm1.bias, up_blocks.0.resnets.1.conv1.bias, down_blocks.0.attentions.0.norm.bias, up_blocks.2.attentions.2.norm.bias, up_blocks.2.attentions.2.proj_out.bias, up_blocks.3.attentions.0.transformer_blocks.0.norm3.weight, up_blocks.3.resnets.2.norm2.weight, up_blocks.0.resnets.1.conv1.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.2.attentions.2.proj_out.weight, down_blocks.2.resnets.1.time_emb_proj.bias, mid_block.resnets.0.norm1.bias, up_blocks.2.resnets.0.conv2.bias, down_blocks.0.resnets.1.norm1.weight, up_blocks.2.attentions.2.transformer_blocks.0.norm3.bias, up_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.2.attentions.1.transformer_blocks.0.norm2.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight, down_blocks.1.attentions.1.transformer_blocks.0.ff.net.2.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight, up_blocks.3.attentions.1.proj_in.bias, up_blocks.3.attentions.1.transformer_blocks.0.norm2.bias, up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2.weight, up_blocks.1.attentions.1.proj_in.weight, down_blocks.2.resnets.1.norm1.bias, time_embedding.linear_1.bias, down_blocks.1.resnets.0.norm2.weight, down_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.bias, up_blocks.0.resnets.0.conv1.bias, up_blocks.0.resnets.1.conv2.weight, down_blocks.0.resnets.1.norm2.bias, up_blocks.1.resnets.0.conv1.weight, up_blocks.2.resnets.1.norm1.bias, up_blocks.3.attentions.1.transformer_blocks.0.ff.net.2.weight, up_blocks.2.attentions.2.transformer_blocks.0.norm1.bias, down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_k.weight, mid_block.resnets.1.conv2.bias, down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, mid_block.resnets.0.norm1.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.0.attentions.1.proj_in.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn1.to_q.weight, mid_block.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.0.resnets.0.norm2.weight, up_blocks.0.resnets.1.norm1.bias, down_blocks.1.attentions.0.norm.bias, up_blocks.1.attentions.1.transformer_blocks.0.norm1.bias, down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.3.attentions.1.transformer_blocks.0.norm1.weight, up_blocks.3.resnets.0.conv2.bias, up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2.bias, down_blocks.2.attentions.0.proj_in.bias, up_blocks.3.resnets.0.time_emb_proj.weight, down_blocks.1.attentions.0.transformer_blocks.0.norm3.weight, up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.0.attentions.0.transformer_blocks.0.norm1.weight, down_blocks.2.resnets.0.time_emb_proj.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight, down_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.1.attentions.2.transformer_blocks.0.ff.net.2.weight, up_blocks.3.attentions.2.proj_in.weight, up_blocks.3.attentions.0.proj_out.bias, down_blocks.2.attentions.1.norm.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn1.to_v.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_k.weight, up_blocks.1.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.0.attentions.0.transformer_blocks.0.norm1.bias, down_blocks.3.resnets.0.norm2.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn1.to_k.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.2.attentions.0.transformer_blocks.0.norm3.bias, down_blocks.2.resnets.0.norm1.bias, up_blocks.3.attentions.0.proj_out.weight, down_blocks.1.attentions.0.norm.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight, up_blocks.2.resnets.0.conv1.weight, up_blocks.0.resnets.0.conv1.weight, up_blocks.2.upsamplers.0.conv.bias, up_blocks.2.attentions.1.transformer_blocks.0.ff.net.2.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight, up_blocks.2.attentions.1.proj_in.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight, mid_block.attentions.0.transformer_blocks.0.norm3.bias, down_blocks.1.attentions.0.transformer_blocks.0.norm1.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight, up_blocks.1.attentions.2.norm.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight, mid_block.attentions.0.norm.bias, up_blocks.2.resnets.0.norm1.weight, mid_block.resnets.1.conv2.weight, up_blocks.0.resnets.0.time_emb_proj.bias, up_blocks.2.attentions.1.proj_in.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.0.resnets.1.time_emb_proj.bias, down_blocks.0.attentions.1.transformer_blocks.0.norm2.bias, down_blocks.3.resnets.1.norm1.bias, up_blocks.1.resnets.0.conv2.bias, up_blocks.1.resnets.2.conv_shortcut.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight, down_blocks.2.attentions.0.transformer_blocks.0.norm1.bias, up_blocks.2.resnets.2.time_emb_proj.bias, down_blocks.2.resnets.0.norm2.weight, down_blocks.2.resnets.0.conv_shortcut.bias, up_blocks.2.resnets.2.norm1.weight, down_blocks.1.attentions.1.transformer_blocks.0.norm1.bias, down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_k.weight, down_blocks.3.resnets.0.norm1.weight, up_blocks.2.resnets.1.time_emb_proj.bias, up_blocks.0.resnets.2.conv2.weight, down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight, up_blocks.3.attentions.1.proj_out.weight, up_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.3.resnets.0.conv_shortcut.bias, down_blocks.0.resnets.1.norm2.weight, down_blocks.1.attentions.1.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight, down_blocks.0.attentions.0.transformer_blocks.0.norm3.weight, up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight, up_blocks.2.attentions.2.transformer_blocks.0.norm3.weight, down_blocks.2.resnets.0.conv2.bias, down_blocks.1.resnets.0.conv_shortcut.bias, mid_block.resnets.0.time_emb_proj.weight, up_blocks.2.attentions.1.norm.bias, up_blocks.3.attentions.2.norm.bias, down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight, down_blocks.3.resnets.1.norm2.weight, mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.weight, up_blocks.0.resnets.0.conv2.bias, up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_q.weight, up_blocks.3.resnets.2.conv1.bias, down_blocks.1.resnets.1.norm1.bias, up_blocks.1.resnets.0.conv_shortcut.bias, up_blocks.1.resnets.2.conv_shortcut.bias, up_blocks.3.attentions.1.norm.bias, down_blocks.0.attentions.0.transformer_blocks.0.norm2.bias, conv_out.bias, up_blocks.1.attentions.1.proj_out.weight, up_blocks.2.resnets.0.norm2.weight, up_blocks.3.resnets.2.time_emb_proj.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight, up_blocks.3.attentions.1.transformer_blocks.0.norm3.bias, down_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight, down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_q.weight, down_blocks.2.resnets.0.norm2.bias, down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight, up_blocks.2.attentions.1.transformer_blocks.0.norm1.bias, mid_block.attentions.0.transformer_blocks.0.attn2.to_q.weight, up_blocks.3.attentions.2.transformer_blocks.0.ff.net.2.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_out.0.bias, down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.1.attentions.1.norm.bias, mid_block.attentions.0.transformer_blocks.0.ff.net.2.bias, down_blocks.2.downsamplers.0.conv.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_q.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_v.weight, up_blocks.2.attentions.2.transformer_blocks.0.attn1.to_k.weight, conv_norm_out.bias, up_blocks.3.resnets.2.conv1.weight, mid_block.attentions.0.transformer_blocks.0.norm1.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight, up_blocks.1.attentions.0.transformer_blocks.0.norm1.bias, up_blocks.1.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.2.attentions.1.transformer_blocks.0.norm3.bias, up_blocks.1.resnets.2.conv2.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_q.weight, up_blocks.0.resnets.1.conv_shortcut.weight, up_blocks.3.resnets.2.norm1.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_q.weight, up_blocks.3.resnets.0.norm2.weight, down_blocks.2.resnets.1.norm1.weight, up_blocks.1.attentions.2.transformer_blocks.0.norm3.bias, up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight, up_blocks.1.resnets.1.norm1.bias, down_blocks.1.downsamplers.0.conv.weight, down_blocks.1.attentions.0.proj_in.bias, up_blocks.1.resnets.0.norm1.weight, up_blocks.1.resnets.1.time_emb_proj.bias, mid_block.resnets.0.norm2.weight, up_blocks.2.resnets.2.conv2.bias, down_blocks.0.attentions.0.transformer_blocks.0.attn1.to_v.weight, down_blocks.1.attentions.0.transformer_blocks.0.norm1.bias, down_blocks.1.resnets.0.conv2.weight, up_blocks.2.resnets.1.norm2.bias, down_blocks.1.attentions.1.norm.bias, up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_out.0.weight, down_blocks.1.resnets.1.norm2.bias, down_blocks.2.resnets.1.time_emb_proj.weight, down_blocks.1.resnets.1.conv2.bias, down_blocks.3.resnets.0.time_emb_proj.weight, up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_k.weight, down_blocks.3.resnets.1.conv1.bias, down_blocks.2.attentions.0.transformer_blocks.0.attn1.to_q.weight, up_blocks.3.attentions.2.transformer_blocks.0.attn1.to_out.0.weight, down_blocks.2.attentions.0.norm.weight, down_blocks.0.resnets.1.norm1.bias, up_blocks.2.resnets.1.conv_shortcut.weight, down_blocks.0.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, down_blocks.2.attentions.0.transformer_blocks.0.norm2.weight, down_blocks.2.attentions.1.transformer_blocks.0.attn1.to_k.weight, up_blocks.3.attentions.0.proj_in.bias, up_blocks.2.resnets.2.conv_shortcut.weight, up_blocks.0.resnets.2.conv1.bias, up_blocks.3.attentions.1.proj_in.weight, up_blocks.2.attentions.2.transformer_blocks.0.ff.net.2.weight, mid_block.attentions.0.proj_in.bias, up_blocks.1.attentions.0.transformer_blocks.0.norm3.bias, up_blocks.1.attentions.1.transformer_blocks.0.norm3.bias, mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight, down_blocks.1.attentions.1.transformer_blocks.0.norm3.weight, down_blocks.0.downsamplers.0.conv.weight, up_blocks.1.attentions.1.transformer_blocks.0.attn1.to_v.weight, up_blocks.1.resnets.2.conv2.bias, down_blocks.0.resnets.0.norm1.bias, up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_q.weight, down_blocks.1.attentions.1.proj_in.weight, down_blocks.0.attentions.1.norm.weight, down_blocks.0.resnets.0.norm2.bias, up_blocks.1.attentions.1.transformer_blocks.0.norm3.weight, up_blocks.0.upsamplers.0.conv.bias, down_blocks.3.resnets.1.norm1.weight, down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.2.attentions.0.transformer_blocks.0.norm2.weight, down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.2.upsamplers.0.conv.weight, down_blocks.0.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, down_blocks.0.attentions.0.proj_in.weight, down_blocks.3.resnets.0.norm2.bias, down_blocks.1.attentions.1.transformer_blocks.0.norm2.bias, down_blocks.1.resnets.0.conv_shortcut.weight, mid_block.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.3.attentions.2.proj_out.weight, up_blocks.3.resnets.1.conv1.weight, up_blocks.0.resnets.0.norm1.bias, up_blocks.3.resnets.2.conv_shortcut.weight, mid_block.resnets.1.norm2.bias, up_blocks.1.resnets.0.conv1.bias, up_blocks.1.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.attentions.0.norm.weight, down_blocks.2.attentions.0.transformer_blocks.0.ff.net.2.weight, up_blocks.2.attentions.1.proj_out.weight, mid_block.attentions.0.transformer_blocks.0.attn1.to_k.weight, up_blocks.2.attentions.1.transformer_blocks.0.ff.net.0.proj.bias, mid_block.resnets.1.time_emb_proj.bias, down_blocks.2.attentions.0.transformer_blocks.0.norm1.weight, conv_norm_out.weight, up_blocks.2.attentions.0.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.3.resnets.1.conv1.bias, down_blocks.0.resnets.0.time_emb_proj.bias, up_blocks.1.resnets.2.norm1.weight, up_blocks.3.resnets.1.norm2.bias, up_blocks.3.attentions.2.transformer_blocks.0.norm3.bias, down_blocks.0.attentions.0.transformer_blocks.0.norm3.bias, down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_q.weight, up_blocks.3.attentions.2.transformer_blocks.0.norm3.weight, up_blocks.0.resnets.0.time_emb_proj.weight, mid_block.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, down_blocks.1.attentions.0.proj_out.weight, up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight, down_blocks.1.resnets.1.conv1.bias, up_blocks.2.resnets.2.conv_shortcut.bias, down_blocks.0.resnets.0.conv2.weight, up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.0.resnets.1.norm2.bias, up_blocks.3.attentions.0.transformer_blocks.0.ff.net.0.proj.weight, up_blocks.1.attentions.2.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.2.attentions.1.transformer_blocks.0.norm3.weight, down_blocks.2.attentions.0.transformer_blocks.0.norm3.weight, mid_block.attentions.0.transformer_blocks.0.norm3.weight, up_blocks.3.resnets.0.norm1.weight, up_blocks.3.resnets.1.conv_shortcut.bias, down_blocks.1.resnets.0.time_emb_proj.bias, down_blocks.1.attentions.0.transformer_blocks.0.norm2.bias, time_embedding.linear_2.weight, up_blocks.2.attentions.1.proj_out.bias, up_blocks.3.resnets.0.conv_shortcut.weight, up_blocks.1.attentions.1.transformer_blocks.0.norm2.weight, down_blocks.0.resnets.1.time_emb_proj.weight, up_blocks.2.attentions.0.transformer_blocks.0.attn1.to_out.0.bias, up_blocks.3.attentions.1.transformer_blocks.0.attn1.to_v.weight, down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_out.0.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_q.weight, up_blocks.1.attentions.2.transformer_blocks.0.ff.net.0.proj.bias, up_blocks.3.attentions.1.transformer_blocks.0.norm2.weight, mid_block.attentions.0.transformer_blocks.0.attn1.to_v.weight, up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight, down_blocks.0.attentions.1.norm.bias, up_blocks.3.attentions.0.transformer_blocks.0.norm1.bias, down_blocks.0.attentions.0.norm.weight, up_blocks.1.attentions.0.norm.weight, up_blocks.2.attentions.0.norm.bias, up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_out.0.weight, up_blocks.1.resnets.0.norm2.bias, down_blocks.2.attentions.1.proj_in.bias.
 Please make sure to pass `low_cpu_mem_usage=False` and `device_map=None` if you want to randomly initialize those weights or else make sure your checkpoint file is correct.
@chflame163
Copy link
Owner

chflame163 commented Sep 19, 2024

I'm glad you like it this node. Its inference part is transplanted from https://github.com/Zheng-Chong/CatVTON, I think you should go talk to him :)

@Siraj-HM
Copy link

Siraj-HM commented Oct 5, 2024

@chflame163 it fails to in-paint hands correctly. Do you have any idea why? Is that a model limitation ?

@chflame163
Copy link
Owner

it fails to in-paint hands correctly. Do you have any idea why? Is that a model limitation ?

Try adjust the mask_grow value.

@Siraj-HM
Copy link

Siraj-HM commented Oct 6, 2024

it fails to in-paint hands correctly. Do you have any idea why? Is that a model limitation ?

Try adjust the mask_grow value.

Sure will do. I am using the original auto-mask generator for masking. Maybe that is causing the issue ?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants