Guest User

AnimateDiff error

a guest
Jul 18th, 2023
505
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 11.66 KB | Software | 0 0
  1. 2023-07-18 08:49:56,434 - AnimateDiff - INFO - Loading motion module mm_sd_v14.ckpt from F:\stable-diffusion-webui\extensions\sd-webui-animatediff\model\mm_sd_v14.ckpt
  2. 2023-07-18 08:49:59,629 - AnimateDiff - WARNING - Missing keys <All keys matched successfully>
  3. 2023-07-18 08:49:59,862 - AnimateDiff - INFO - Injecting motion module mm_sd_v14.ckpt into SD1.5 UNet input blocks.
  4. 2023-07-18 08:49:59,862 - AnimateDiff - INFO - Injecting motion module mm_sd_v14.ckpt into SD1.5 UNet output blocks.
  5. 2023-07-18 08:49:59,862 - AnimateDiff - INFO - Injection finished.
  6. 0%| | 0/40 [00:00<?, ?it/s]
  7. *** Error completing request
  8. *** Arguments: ('task(w5jkpc6lpqyhvxk)', 'a man dancing in the rain', 'disconnected limbs, overexposed, jpeg artifacts, long neck, BadDream, UnrealisticDream', [], 40, 0, False, False, 1, 1, 7, -1.0, -1.0, 0, 0, 0, False, 512, 512, False, 0.7, 2, 'Latent', 0, 0, 0, 0, '', '', [], 0, False, {'ad_model': 'face_yolov8n.pt', 'ad_prompt': '', 'ad_negative_prompt': '', 'ad_confidence': 0.3, 'ad_mask_min_ratio': 0, 'ad_mask_max_ratio': 1, 'ad_x_offset': 0, 'ad_y_offset': 0, 'ad_dilate_erode': 4, 'ad_mask_merge_invert': 'None', 'ad_mask_blur': 4, 'ad_denoising_strength': 0.4, 'ad_inpaint_only_masked': True, 'ad_inpaint_only_masked_padding': 32, 'ad_use_inpaint_width_height': False, 'ad_inpaint_width': 512, 'ad_inpaint_height': 512, 'ad_use_steps': False, 'ad_steps': 28, 'ad_use_cfg_scale': False, 'ad_cfg_scale': 7, 'ad_use_noise_multiplier': False, 'ad_noise_multiplier': 1, 'ad_restore_face': False, 'ad_controlnet_model': 'None', 'ad_controlnet_module': 'inpaint_global_harmonious', 'ad_controlnet_weight': 1, 'ad_controlnet_guidance_start': 0, 'ad_controlnet_guidance_end': 1, 'is_api': ()}, {'ad_model': 'None', 'ad_prompt': '', 'ad_negative_prompt': '', 'ad_confidence': 0.3, 'ad_mask_min_ratio': 0, 'ad_mask_max_ratio': 1, 'ad_x_offset': 0, 'ad_y_offset': 0, 'ad_dilate_erode': 4, 'ad_mask_merge_invert': 'None', 'ad_mask_blur': 4, 'ad_denoising_strength': 0.4, 'ad_inpaint_only_masked': True, 'ad_inpaint_only_masked_padding': 32, 'ad_use_inpaint_width_height': False, 'ad_inpaint_width': 512, 'ad_inpaint_height': 512, 'ad_use_steps': False, 'ad_steps': 28, 'ad_use_cfg_scale': False, 'ad_cfg_scale': 7, 'ad_use_noise_multiplier': False, 'ad_noise_multiplier': 1, 'ad_restore_face': False, 'ad_controlnet_model': 'None', 'ad_controlnet_module': 'inpaint_global_harmonious', 'ad_controlnet_weight': 1, 'ad_controlnet_guidance_start': 0, 'ad_controlnet_guidance_end': 1, 'is_api': ()}, False, 'MultiDiffusion', False, True, 1024, 1024, 96, 96, 48, 4, 'None', 2, False, 10, 1, 1, 64, False, False, False, False, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 0.4, 0.4, 0.2, 0.2, '', '', 'Background', 0.2, -1.0, False, 3072, 192, True, True, True, False, 'Send to Canvas Editor', False, 7, 100, 'Constant', 0, 'Constant', 0, 4, True, 16, 8, 'mm_sd_v14.ckpt', <scripts.controlnet_ui.controlnet_ui_group.UiControlNetUnit object at 0x0000024D7C2C8310>, <scripts.controlnet_ui.controlnet_ui_group.UiControlNetUnit object at 0x0000024D7F4E49A0>, <scripts.controlnet_ui.controlnet_ui_group.UiControlNetUnit object at 0x0000024D7F4E4940>, False, '', 0.5, True, False, '', 'Lerp', False, False, 1, 0.15, False, 'OUT', ['OUT'], 5, 0, 'Bilinear', False, 'Bilinear', False, 'Lerp', '', '', False, None, False, '0', '0', 'F:\\stable-diffusion-webui\\models/roop\\inswapper_128.onnx', 'CodeFormer', 1, True, '', 1, 1, False, True, False, False, False, False, '1:1,1:2,1:2', '0:0,0:0,0:1', '0.2,0.8,0.8', 150, 0.2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, False, False, 'positive', 'comma', 0, False, False, '', 1, '', [], 0, '', [], 0, '', [], True, False, False, False, 0, None, None, False, None, None, False, None, None, False, 50) {}
  9. Traceback (most recent call last):
  10. File "F:\stable-diffusion-webui\modules\call_queue.py", line 55, in f
  11. res = list(func(*args, **kwargs))
  12. File "F:\stable-diffusion-webui\modules\call_queue.py", line 35, in f
  13. res = func(*args, **kwargs)
  14. File "F:\stable-diffusion-webui\modules\txt2img.py", line 57, in txt2img
  15. processed = processing.process_images(p)
  16. File "F:\stable-diffusion-webui\modules\processing.py", line 620, in process_images
  17. res = process_images_inner(p)
  18. File "F:\stable-diffusion-webui\extensions\sd-webui-controlnet\scripts\batch_hijack.py", line 42, in processing_process_images_hijack
  19. return getattr(processing, '__controlnet_original_process_images_inner')(p, *args, **kwargs)
  20. File "F:\stable-diffusion-webui\modules\processing.py", line 739, in process_images_inner
  21. samples_ddim = p.sample(conditioning=p.c, unconditional_conditioning=p.uc, seeds=p.seeds, subseeds=p.subseeds, subseed_strength=p.subseed_strength, prompts=p.prompts)
  22. File "F:\stable-diffusion-webui\modules\processing.py", line 992, in sample
  23. samples = self.sampler.sample(self, x, conditioning, unconditional_conditioning, image_conditioning=self.txt2img_image_conditioning(x))
  24. File "F:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 439, in sample
  25. samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
  26. File "F:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 278, in launch_sampling
  27. return func()
  28. File "F:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 439, in <lambda>
  29. samples = self.launch_sampling(steps, lambda: self.func(self.model_wrap_cfg, x, extra_args={
  30. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context
  31. return func(*args, **kwargs)
  32. File "F:\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\sampling.py", line 145, in sample_euler_ancestral
  33. denoised = model(x, sigmas[i] * s_in, **extra_args)
  34. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  35. return forward_call(*args, **kwargs)
  36. File "F:\stable-diffusion-webui\modules\sd_samplers_kdiffusion.py", line 177, in forward
  37. x_out[a:b] = self.inner_model(x_in[a:b], sigma_in[a:b], cond=make_condition_dict(c_crossattn, image_cond_in[a:b]))
  38. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  39. return forward_call(*args, **kwargs)
  40. File "F:\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 112, in forward
  41. eps = self.get_eps(input * c_in, self.sigma_to_t(sigma), **kwargs)
  42. File "F:\stable-diffusion-webui\repositories\k-diffusion\k_diffusion\external.py", line 138, in get_eps
  43. return self.inner_model.apply_model(*args, **kwargs)
  44. File "F:\stable-diffusion-webui\modules\sd_hijack_utils.py", line 17, in <lambda>
  45. setattr(resolved_obj, func_path[-1], lambda *args, **kwargs: self(*args, **kwargs))
  46. File "F:\stable-diffusion-webui\modules\sd_hijack_utils.py", line 28, in __call__
  47. return self.__orig_func(*args, **kwargs)
  48. File "F:\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 858, in apply_model
  49. x_recon = self.model(x_noisy, t, **cond)
  50. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  51. return forward_call(*args, **kwargs)
  52. File "F:\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 1335, in forward
  53. out = self.diffusion_model(x, t, context=cc)
  54. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  55. return forward_call(*args, **kwargs)
  56. File "F:\stable-diffusion-webui\modules\sd_unet.py", line 91, in UNetModel_forward
  57. return ldm.modules.diffusionmodules.openaimodel.copy_of_UNetModel_forward_for_webui(self, x, timesteps, context, *args, **kwargs)
  58. File "F:\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\openaimodel.py", line 802, in forward
  59. h = module(h, emb, context)
  60. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  61. return forward_call(*args, **kwargs)
  62. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\scripts\animatediff.py", line 21, in mm_tes_forward
  63. x = layer(x, context)
  64. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  65. return forward_call(*args, **kwargs)
  66. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\motion_module.py", line 79, in forward
  67. hidden_states = self.temporal_transformer(hidden_states, encoder_hidden_states, attention_mask)
  68. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  69. return forward_call(*args, **kwargs)
  70. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\motion_module.py", line 151, in forward
  71. hidden_states = block(hidden_states, encoder_hidden_states=encoder_hidden_states, video_length=video_length)
  72. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  73. return forward_call(*args, **kwargs)
  74. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\motion_module.py", line 215, in forward
  75. hidden_states = attention_block(
  76. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1501, in _call_impl
  77. return forward_call(*args, **kwargs)
  78. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\motion_module.py", line 539, in forward
  79. hidden_states = self._memory_efficient_attention_xformers(query, key, value, attention_mask)
  80. File "F:\stable-diffusion-webui\extensions\sd-webui-animatediff\motion_module.py", line 468, in _memory_efficient_attention_xformers
  81. hidden_states = xformers.ops.memory_efficient_attention(query, key, value, attn_bias=attention_mask,
  82. File "F:\stable-diffusion-webui\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 192, in memory_efficient_attention
  83. return _memory_efficient_attention(
  84. File "F:\stable-diffusion-webui\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 290, in _memory_efficient_attention
  85. return _memory_efficient_attention_forward(
  86. File "F:\stable-diffusion-webui\venv\lib\site-packages\xformers\ops\fmha\__init__.py", line 310, in _memory_efficient_attention_forward
  87. out, *_ = op.apply(inp, needs_gradient=False)
  88. File "F:\stable-diffusion-webui\venv\lib\site-packages\xformers\ops\fmha\cutlass.py", line 175, in apply
  89. out, lse, rng_seed, rng_offset = cls.OPERATOR(
  90. File "F:\stable-diffusion-webui\venv\lib\site-packages\torch\_ops.py", line 502, in __call__
  91. return self._op(*args, **kwargs or {})
  92. RuntimeError: CUDA error: invalid configuration argument
  93. CUDA kernel errors might be asynchronously reported at some other API call, so the stacktrace below might be incorrect.
  94. For debugging consider passing CUDA_LAUNCH_BLOCKING=1.
  95. Compile with `TORCH_USE_CUDA_DSA` to enable device-side assertions.
Advertisement
Add Comment
Please, Sign In to add comment