Advertisement
Guest User

Untitled

a guest
Dec 17th, 2022
81
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 3.39 KB | None | 0 0
  1. Arguments: (0, 'analog style, photorealistic, RAW, 8k, perfect eyes', '', 'None', 'None', <PIL.Image.Image image mode=RGB size=1920x1080 at 0x1301DAB99F0>, None, None, None, None, 0, 20, 0, 4, 0, 1, False, False, 1, 1, 7, 0.5, -1.0, -1.0, 0, 0, 0, False, 512, 512, 0, False, 32, 0, '', '', 0, '<ul>\n<li><code>CFG Scale</code> should be 2 or lower.</li>\n</ul>\n', True, True, '', '', True, 50, True, 1, 0, False, 4, 1, '<p style="margin-bottom:0.75em">Recommended settings: Sampling Steps: 80-100, Sampler: Euler a, Denoising strength: 0.8</p>', 128, 8, ['left', 'right', 'up', 'down'], 1, 0.05, 128, 4, 0, ['left', 'right', 'up', 'down'], False, False, False, False, '', '<p style="margin-bottom:0.75em">Will upscale the image by the selected scale factor; use width and height sliders to set tile size</p>', 64, 0, 2, 1, '', 0, '', True, False, False) {}
  2. Traceback (most recent call last):
  3. File "D:\AI\stable-diffusion-webui\modules\call_queue.py", line 45, in f
  4. res = list(func(*args, **kwargs))
  5. File "D:\AI\stable-diffusion-webui\modules\call_queue.py", line 28, in f
  6. res = func(*args, **kwargs)
  7. File "D:\AI\stable-diffusion-webui\modules\img2img.py", line 152, in img2img
  8. processed = process_images(p)
  9. File "D:\AI\stable-diffusion-webui\modules\processing.py", line 464, in process_images
  10. res = process_images_inner(p)
  11. File "D:\AI\stable-diffusion-webui\modules\processing.py", line 532, in process_images_inner
  12. p.init(p.all_prompts, p.all_seeds, p.all_subseeds)
  13. File "D:\AI\stable-diffusion-webui\modules\processing.py", line 877, in init
  14. self.init_latent = self.sd_model.get_first_stage_encoding(self.sd_model.encode_first_stage(image))
  15. File "D:\AI\stable-diffusion-webui\venv\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
  16. return func(*args, **kwargs)
  17. File "D:\AI\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\diffusion\ddpm.py", line 830, in encode_first_stage
  18. return self.first_stage_model.encode(x)
  19. File "D:\AI\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\models\autoencoder.py", line 83, in encode
  20. h = self.encoder(x)
  21. File "D:\AI\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
  22. return forward_call(*input, **kwargs)
  23. File "D:\AI\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\model.py", line 536, in forward
  24. h = self.mid.attn_1(h)
  25. File "D:\AI\stable-diffusion-webui\venv\lib\site-packages\torch\nn\modules\module.py", line 1130, in _call_impl
  26. return forward_call(*input, **kwargs)
  27. File "D:\AI\stable-diffusion-webui\repositories\stable-diffusion-stability-ai\ldm\modules\diffusionmodules\model.py", line 258, in forward
  28. out = xformers.ops.memory_efficient_attention(q, k, v, attn_bias=None, op=self.attention_op)
  29. File "D:\AI\stable-diffusion-webui\venv\lib\site-packages\xformers\ops.py", line 858, in memory_efficient_attention
  30. ).op
  31. File "D:\AI\stable-diffusion-webui\venv\lib\site-packages\xformers\ops.py", line 726, in op
  32. raise NotImplementedError(f"No operator found for this attention: {self}")
  33. NotImplementedError: No operator found for this attention: AttentionOpDispatch(dtype=torch.float32, device=device(type='cpu'), k=512, has_dropout=False, attn_bias_type=<class 'NoneType'>, kv_len=4096, q_len=4096, kv=512, batch_size=1, num_heads=1)
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement