Advertisement
Guest User

Untitled

a guest
Jun 9th, 2025
19
0
Never
Not a member of Pastebin yet? Sign Up, it unlocks many cool features!
text 6.36 KB | None | 0 0
  1. 2025-06-09 10:45:27,211]::[InvokeAI]::INFO --> Executing queue item 532, session 9523b9bf-1d9b-423c-ac4d-874cd211e386
  2. [2025-06-09 10:45:31,389]::[ModelManagerService]::INFO --> [MODEL CACHE] Loaded model '531c0e81-9165-42e3-97f3-9eb7ee890093:text_encoder_2' (T5EncoderModel) onto cuda device in 3.96s. Total model size: 4667.39MB, VRAM: 4667.39MB (100.0%)
  3. [2025-06-09 10:45:31,532]::[ModelManagerService]::INFO --> [MODEL CACHE] Loaded model '531c0e81-9165-42e3-97f3-9eb7ee890093:tokenizer_2' (T5Tokenizer) onto cuda device in 0.00s. Total model size: 0.03MB, VRAM: 0.00MB (0.0%)
  4. /opt/venv/lib/python3.12/site-packages/bitsandbytes/autograd/_functions.py:315: UserWarning: MatMul8bitLt: inputs will be cast from torch.bfloat16 to float16 during quantization
  5. warnings.warn(f"MatMul8bitLt: inputs will be cast from {A.dtype} to float16 during quantization")
  6. [2025-06-09 10:45:32,541]::[ModelManagerService]::INFO --> [MODEL CACHE] Loaded model 'fff14f82-ca21-486f-90b5-27c224ac4e59:text_encoder' (CLIPTextModel) onto cuda device in 0.11s. Total model size: 469.44MB, VRAM: 469.44MB (100.0%)
  7. [2025-06-09 10:45:32,603]::[ModelManagerService]::INFO --> [MODEL CACHE] Loaded model 'fff14f82-ca21-486f-90b5-27c224ac4e59:tokenizer' (CLIPTokenizer) onto cuda device in 0.00s. Total model size: 0.00MB, VRAM: 0.00MB (0.0%)
  8. [2025-06-09 10:45:50,174]::[ModelManagerService]::WARNING --> [MODEL CACHE] Insufficient GPU memory to load model. Aborting
  9. [2025-06-09 10:45:50,179]::[ModelManagerService]::WARNING --> [MODEL CACHE] Insufficient GPU memory to load model. Aborting
  10. [2025-06-09 10:45:50,211]::[InvokeAI]::ERROR --> Error while invoking session 9523b9bf-1d9b-423c-ac4d-874cd211e386, invocation b1c4de60-6b49-4a0a-bb10-862154b16d74 (flux_denoise): CUDA out of memory. Tried to allocate 126.00 MiB. GPU 0 has a total capacity of 23.65 GiB of which 67.50 MiB is free. Process 2287 has 258.00 MiB memory in use. Process 1850797 has 554.22 MiB memory in use. Process 1853540 has 21.97 GiB memory in use. Of the allocated memory 21.63 GiB is allocated by PyTorch, and 31.44 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
  11. [2025-06-09 10:45:50,211]::[InvokeAI]::ERROR --> Traceback (most recent call last):
  12. File "/opt/invokeai/invokeai/app/services/session_processor/session_processor_default.py", line 129, in run_node
  13. output = invocation.invoke_internal(context=context, services=self._services)
  14. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  15. File "/opt/invokeai/invokeai/app/invocations/baseinvocation.py", line 241, in invoke_internal
  16. output = self.invoke(context)
  17. ^^^^^^^^^^^^^^^^^^^^
  18. File "/opt/venv/lib/python3.12/site-packages/torch/utils/_contextlib.py", line 116, in decorate_context
  19. return func(*args, **kwargs)
  20. ^^^^^^^^^^^^^^^^^^^^^
  21. File "/opt/invokeai/invokeai/app/invocations/flux_denoise.py", line 155, in invoke
  22. latents = self._run_diffusion(context)
  23. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  24. File "/opt/invokeai/invokeai/app/invocations/flux_denoise.py", line 335, in _run_diffusion
  25. (cached_weights, transformer) = exit_stack.enter_context(
  26. ^^^^^^^^^^^^^^^^^^^^^^^^^
  27. File "/root/.local/share/uv/python/cpython-3.12.9-linux-x86_64-gnu/lib/python3.12/contextlib.py", line 526, in enter_context
  28. result = _enter(cm)
  29. ^^^^^^^^^^
  30. File "/root/.local/share/uv/python/cpython-3.12.9-linux-x86_64-gnu/lib/python3.12/contextlib.py", line 137, in __enter__
  31. return next(self.gen)
  32. ^^^^^^^^^^^^^^
  33. File "/opt/invokeai/invokeai/backend/model_manager/load/load_base.py", line 74, in model_on_device
  34. self._cache.lock(self._cache_record, working_mem_bytes)
  35. File "/opt/invokeai/invokeai/backend/model_manager/load/model_cache/model_cache.py", line 53, in wrapper
  36. return method(self, *args, **kwargs)
  37. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  38. File "/opt/invokeai/invokeai/backend/model_manager/load/model_cache/model_cache.py", line 336, in lock
  39. self._load_locked_model(cache_entry, working_mem_bytes)
  40. File "/opt/invokeai/invokeai/backend/model_manager/load/model_cache/model_cache.py", line 408, in _load_locked_model
  41. model_bytes_loaded = self._move_model_to_vram(cache_entry, vram_available + MB)
  42. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  43. File "/opt/invokeai/invokeai/backend/model_manager/load/model_cache/model_cache.py", line 432, in _move_model_to_vram
  44. return cache_entry.cached_model.full_load_to_vram()
  45. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  46. File "/opt/invokeai/invokeai/backend/model_manager/load/model_cache/cached_model/cached_model_only_full_load.py", line 79, in full_load_to_vram
  47. new_state_dict[k] = v.to(self._compute_device, copy=True)
  48. ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  49. torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 126.00 MiB. GPU 0 has a total capacity of 23.65 GiB of which 67.50 MiB is free. Process 2287 has 258.00 MiB memory in use. Process 1850797 has 554.22 MiB memory in use. Process 1853540 has 21.97 GiB memory in use. Of the allocated memory 21.63 GiB is allocated by PyTorch, and 31.44 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)
  50. [2025-06-09 10:45:51,961]::[InvokeAI]::INFO --> Graph stats: 9523b9bf-1d9b-423c-ac4d-874cd211e386
  51. Node Calls Seconds VRAM Used
  52. flux_model_loader 1 0.008s 0.000G
  53. flux_text_encoder 1 5.487s 5.038G
  54. collect 1 0.000s 5.034G
  55. flux_denoise 1 17.466s 21.628G
  56. TOTAL GRAPH EXECUTION TIME: 22.961s
  57. TOTAL GRAPH WALL TIME: 22.965s
  58. RAM used by InvokeAI process: 22.91G (+22.289G)
  59. RAM used to load models: 27.18G
  60. VRAM in use: 0.012G
  61. RAM cache statistics:
  62. Model cache hits: 5
  63. Model cache misses: 5
  64. Models cached: 1
  65. Models cleared from cache: 3
  66. Cache high water mark: 22.17/0.00G
Advertisement
Add Comment
Please, Sign In to add comment
Advertisement