You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
it was working fine yesterday but now I am having this error... I don't know why, it's my first time using an image-generation model so I don't know what to do. it is working fine using an image from ComfyUI_examples as a workflow, it is occurring when I am using a workflow with LoRA or I don't know if something else is happening in the background, but it was working fine before even with the LoRA and now its giving Allocation on device error.
2024-09-16 01:15:35,690 - root - INFO - Total VRAM 24576 MB, total RAM 16305 MB
2024-09-16 01:15:35,691 - root - INFO - pytorch version: 2.4.1+cu124
2024-09-16 01:15:35,691 - root - INFO - Set vram state to: NORMAL_VRAM
2024-09-16 01:15:35,691 - root - INFO - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-09-16 01:15:36,433 - root - INFO - Using pytorch cross attention
2024-09-16 01:15:37,505 - root - INFO - [Prompt Server] web root: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\web
2024-09-16 01:15:38,351 - root - INFO -
Import times for custom nodes:
2024-09-16 01:15:38,352 - root - INFO - 0.0 seconds: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
2024-09-16 01:15:38,352 - root - INFO - 0.0 seconds: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\comfy-image-saver
2024-09-16 01:15:38,352 - root - INFO - 0.1 seconds: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Crystools
2024-09-16 01:15:38,352 - root - INFO - 0.2 seconds: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\x-flux-comfyui
2024-09-16 01:15:38,352 - root - INFO - 0.3 seconds: E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
2024-09-16 01:15:38,352 - root - INFO -
2024-09-16 01:15:38,357 - root - INFO - Starting server
2024-09-16 01:15:38,357 - root - INFO - To see the GUI go to: http://127.0.0.1:8188
2024-09-16 01:17:48,718 - root - INFO - got prompt
2024-09-16 01:17:48,720 - root - ERROR - Failed to validate prompt for output 9:
2024-09-16 01:17:48,720 - root - ERROR - * UNETLoader 12:
2024-09-16 01:17:48,720 - root - ERROR - - Value not in list: unet_name: 'flux1-dev-fp8.safetensors' not in ['flux1-dev.safetensors', 'flux1-schnell.safetensors']
2024-09-16 01:17:48,721 - root - ERROR - Output will be ignored
2024-09-16 01:17:48,721 - root - WARNING - invalid prompt: {'type': 'prompt_outputs_failed_validation', 'message': 'Prompt outputs failed validation', 'details': '', 'extra_info': {}}
2024-09-16 01:18:08,877 - root - INFO - got prompt
2024-09-16 01:18:08,923 - root - INFO - Using pytorch attention in VAE
2024-09-16 01:18:08,925 - root - INFO - Using pytorch attention in VAE
2024-09-16 01:18:14,244 - root - INFO - model weight dtype torch.bfloat16, manual cast: None
2024-09-16 01:18:14,252 - root - INFO - model_type FLUX
2024-09-16 01:19:32,926 - root - INFO - Requested to load FluxClipModel_
2024-09-16 01:19:32,926 - root - INFO - Loading 1 new model
2024-09-16 01:19:32,936 - root - INFO - loaded completely 0.0 9319.23095703125 True
2024-09-16 01:19:44,434 - root - WARNING - clip missing: ['text_projection.weight']
2024-09-16 01:19:46,098 - root - INFO - Requested to load Flux
2024-09-16 01:19:46,098 - root - INFO - Loading 1 new model
2024-09-16 01:20:26,283 - root - ERROR - ERROR lora diffusion_model.double_blocks.15.img_mlp.0.weight Allocation on device
2024-09-16 01:20:26,424 - root - ERROR - ERROR lora diffusion_model.double_blocks.14.txt_mlp.0.weight Allocation on device
2024-09-16 01:20:26,576 - root - ERROR - ERROR lora diffusion_model.double_blocks.14.img_mlp.0.weight Allocation on device
2024-09-16 01:20:26,858 - root - ERROR - !!! Exception during processing !!! Allocation on device
2024-09-16 01:20:26,905 - root - ERROR - Traceback (most recent call last):
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy_extras\nodes_custom_sampler.py", line 612, in sample
samples = guider.sample(noise.generate_noise(latent), latent_image, sampler, sigmas, denoise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=noise.seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in sample
self.inner_model, self.conds, self.loaded_models = comfy.sampler_helpers.prepare_sampling(self.model_patcher, noise.shape, self.conds)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\sampler_helpers.py", line 66, in prepare_sampling
comfy.model_management.load_models_gpu([model] + models, memory_required=memory_required, minimum_memory_required=minimum_memory_required)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 545, in load_models_gpu
cur_loaded_model = loaded_model.model_load(lowvram_model_memory, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 327, in model_load
raise e
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 323, in model_load
self.real_model = self.model.patch_model(device_to=patch_model_to, lowvram_model_memory=lowvram_model_memory, load_weights=load_weights, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 427, in patch_model
self.load(device_to, lowvram_model_memory=lowvram_model_memory, force_patch_weights=force_patch_weights, full_load=full_load)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 393, in load
self.patch_weight_to_device(weight_key, device_to=device_to)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 320, in patch_weight_to_device
temp_weight = comfy.model_management.cast_to_device(weight, device_to, torch.float32, copy=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 857, in cast_to_device
return tensor.to(device, copy=copy, non_blocking=non_blocking).to(dtype, non_blocking=non_blocking)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
torch.OutOfMemoryError: Allocation on device
2024-09-16 01:20:26,910 - root - ERROR - Got an OOM, unloading all loaded models.
2024-09-16 01:20:27,032 - root - INFO - Prompt executed in 138.15 seconds
2024-09-16 01:37:11,648 - root - INFO - got prompt
2024-09-16 01:37:11,717 - root - INFO - Requested to load FluxClipModel_
2024-09-16 01:37:11,717 - root - INFO - Loading 1 new model
2024-09-16 01:37:29,424 - root - INFO - loaded completely 0.0 9319.23095703125 True
2024-09-16 01:37:29,744 - root - INFO - Requested to load Flux
2024-09-16 01:37:29,744 - root - INFO - Loading 1 new model
2024-09-16 01:38:12,806 - root - ERROR - ERROR lora diffusion_model.double_blocks.17.img_mlp.0.weight Allocation on device
2024-09-16 01:38:12,956 - root - ERROR - ERROR lora diffusion_model.double_blocks.16.txt_mlp.0.weight Allocation on device
2024-09-16 01:38:13,268 - root - ERROR - ERROR lora diffusion_model.double_blocks.15.txt_mlp.0.weight Allocation on device
2024-09-16 01:38:13,404 - root - ERROR - ERROR lora diffusion_model.double_blocks.15.img_mlp.0.weight Allocation on device
2024-09-16 01:38:13,555 - root - ERROR - ERROR lora diffusion_model.double_blocks.14.txt_mlp.0.weight Allocation on device
2024-09-16 01:38:13,839 - root - ERROR - !!! Exception during processing !!! Allocation on device
2024-09-16 01:38:13,873 - root - ERROR - Traceback (most recent call last):
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy_extras\nodes_custom_sampler.py", line 612, in sample
samples = guider.sample(noise.generate_noise(latent), latent_image, sampler, sigmas, denoise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=noise.seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in sample
self.inner_model, self.conds, self.loaded_models = comfy.sampler_helpers.prepare_sampling(self.model_patcher, noise.shape, self.conds)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\sampler_helpers.py", line 66, in prepare_sampling
comfy.model_management.load_models_gpu([model] + models, memory_required=memory_required, minimum_memory_required=minimum_memory_required)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 545, in load_models_gpu
cur_loaded_model = loaded_model.model_load(lowvram_model_memory, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 327, in model_load
raise e
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 323, in model_load
self.real_model = self.model.patch_model(device_to=patch_model_to, lowvram_model_memory=lowvram_model_memory, load_weights=load_weights, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 427, in patch_model
self.load(device_to, lowvram_model_memory=lowvram_model_memory, force_patch_weights=force_patch_weights, full_load=full_load)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 393, in load
self.patch_weight_to_device(weight_key, device_to=device_to)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 320, in patch_weight_to_device
temp_weight = comfy.model_management.cast_to_device(weight, device_to, torch.float32, copy=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 857, in cast_to_device
return tensor.to(device, copy=copy, non_blocking=non_blocking).to(dtype, non_blocking=non_blocking)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
torch.OutOfMemoryError: Allocation on device
2024-09-16 01:38:13,877 - root - ERROR - Got an OOM, unloading all loaded models.
2024-09-16 01:38:14,002 - root - INFO - Prompt executed in 62.35 seconds
2024-09-16 01:40:06,294 - root - INFO - got prompt
2024-09-16 01:40:06,546 - root - INFO - Requested to load FluxClipModel_
2024-09-16 01:40:06,546 - root - INFO - Loading 1 new model
2024-09-16 01:40:25,405 - root - INFO - loaded completely 0.0 9319.23095703125 True
2024-09-16 01:40:25,736 - root - INFO - Requested to load Flux
2024-09-16 01:40:25,737 - root - INFO - Loading 1 new model
2024-09-16 01:41:07,730 - root - ERROR - ERROR lora diffusion_model.double_blocks.16.txt_mlp.0.weight Allocation on device
2024-09-16 01:41:07,866 - root - ERROR - ERROR lora diffusion_model.double_blocks.16.img_mlp.0.weight Allocation on device
2024-09-16 01:41:08,000 - root - ERROR - ERROR lora diffusion_model.double_blocks.15.txt_mlp.0.weight Allocation on device
2024-09-16 01:41:08,215 - root - ERROR - !!! Exception during processing !!! Allocation on device
2024-09-16 01:41:08,237 - root - ERROR - Traceback (most recent call last):
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 323, in execute
output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 198, in get_output_data
return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 169, in _map_node_over_list
process_inputs(input_dict, i)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 158, in process_inputs
results.append(getattr(obj, func)(**inputs))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy_extras\nodes_custom_sampler.py", line 612, in sample
samples = guider.sample(noise.generate_noise(latent), latent_image, sampler, sigmas, denoise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=noise.seed)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\samplers.py", line 706, in sample
self.inner_model, self.conds, self.loaded_models = comfy.sampler_helpers.prepare_sampling(self.model_patcher, noise.shape, self.conds)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\sampler_helpers.py", line 66, in prepare_sampling
comfy.model_management.load_models_gpu([model] + models, memory_required=memory_required, minimum_memory_required=minimum_memory_required)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 545, in load_models_gpu
cur_loaded_model = loaded_model.model_load(lowvram_model_memory, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 327, in model_load
raise e
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 323, in model_load
self.real_model = self.model.patch_model(device_to=patch_model_to, lowvram_model_memory=lowvram_model_memory, load_weights=load_weights, force_patch_weights=force_patch_weights)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 427, in patch_model
self.load(device_to, lowvram_model_memory=lowvram_model_memory, force_patch_weights=force_patch_weights, full_load=full_load)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 393, in load
self.patch_weight_to_device(weight_key, device_to=device_to)
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_patcher.py", line 320, in patch_weight_to_device
temp_weight = comfy.model_management.cast_to_device(weight, device_to, torch.float32, copy=True)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "E:\ComfyUI\ComfyUI_windows_portable\ComfyUI\comfy\model_management.py", line 857, in cast_to_device
return tensor.to(device, copy=copy, non_blocking=non_blocking).to(dtype, non_blocking=non_blocking)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
torch.OutOfMemoryError: Allocation on device
2024-09-16 01:41:08,240 - root - ERROR - Got an OOM, unloading all loaded models.
2024-09-16 01:41:08,368 - root - INFO - Prompt executed in 62.07 seconds
Attached Workflow
Please make sure that workflow does not contain any sensitive information such as API keys or passwords.
{"last_node_id":72,"last_link_id":108,"nodes":[{"id":6,"type":"CLIPTextEncode","pos":{"0":424.71875,"1":618.052001953125},"size":{"0":210,"1":54},"flags":{"collapsed":false},"order":12,"mode":0,"inputs":[{"name":"clip","type":"CLIP","link":108},{"name":"text","type":"STRING","link":47,"slot_index":1,"widget":{"name":"text"}}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[86],"slot_index":0}],"properties":{"Node name for S&R":"CLIPTextEncode"},"widgets_values":[""]},{"id":25,"type":"RandomNoise","pos":{"0":424,"1":236},"size":{"0":327.1990661621094,"1":94.58134460449219},"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"NOISE","type":"NOISE","links":[37],"shape":3}],"properties":{"Node name for S&R":"RandomNoise"},"widgets_values":[655720647218432,"randomize"]},{"id":5,"type":"EmptyLatentImage","pos":{"0":422,"1":101},"size":{"0":330.5548400878906,"1":78},"flags":{},"order":10,"mode":0,"inputs":[{"name":"width","type":"INT","link":101,"widget":{"name":"width"}},{"name":"height","type":"INT","link":103,"widget":{"name":"height"}}],"outputs":[{"name":"LATENT","type":"LATENT","links":[23],"slot_index":0}],"properties":{"Node name for S&R":"EmptyLatentImage"},"widgets_values":[832,1216,1]},{"id":16,"type":"KSamplerSelect","pos":{"0":809,"1":261},"size":{"0":268.2277526855469,"1":58},"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"SAMPLER","type":"SAMPLER","links":[19],"shape":3}],"properties":{"Node name for S&R":"KSamplerSelect"},"widgets_values":["euler"]},{"id":22,"type":"BasicGuider","pos":{"0":893.71875,"1":612.052001953125},"size":{"0":196.9998779296875,"1":62.66668701171875},"flags":{"collapsed":false},"order":16,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":94,"slot_index":0},{"name":"conditioning","type":"CONDITIONING","link":87,"slot_index":1}],"outputs":[{"name":"GUIDER","type":"GUIDER","links":[30],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"BasicGuider"}},{"id":52,"type":"Note","pos":{"0":1148.09375,"1":611.84375},"size":{"0":346.2236022949219,"1":58},"flags":{},"order":2,"mode":0,"inputs":[],"outputs":[],"properties":{"text":""},"widgets_values":["To see the preview, update your ComfyUI and go into the Manager menu. Set \"Preview Method\" to \"Auto\""],"color":"#432","bgcolor":"#653"},{"id":13,"type":"SamplerCustomAdvanced","pos":{"0":1143.252685546875,"1":89.17115783691406},"size":{"0":352.4039611816406,"1":463.3393859863281},"flags":{},"order":17,"mode":0,"inputs":[{"name":"noise","type":"NOISE","link":37,"slot_index":0},{"name":"guider","type":"GUIDER","link":30,"slot_index":1},{"name":"sampler","type":"SAMPLER","link":19,"slot_index":2},{"name":"sigmas","type":"SIGMAS","link":20,"slot_index":3},{"name":"latent_image","type":"LATENT","link":23,"slot_index":4}],"outputs":[{"name":"output","type":"LATENT","links":[24],"slot_index":0,"shape":3},{"name":"denoised_output","type":"LATENT","links":null,"shape":3}],"properties":{"Node name for S&R":"SamplerCustomAdvanced"}},{"id":53,"type":"Note","pos":{"0":381,"1":738},"size":{"0":548.0299072265625,"1":283.5565490722656},"flags":{},"order":3,"mode":0,"inputs":[],"outputs":[],"properties":{"text":""},"widgets_values":["The checkpoint goes in ComfyUI/models/unet (not checkpoints)\nDownload the original weights here:\nhttps://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/flux1-dev.sft\n\nDownload the fp8 version for <24gb vram systems:\nhttps://huggingface.co/Kijai/flux-fp8/blob/main/flux1-dev-fp8.safetensors\n\nText encoders go in ComfyUI/models/clip:\nhttps://huggingface.co/comfyanonymous/flux_text_encoders/tree/main\n\nVAE (ae.sft) goes in ComfyUI/models/vae:\nhttps://huggingface.co/black-forest-labs/FLUX.1-schnell/blob/main/ae.sft\n\nDownload the fp8 t5xxl for degraded quality but less RAM use\nLaunch ComfyUI with \"--lowvram\" arg (in the .bat file) to offload text encoder to CPU.\n\nI can confirm this runs on:\n- RTX 3090 (24gb) 1.29s/it\n- RTX 4070 (12gb) 85s/it\nBoth running the fp8 quantized version. The 4070 is very slow though."],"color":"#432","bgcolor":"#653"},{"id":8,"type":"VAEDecode","pos":{"0":1546,"1":88},"size":{"0":210,"1":46},"flags":{},"order":18,"mode":0,"inputs":[{"name":"samples","type":"LATENT","link":24},{"name":"vae","type":"VAE","link":12}],"outputs":[{"name":"IMAGE","type":"IMAGE","links":[9],"slot_index":0}],"properties":{"Node name for S&R":"VAEDecode"}},{"id":9,"type":"SaveImage","pos":{"0":1784,"1":88},"size":{"0":399.1837463378906,"1":508.5245666503906},"flags":{},"order":19,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":9}],"outputs":[],"properties":{"Node name for S&R":"SaveImage"},"widgets_values":["MarkuryFLUX"]},{"id":61,"type":"ModelSamplingFlux","pos":{"0":754,"1":383},"size":{"0":321.8402404785156,"1":122},"flags":{},"order":13,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":106},{"name":"width","type":"INT","link":102,"widget":{"name":"width"}},{"name":"height","type":"INT","link":104,"widget":{"name":"height"}}],"outputs":[{"name":"MODEL","type":"MODEL","links":[93,94],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"ModelSamplingFlux"},"widgets_values":[1.15,0.5,1024,1024]},{"id":70,"type":"Int Literal","pos":{"0":25,"1":495},"size":{"0":315,"1":58},"flags":{},"order":4,"mode":0,"inputs":[],"outputs":[{"name":"INT","type":"INT","links":[101,102],"slot_index":0,"shape":3}],"title":"Width","properties":{"Node name for S&R":"Int Literal"},"widgets_values":[832]},{"id":71,"type":"Int Literal","pos":{"0":28,"1":610},"size":{"0":315,"1":58},"flags":{},"order":5,"mode":0,"inputs":[],"outputs":[{"name":"INT","type":"INT","links":[103,104],"slot_index":0,"shape":3}],"title":"Height","properties":{"Node name for S&R":"Int Literal"},"widgets_values":[1216]},{"id":60,"type":"FluxGuidance","pos":{"0":659,"1":614},"size":{"0":211.60000610351562,"1":58},"flags":{},"order":14,"mode":0,"inputs":[{"name":"conditioning","type":"CONDITIONING","link":86}],"outputs":[{"name":"CONDITIONING","type":"CONDITIONING","links":[87],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"FluxGuidance"},"widgets_values":[3.5],"color":"#323","bgcolor":"#535"},{"id":11,"type":"DualCLIPLoader","pos":{"0":22,"1":214},"size":{"0":315,"1":106},"flags":{},"order":6,"mode":0,"inputs":[],"outputs":[{"name":"CLIP","type":"CLIP","links":[108],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"DualCLIPLoader"},"widgets_values":["t5xxl_fp16.safetensors","clip_l.safetensors","flux"]},{"id":10,"type":"VAELoader","pos":{"0":26,"1":379},"size":{"0":315,"1":58},"flags":{},"order":7,"mode":0,"inputs":[],"outputs":[{"name":"VAE","type":"VAE","links":[12],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"VAELoader"},"widgets_values":["ae.safetensors"]},{"id":17,"type":"BasicScheduler","pos":{"0":797,"1":94},"size":{"0":281.2428283691406,"1":106},"flags":{},"order":15,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":93,"slot_index":0}],"outputs":[{"name":"SIGMAS","type":"SIGMAS","links":[20],"shape":3}],"properties":{"Node name for S&R":"BasicScheduler"},"widgets_values":["simple",20,1]},{"id":12,"type":"UNETLoader","pos":{"0":18,"1":84},"size":{"0":315,"1":82},"flags":{},"order":8,"mode":0,"inputs":[],"outputs":[{"name":"MODEL","type":"MODEL","links":[107],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"UNETLoader"},"widgets_values":["flux1-dev.safetensors","default"]},{"id":72,"type":"LoraLoaderModelOnly","pos":{"0":419,"1":403},"size":{"0":315,"1":82},"flags":{},"order":11,"mode":0,"inputs":[{"name":"model","type":"MODEL","link":107}],"outputs":[{"name":"MODEL","type":"MODEL","links":[106],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"LoraLoaderModelOnly"},"widgets_values":["nami_v1.safetensors",1]},{"id":28,"type":"String Literal","pos":{"0":26.5353946685791,"1":790.639892578125},"size":{"0":317.8795471191406,"1":202.01535034179688},"flags":{},"order":9,"mode":0,"inputs":[],"outputs":[{"name":"STRING","type":"STRING","links":[47],"slot_index":0,"shape":3}],"properties":{"Node name for S&R":"String Literal"},"widgets_values":["Nami from one piece"]}],"links":[[9,8,0,9,0,"IMAGE"],[12,10,0,8,1,"VAE"],[19,16,0,13,2,"SAMPLER"],[20,17,0,13,3,"SIGMAS"],[23,5,0,13,4,"LATENT"],[24,13,0,8,0,"LATENT"],[30,22,0,13,1,"GUIDER"],[37,25,0,13,0,"NOISE"],[47,28,0,6,1,"STRING"],[86,6,0,60,0,"CONDITIONING"],[87,60,0,22,1,"CONDITIONING"],[93,61,0,17,0,"MODEL"],[94,61,0,22,0,"MODEL"],[101,70,0,5,0,"INT"],[102,70,0,61,1,"INT"],[103,71,0,5,1,"INT"],[104,71,0,61,2,"INT"],[106,72,0,61,0,"MODEL"],[107,12,0,72,0,"MODEL"],[108,11,0,6,0,"CLIP"]],"groups":[{"title":"Load FLUX.1","bounding":[1,2,369,693],"color":"#3f789e","font_size":24,"flags":{}},{"title":"Set Parameters","bounding":[379,0,733,526],"color":"#3f789e","font_size":24,"flags":{}},{"title":"FLUX Prompt","bounding":[1,704,368,318],"color":"#3f789e","font_size":24,"flags":{}},{"title":"Conditioning","bounding":[379,535,732,159],"color":"#3f789e","font_size":24,"flags":{}},{"title":"1st Pass","bounding":[1119,0,402,693],"color":"#3f789e","font_size":24,"flags":{}}],"config":{},"extra":{"ds":{"scale":0.9090909090909091,"offset":[239.11152152765774,97.91972990135856]}},"version":0.4}
Additional Context
(Please add any additional context or steps to reproduce the error here)
The text was updated successfully, but these errors were encountered:
Your question
it was working fine yesterday but now I am having this error... I don't know why, it's my first time using an image-generation model so I don't know what to do. it is working fine using an image from ComfyUI_examples as a workflow, it is occurring when I am using a workflow with LoRA or I don't know if something else is happening in the background, but it was working fine before even with the LoRA and now its giving Allocation on device error.
Logs
System Information
Devices
Logs
Attached Workflow
Please make sure that workflow does not contain any sensitive information such as API keys or passwords.
Additional Context
(Please add any additional context or steps to reproduce the error here)
The text was updated successfully, but these errors were encountered: