Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[My workflow includes Facereactor and total on a good day is 150 seconds per image generations. and now it is just 50 seconds on average just because of installing Flash Attention WITHOUT making any special changes to my setup???? I DO NOT GET IT I DON"T UNDERSTAND!?!?!?!?!] I had a hard time using ComfyUI on Windows 11 previously... so this is what I changed... #4923

Closed
kairin opened this issue Sep 14, 2024 · 26 comments
Labels
User Support A user needs help with something, probably not a bug.

Comments

@kairin
Copy link

kairin commented Sep 14, 2024

Screenshot 2024-09-14 214934

I installed Ubuntu but running from Windows 11 itself.

I am not sure if there are any real benefits to doing this... as I have no idea how to run benchmarks to verify if there are any improvement gains.

@kairin
Copy link
Author

kairin commented Sep 14, 2024

Screenshot 2024-09-14 212805
Screenshot 2024-09-14 213505

but... since i technically have a linux machine now, i can technically get flash attention 2 running.

even though it is not supported in windows.

can someone direct me how i can benchmark and verify if running ComfyUI in WSL2 via Ubuntu have any speed gains or improvements or something?

@kairin
Copy link
Author

kairin commented Sep 14, 2024

Screenshot 2024-09-14 215737

okay. so first issue that i encounter is the following:

[START] Security scan
DEPRECATION: Loading egg at /home/ks/comfy/venv/lib/python3.12/site-packages/flash_attn-2.6.3-py3.12-linux-x86_64.egg is deprecated. pip 24.3 will enforce this behaviour change. A possible replacement is to use pip for package installation. Discussion can be found at https://github.com/pypa/pip/issues/12330
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-09-14 21:56:00.044347
** Platform: Linux
** Python version: 3.12.3 (main, Jul 31 2024, 17:43:48) [GCC 13.2.0]

pypa/pip#12330

i have no idea what i need to do. can i simply reinstall based on the suggested steps?

i am unsure.

@kairin
Copy link
Author

kairin commented Sep 14, 2024

from the rest of the startup process... i can't actually tell if there's any difference...

i just wish i can figure out how to get rid of all the warnings... they are causing a lot of anxiety.

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/rgthree-comfy
  15.6 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.6.0.dev20240913+cu124
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using pytorch cross attention
[Prompt Server] web root: /home/ks/ComfyUI/web
/home/ks/comfy/venv/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ComfyUI/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Art is the breath of life." - Liza Donnelly

### Loading: ComfyUI-Manager (V2.50.3)
### ComfyUI Revision: 2691 [369a6dd2] | Released on '2024-09-13'

[rgthree] Loaded 42 epic nodes.
[rgthree] NOTE: Will NOT use rgthree's optimized recursive execution as ComfyUI has changed.

/home/ks/comfy/venv/lib/python3.12/site-packages/diffusers/utils/outputs.py:63: FutureWarning: `torch.utils._pytree._register_pytree_node` is deprecated. Please use `torch.utils._pytree.register_pytree_node` instead.
  torch.utils._pytree._register_pytree_node(
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.6.0.dev20240913+cu124
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
[Crystools INFO] Crystools version: 1.16.6
[Crystools INFO] CPU: AMD Ryzen 5 5600X 6-Core Processor - Arch: x86_64 - OS: Linux 5.15.153.1-microsoft-standard-WSL2
[Crystools INFO] Pynvml (Nvidia) initialized.
[Crystools INFO] GPU/s:
[Crystools INFO] 0) NVIDIA GeForce RTX 4080 SUPER
[Crystools INFO] NVIDIA Driver: 560.81

Import times for custom nodes:
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/cg-use-everywhere
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI_Noise
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/cg-noisetools
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/comfy-image-saver
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI_UltimateSDUpscale
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-post-processing-nodes
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-GGUF
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI_essentials
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-GlifNodes
   0.0 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ComfyUI/custom_nodes/ComfyUI-Crystools
   0.9 seconds: /home/ks/ComfyUI/custom_nodes/was-node-suite-comfyui

Starting server

To see the GUI go to: http://127.0.0.1:8188
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json

Copy link

This issue is being marked stale because it has not had any activity for 30 days. Reply below within 7 days if your issue still isn't solved, and it will be left open. Otherwise, the issue will be closed automatically.

@github-actions github-actions bot added the Stale This issue is stale and will be autoclosed soon. label Oct 24, 2024
@kairin
Copy link
Author

kairin commented Oct 26, 2024

image

Okay i am in the process of installing flash attention on my windows 11 pc


ddd  insightface  ven
Sat Oct 26 17:29:25 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   37C    P0             18W /  320W |    1351MiB /  16376MiB |      9%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A       435      G   /Xwayland                                   N/A      |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/ks/ven
VIRTUAL_ENV_PROMPT=(ven)
ddd  insightface  ven
CODEOWNERS       app               comfyui.prev2.log               fix_torch.py       models           pytest.ini        tests
CONTRIBUTING.md  comfy             comfyui_screenshot.png          folder_paths.py    new_updater.py   requirements.txt  tests-unit
LICENSE          comfy_execution   cuda_malloc.py                  input              node_helpers.py  script_examples   user
README.md        comfy_extras      custom_nodes                    latent_preview.py  nodes.py         server.py         utils
__pycache__      comfyui.log       execution.py                    main.py            notebooks        styles            web
api_server       comfyui.prev.log  extra_model_paths.yaml.example  model_filemanager  output           temp
  on    bash   ven 3.12.3    master ≡  ?1                                                                                0ms 
╭─ 17:29:26 |  26 Oct, Saturday |   in     ven  comf
╰─❯ python main.py
[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-26 17:29:34.939822
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/ven/bin/python
** ComfyUI Path: /home/ks/ven/comf
** Log path: /home/ks/ven/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.9 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2787 [d605677b] | Released on '2024-10-25'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.5.0+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 exciting nodes. 🎉

WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ven/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "The artist's world is limitless. It can be found anywhere, far from where he lives or a few feet away. It is always on his doorstep." - Paul Strand

FaceDetailer: Model directory already exists
FaceDetailer: Model already exists

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ven/comf/custom_nodes/Skimmed_CFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-AutomaticCFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
   0.0 seconds: /home/ks/ven/comf/custom_nodes/facerestore_cf
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_ExtraModels
   0.3 seconds: /home/ks/ven/comf/custom_nodes/comfyui-reactor-node
   0.3 seconds: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
   0.3 seconds: /home/ks/ven/comf/custom_nodes/DZ-FaceDetailer
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
   0.6 seconds: /home/ks/ven/comf/custom_nodes/was-node-suite-comfyui
   0.8 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR
   2.4 seconds: /home/ks/ven/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
got prompt
model weight dtype torch.float16, manual cast: None
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
loaded straight to GPU
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
[ReActor] 17:30:13 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
got prompt


0: 640x480 1 face, 55.8ms
0: 640x480 1 face, 55.8ms
Speed: 4.5ms preprocess, 55.8ms inference, 15.3ms postprocess per image at shape (1, 3, 640, 480)
Speed: 4.5ms preprocess, 55.8ms inference, 15.3ms postprocess per image at shape (1, 3, 640, 480)
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:51<00:00,  1.16it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
[ReActor] 17:31:37 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 17:31:44 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 17:31:44 - STATUS - Analyzing Source Image...
[ReActor] 17:31:47 - STATUS - Trying to halve 'det_size' parameter
[ReActor] 17:31:53 - STATUS - Analyzing Target Image...
[ReActor] 17:31:57 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 17:31:57 - STATUS - OK - Detected Gender matches Condition
[ReActor] 17:32:00 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 17:32:00 - STATUS - OK - Detected Gender matches Condition
[ReActor] 17:32:00 - STATUS - Swapping...
[ReActor] 17:32:00 - STATUS - Face Boost is enabled
[ReActor] 17:32:01 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 17:32:04 - STATUS - --Done!--
[ReActor] 17:32:04 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 17:32:21 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  1,646,826,258
VRAMdebug: free memory after:  1,613,767,442
VRAMdebug: freed memory:  -33,058,816
Prompt executed in 150.53 seconds
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:53<00:00,  1.12it/s]
[ReActor] 17:33:33 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 17:33:37 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 17:33:37 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 17:33:37 - STATUS - Analyzing Target Image...
[ReActor] 17:33:37 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 17:33:37 - STATUS - OK - Detected Gender matches Condition
[ReActor] 17:33:37 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 17:33:37 - STATUS - OK - Detected Gender matches Condition
[ReActor] 17:33:37 - STATUS - Swapping...
[ReActor] 17:33:37 - STATUS - Face Boost is enabled
[ReActor] 17:33:38 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 17:33:39 - STATUS - --Done!--
[ReActor] 17:33:39 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 17:33:48 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,540,433,906
VRAMdebug: free memory after:  5,540,433,906
VRAMdebug: freed memory:  0
Prompt executed in 92.59 seconds
got prompt
Killed
  on    bash   ven 3.12.3    master ≡  ?1                                                                        5m 5s 251ms 
╭─ 17:34:39 |  26 Oct, Saturday |   in     ven  comf
╰─❯

if you can see, currently without flash attention and using xformers, my image generations are about 150 seconds or so...

@kairin
Copy link
Author

kairin commented Oct 26, 2024

image

the other window is still installing flash-attn as i type this.

let's see if it makes any difference once it is installed...

i am assuming once it is installed i don't have to do anything... and just run comfy as usual.

image

and it completed the installation.

and it states that it has:


Successfully built flash-attn
Installing collected packages: flash-attn
Successfully installed flash-attn-2.6.3

@kairin
Copy link
Author

kairin commented Oct 26, 2024

image

You are kidding me right?

that's all i got to do and now my image generation based on my existing workflows... wiithout any other special modification...

from the normal 150seconds per generation...

i shaved off 100seconds just like that?????????????????

@github-actions github-actions bot removed the Stale This issue is stale and will be autoclosed soon. label Oct 26, 2024
@kairin
Copy link
Author

kairin commented Oct 26, 2024


ddd  insightface  ven
Sat Oct 26 19:08:12 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   37C    P0             18W /  320W |    1157MiB /  16376MiB |      8%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A       435      G   /Xwayland                                   N/A      |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/ks/ven
VIRTUAL_ENV_PROMPT=(ven)
ddd  insightface  ven
CODEOWNERS       app               comfyui.prev2.log               fix_torch.py       models           pytest.ini        tests
CONTRIBUTING.md  comfy             comfyui_screenshot.png          folder_paths.py    new_updater.py   requirements.txt  tests-unit
LICENSE          comfy_execution   cuda_malloc.py                  input              node_helpers.py  script_examples   user
README.md        comfy_extras      custom_nodes                    latent_preview.py  nodes.py         server.py         utils
__pycache__      comfyui.log       execution.py                    main.py            notebooks        styles            web
api_server       comfyui.prev.log  extra_model_paths.yaml.example  model_filemanager  output           temp
  on    bash   ven 3.12.3    master ≡  ?1                                                                                0ms 
╭─ 19:08:12 |  26 Oct, Saturday |   in     ven  comf
╰─❯ python main.py
[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-26 19:08:19.317467
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/ven/bin/python
** ComfyUI Path: /home/ks/ven/comf
** Log path: /home/ks/ven/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2787 [d605677b] | Released on '2024-10-25'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.5.0+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 magnificent nodes. 🎉

WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ven/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Success is not about being better than someone else. It's about being better than you used to be." - Unknown

FaceDetailer: Model directory already exists
FaceDetailer: Model already exists

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ven/comf/custom_nodes/Skimmed_CFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-AutomaticCFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
   0.0 seconds: /home/ks/ven/comf/custom_nodes/facerestore_cf
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_ExtraModels
   0.1 seconds: /home/ks/ven/comf/custom_nodes/comfyui-reactor-node
   0.2 seconds: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
   0.3 seconds: /home/ks/ven/comf/custom_nodes/DZ-FaceDetailer
   0.3 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
   0.5 seconds: /home/ks/ven/comf/custom_nodes/was-node-suite-comfyui
   0.6 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR
   1.7 seconds: /home/ks/ven/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/1742899825_extension-node-map.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/1514988643_custom-node-list.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/746607195_github-stats.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/1742899825_extension-node-map.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/custom-node-list.json [DONE]
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/1241942342_custom-node-list.json [DONE]
Install custom node 'Phi-3-mini in ComfyUI'
Install: ['https://github.com/ZHO-ZHO-ZHO/ComfyUI-Phi-3-mini']
Download: git clone 'https://github.com/ZHO-ZHO-ZHO/ComfyUI-Phi-3-mini'
100%|██████████████████████████████████████████████████████████████████████████████████████████████████████| 7.0/7.0 [00:00<00:00,  8.66it/s]
Install: pip packages

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'transformers>=4.40.0']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
                 Installation was successful.
After restarting ComfyUI, please refresh the browser.

Restarting... [Legacy Mode]


[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-26 19:10:37.768441
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/ven/bin/python
** ComfyUI Path: /home/ks/ven/comf
** Log path: /home/ks/ven/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2787 [d605677b] | Released on '2024-10-25'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.5.0+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 fantastic nodes. 🎉

WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ven/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Success usually comes to those who are too busy to be looking for it." - Henry David Thoreau

FaceDetailer: Model directory already exists
FaceDetailer: Model already exists

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ven/comf/custom_nodes/Skimmed_CFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-AutomaticCFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/facerestore_cf
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_ExtraModels
   0.1 seconds: /home/ks/ven/comf/custom_nodes/comfyui-reactor-node
   0.2 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
   0.2 seconds: /home/ks/ven/comf/custom_nodes/DZ-FaceDetailer
   0.4 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR
   0.5 seconds: /home/ks/ven/comf/custom_nodes/was-node-suite-comfyui
   0.6 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Phi-3-mini
   1.3 seconds: /home/ks/ven/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
got prompt
model weight dtype torch.float16, manual cast: None
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:34<00:00,  1.75it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
[ReActor] 19:24:08 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:24:13 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512


0: 640x480 1 face, 31.9ms
0: 640x480 1 face, 31.9ms
Speed: 1.5ms preprocess, 31.9ms inference, 7.9ms postprocess per image at shape (1, 3, 640, 480)
Speed: 1.5ms preprocess, 31.9ms inference, 7.9ms postprocess per image at shape (1, 3, 640, 480)
[ReActor] 19:24:15 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:24:15 - STATUS - Analyzing Source Image...
[ReActor] 19:24:17 - STATUS - Trying to halve 'det_size' parameter
[ReActor] 19:24:21 - STATUS - Analyzing Target Image...
[ReActor] 19:24:24 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:24:24 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:24:24 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:24:24 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:24:24 - STATUS - Swapping...
[ReActor] 19:24:24 - STATUS - Face Boost is enabled
[ReActor] 19:24:25 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:24:26 - STATUS - --Done!--
[ReActor] 19:24:26 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:24:45 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  1,675,718,354
VRAMdebug: free memory after:  1,626,177,234
VRAMdebug: freed memory:  -49,541,120
Prompt executed in 93.70 seconds
got prompt
model weight dtype torch.float16, manual cast: None
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:33<00:00,  1.77it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
[ReActor] 19:25:45 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:25:48 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512


0: 640x480 1 face, 7.5ms
0: 640x480 1 face, 7.5ms
Speed: 1.3ms preprocess, 7.5ms inference, 1.6ms postprocess per image at shape (1, 3, 640, 480)
Speed: 1.3ms preprocess, 7.5ms inference, 1.6ms postprocess per image at shape (1, 3, 640, 480)
[ReActor] 19:25:50 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:25:50 - STATUS - Analyzing Source Image...
[ReActor] 19:25:50 - STATUS - Trying to halve 'det_size' parameter
[ReActor] 19:25:50 - STATUS - Analyzing Target Image...
[ReActor] 19:25:50 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:25:50 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:25:50 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:25:50 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:25:50 - STATUS - Swapping...
[ReActor] 19:25:50 - STATUS - Face Boost is enabled
[ReActor] 19:25:50 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:25:51 - STATUS - --Done!--
[ReActor] 19:25:51 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:25:59 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,276,451,682
VRAMdebug: free memory after:  5,276,451,682
VRAMdebug: freed memory:  0
Prompt executed in 61.69 seconds
got prompt
loaded completely 7571.969965744019 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:33<00:00,  1.78it/s]
[ReActor] 19:26:44 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:26:47 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:26:47 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 19:26:47 - STATUS - Analyzing Target Image...
[ReActor] 19:26:47 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:26:47 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:26:47 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:26:47 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:26:47 - STATUS - Swapping...
[ReActor] 19:26:47 - STATUS - Face Boost is enabled
[ReActor] 19:26:47 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:26:48 - STATUS - --Done!--
[ReActor] 19:26:48 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:26:56 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,553,085,362
VRAMdebug: free memory after:  5,553,085,362
VRAMdebug: freed memory:  0
Prompt executed in 50.14 seconds
got prompt
loaded completely 7310.373362350464 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:34<00:00,  1.74it/s]
[ReActor] 19:27:39 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:27:42 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:27:42 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 19:27:42 - STATUS - Analyzing Target Image...
[ReActor] 19:27:42 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:27:42 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:27:42 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:27:42 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:27:42 - STATUS - Swapping...
[ReActor] 19:27:42 - STATUS - Face Boost is enabled
[ReActor] 19:27:42 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:27:43 - STATUS - --Done!--
[ReActor] 19:27:43 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:27:51 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,843,540,914
VRAMdebug: free memory after:  5,843,540,914
VRAMdebug: freed memory:  0
Prompt executed in 52.60 seconds
got prompt
loaded completely 7312.373362350464 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:35<00:00,  1.67it/s]
[ReActor] 19:28:36 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:28:39 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:28:39 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 19:28:39 - STATUS - Analyzing Target Image...
[ReActor] 19:28:39 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:28:39 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:28:39 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:28:39 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:28:39 - STATUS - Swapping...
[ReActor] 19:28:39 - STATUS - Face Boost is enabled
[ReActor] 19:28:39 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:28:39 - STATUS - --Done!--
[ReActor] 19:28:39 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:28:52 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,885,483,954
VRAMdebug: free memory after:  5,885,483,954
VRAMdebug: freed memory:  0
Prompt executed in 59.41 seconds
got prompt
loaded completely 7312.373362350464 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:33<00:00,  1.81it/s]
[ReActor] 19:29:35 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:29:38 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:29:38 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 19:29:38 - STATUS - Analyzing Target Image...
[ReActor] 19:29:38 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:29:38 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:29:38 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:29:38 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:29:38 - STATUS - Swapping...
[ReActor] 19:29:38 - STATUS - Face Boost is enabled
[ReActor] 19:29:38 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:29:39 - STATUS - --Done!--
[ReActor] 19:29:39 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:29:54 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,847,735,218
VRAMdebug: free memory after:  5,847,735,218
VRAMdebug: freed memory:  0
Prompt executed in 58.80 seconds
got prompt
loaded completely 7306.373362350464 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:34<00:00,  1.75it/s]
[ReActor] 19:30:39 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.
[ReActor] 19:30:41 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 19:30:41 - STATUS - Using Hashed Source Face(s) Model...
[ReActor] 19:30:41 - STATUS - Analyzing Target Image...
[ReActor] 19:30:41 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 19:30:41 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:30:41 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 19:30:41 - STATUS - OK - Detected Gender matches Condition
[ReActor] 19:30:41 - STATUS - Swapping...
[ReActor] 19:30:41 - STATUS - Face Boost is enabled
[ReActor] 19:30:41 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 19:30:42 - STATUS - --Done!--
[ReActor] 19:30:42 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 19:30:48 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  5,841,443,762
VRAMdebug: free memory after:  5,841,443,762
VRAMdebug: freed memory:  0
Prompt executed in 50.94 seconds
got prompt
loaded completely 7290.373362350464 4897.0483474731445 True
 12%|████████████▎                                                                                            | 7/60 [00:04<00:29,  1.83it/s]


@kairin kairin changed the title I had a hard time using ComfyUI on Windows 11 previously... so this is what I changed... [My workflow includes Facereactor and total on a good day is 150 seconds per image generations. and now it is just 50 seconds on average just because of installing Flash Attention WITHOUT making any special changes to my setup???? I DO NOT GET IT I DON"T UNDERSTAND!?!?!?!?!] I had a hard time using ComfyUI on Windows 11 previously... so this is what I changed... Oct 26, 2024
@kairin
Copy link
Author

kairin commented Oct 27, 2024

https://huggingface.co/microsoft/Phi-3-mini-4k-instruct

from MS's hugging face it mentioned briefly.

flash_attn==2.5.8 <<--- it seems this is the only mention of flash_attn mentioned in any of the documentations...--->>

below i have updated to show my current flash-attn built from source after installing via pip.


How to Use
Phi-3 Mini-4K-Instruct has been integrated in the 4.41.2 version of transformers. The current transformers version can be verified with: pip list | grep transformers.

Examples of required packages:

**_flash_attn==2.5.8  <<--- it seems this is the only mention of flash_attn mentioned in any of the documentations...--->>_**
torch==2.3.1
accelerate==0.31.0
transformers==4.41.2

Phi-3 Mini-4K-Instruct is also available in [Azure AI Studio](https://aka.ms/try-phi3)

Tokenizer
Phi-3 Mini-4K-Instruct supports a vocabulary size of up to 32064 tokens. The [tokenizer files](https://huggingface.co/microsoft/Phi-3-mini-4k-instruct/blob/main/added_tokens.json) already provide placeholder tokens that can be used for downstream fine-tuning, but they can also be extended up to the model's vocabulary size.

@kairin
Copy link
Author

kairin commented Oct 27, 2024

#3265 (comment)

this is the only other person who indicated a possibility on how the flash_attn is being used.

@kairin
Copy link
Author

kairin commented Oct 27, 2024

#3265 (comment)

this is the only other person who indicated a possibility on how the flash_attn is being used.

2 things i gather:

  1. Where did my flash-attn build go? is it somewhere that xformers picked up along the way? what version of xformers do i have?

Try installing the flash-attention-2.3.6 py311 ada / sm_89 (not xformers) wheel from my link on the discussions page you posted on (yeah it's from last December, doesn't seem to matter). xformers builds flash-attn-2 in-tree instead of as a dependency and tosses it in an anonymously named .pyd file that nothing else can use.

  1. if i have an xformers that is not compatible i should be seeing an error too.

Torch (seems) to look for one in the base install via lazy module loading; the error is just misleading. I think I'm still using 2.3.6 because I built a newer version like 2.4.2 around february and tried installing it and something started complaining but xformers is on v2.5.6 now so a newer build might work unless torch requires a specific version.

@kairin
Copy link
Author

kairin commented Oct 27, 2024

let's take a look:


ddd  insightface  ven
Sun Oct 27 09:00:42 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   39C    P0             33W /  320W |     411MiB /  16376MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|  No running processes found                                                             |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/ks/ven
VIRTUAL_ENV_PROMPT=(ven)
ddd  insightface  ven
CODEOWNERS       app               comfyui.prev2.log               folder_paths.py    new_updater.py   requirements.txt  tests-unit
CONTRIBUTING.md  comfy             cuda_malloc.py                  input              node_helpers.py  script_examples   user
LICENSE          comfy_execution   custom_nodes                    latent_preview.py  nodes.py         server.py         utils
README.md        comfy_extras      execution.py                    main.py            notebooks        styles            web
__pycache__      comfyui.log       extra_model_paths.yaml.example  model_filemanager  output           temp
api_server       comfyui.prev.log  fix_torch.py                    models             pytest.ini       tests
  on    bash   ven 3.12.3    master ≡  ?1 -1                                               ╰─❯
╭─ 09:00:42 |  27 Oct, Sunday |   in     ven  comf
╰─❯
╰─❯ nvidia-smi
Sun Oct 27 09:04:55 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.03              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   40C    P8              3W /  320W |     500MiB /  16376MiB |      1%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A       447      G   /Xwayland                                   N/A      |
+-----------------------------------------------------------------------------------------+
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                244ms 
╭─ 09:04:55 |  27 Oct, Sunday |   in     ven  comf
╰─❯ xformers --version
-bash: xformers: command not found
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                 19ms 
╭─ 09:23:18 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show xformers
Name: xformers
Version: 0.0.28.post2
Summary: XFormers: A collection of composable Transformer building blocks.
Home-page: https://facebookresearch.github.io/xformers/
Author: Facebook AI Research
Author-email: [email protected]
License:
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: numpy, torch
Required-by:
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                684ms 
╭─ 09:23:31 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show flash
WARNING: Package(s) not found: flash
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                202ms 
╭─ 09:23:56 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show flash-attn
Name: flash_attn
Version: 2.6.3
Summary: Flash Attention: Fast and Memory-Efficient Exact Attention
Home-page: https://github.com/Dao-AILab/flash-attention
Author: Tri Dao
Author-email: [email protected]
License:
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: einops, torch
Required-by:
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                422ms 
╭─ 09:24:02 |  27 Oct, Sunday |   in     ven  comf
╰─❯

@kairin
Copy link
Author

kairin commented Oct 27, 2024


  on    bash   ven 3.12.3    master ≡  ?1 -1                                                202ms 
╭─ 09:23:56 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show flash-attn
Name: flash_attn
Version: 2.6.3
Summary: Flash Attention: Fast and Memory-Efficient Exact Attention
Home-page: https://github.com/Dao-AILab/flash-attention
Author: Tri Dao
Author-email: [email protected]
License:
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: einops, torch
Required-by:
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                422ms 
╭─ 09:24:02 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show einops
Name: einops
Version: 0.8.0
Summary: A new flavour of deep learning operations
Home-page: https://github.com/arogozhnikov/einops
Author: Alex Rogozhnikov
Author-email:
License: MIT
Location: /home/ks/ven/lib/python3.12/site-packages
Requires:
Required-by: flash_attn, spandrel
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                408ms 
╭─ 09:25:13 |  27 Oct, Sunday |   in     ven  comf
╰─❯

@kairin
Copy link
Author

kairin commented Oct 27, 2024


  on    bash   ven 3.12.3    master ≡  ?1 -1                                                279ms 
╭─ 09:26:16 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show flash-attn
Name: flash_attn
Version: 2.6.3
Summary: Flash Attention: Fast and Memory-Efficient Exact Attention
Home-page: https://github.com/Dao-AILab/flash-attention
Author: Tri Dao
Author-email: [email protected]
License:
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: einops, torch
Required-by:
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                390ms 
╭─ 09:26:21 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show torch
Name: torch
Version: 2.5.0
Summary: Tensors and Dynamic neural networks in Python with strong GPU acceleration
Home-page: https://pytorch.org/
Author: PyTorch Team
Author-email: [email protected]
License: BSD-3-Clause
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: filelock, fsspec, jinja2, networkx, nvidia-cublas-cu12, nvidia-cuda-cupti-cu12, nvidia-cuda-nvrtc-cu12, nvidia-cuda-runtime-cu12, nvidia-cudnn-cu12, nvidia-cufft-cu12, nvidia-curand-cu12, nvidia-cusolver-cu12, nvidia-cusparse-cu12, nvidia-nccl-cu12, nvidia-nvjitlink-cu12, nvidia-nvtx-cu12, setuptools, sympy, triton, typing-extensions
Required-by: accelerate, bitsandbytes, fairscale, flash_attn, kornia, lpips, open_clip_torch, pytorch-lightning, spandrel, timm, torch_tensorrt, torchaudio, torchmetrics, torchsde, torchvision, ultralytics, ultralytics-thop, xformers
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                456ms 
╭─ 09:26:27 |  27 Oct, Sunday |   in     ven  comf
╰─❯ pip show accelerate
Name: accelerate
Version: 1.0.1
Summary: Accelerate
Home-page: https://github.com/huggingface/accelerate
Author: The HuggingFace team
Author-email: [email protected]
License: Apache
Location: /home/ks/ven/lib/python3.12/site-packages
Requires: huggingface-hub, numpy, packaging, psutil, pyyaml, safetensors, torch
Required-by:
  on    bash   ven 3.12.3    master ≡  ?1 -1                                                436ms 
╭─ 09:27:18 |  27 Oct, Sunday |   in     ven  comf
╰─❯

@kairin
Copy link
Author

kairin commented Oct 27, 2024

from here... my assumption is:

torch, accelerate and xformers will somehow figure out where flash-attn is, and somehow would use it automatically. contributing to the speed gain that i am experiencing now.

i have no way of confirming this.

can someone point me in the right direction to confirm?

image

@kairin
Copy link
Author

kairin commented Oct 27, 2024

https://huggingface.co/docs/transformers/en/perf_infer_gpu_one

i feel dumb.

huggingface have descriptions on support for flash_attn.

FlashAttention-2 can only be used when the model’s dtype is fp16 or bf16. Make sure to cast your model to the appropriate dtype and load them on a supported device before using FlashAttention-2.

You can also set use_flash_attention_2=True to enable FlashAttention-2 but it is deprecated in favor of attn_implementation="flash_attention_2".

@kairin
Copy link
Author

kairin commented Oct 27, 2024

FlashAttention-2 can only be used when the model’s dtype is fp16 or bf16. Make sure to cast your model to the appropriate dtype and load them on a supported device before using FlashAttention-2.

You can also set use_flash_attention_2=True to enable FlashAttention-2 but it is deprecated in favor of attn_implementation="flash_attention_2".

i guess the below kinda answers my question.

import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("facebook/opt-350m")
model = AutoModelForCausalLM.from_pretrained("facebook/opt-350m", torch_dtype=torch.float16).to("cuda")

input_text = "Hello my dog is cute and"
inputs = tokenizer(input_text, return_tensors="pt").to("cuda")

  • with torch.backends.cuda.sdp_kernel(enable_flash=True, enable_math=False, enable_mem_efficient=False):
    outputs = model.generate(**inputs)

print(tokenizer.decode(outputs[0], skip_special_tokens=True))
If you see a bug with the traceback below, try using the nightly version of PyTorch which may have broader coverage for FlashAttention:

Copied
RuntimeError: No available kernel. Aborting execution.

install PyTorch nightly

pip3 install -U --pre torch torchvision torchaudio --index-url https://download.pytorch.org/whl/nightly/cu118
BetterTransformer

@kairin
Copy link
Author

kairin commented Oct 27, 2024

@kairin
Copy link
Author

kairin commented Oct 27, 2024

https://huggingface.co/docs/transformers/en/perf_infer_gpu_one

i feel dumb.

huggingface have descriptions on support for flash_attn.

FlashAttention-2 can only be used when the model’s dtype is fp16 or bf16. Make sure to cast your model to the appropriate dtype and load them on a supported device before using FlashAttention-2.

You can also set use_flash_attention_2=True to enable FlashAttention-2 but it is deprecated in favor of attn_implementation="flash_attention_2".
BetterTransformer also converts all attention operations to use the more memory-efficient [scaled dot product attention (SDPA)](https://pytorch.org/docs/master/generated/torch.nn.functional.scaled_dot_product_attention), and it calls optimized kernels like [FlashAttention](https://huggingface.co/papers/2205.14135) under the hood.

okay so someone has made this job easier for me. so this part answers the question how does flash_attn gets loaded into comfyUI when i did not make any changes to my workflow.

so apparently they have updated in pytorch. installing updated pytorch is then key.

https://pytorch.org/blog/a-better-transformer-for-fast-transformer-encoder-inference/

-->> Attention Is All You Need <<--

@kairin
Copy link
Author

kairin commented Oct 27, 2024

https://github.com/pytorch/pytorch/tree/main/aten/src/ATen/native/transformers/cuda/flash_attn

so this part will call flash_attn.

pytorch/pytorch@d41558f#diff-d1bfb425f2c653ff16c5f553eec51cae8be05c881259ec72f84fbd7d929f92b0R70-R80

#ifdef USE_FLASH_ATTENTION
// FlashAttention Specific Imports
#include <ATen/native/transformers/cuda/flash_attn/flash_api.h>
#endif
#ifdef USE_MEM_EFF_ATTENTION
#ifndef USE_ROCM
// MemoryEfficient Attention Specific Imports for CUDA
#include <ATen/native/transformers/cuda/mem_eff_attention/kernel_forward.h>
#include <ATen/native/transformers/cuda/mem_eff_attention/kernels/cutlassF.h>
#include <ATen/native/transformers/cuda/mem_eff_attention/pytorch_utils.h>
#else

ok ok... i'm just surprised that this is not a common knowledge kind of thing... judging from not seeing much information about enabling flash_attention when using ComfyUI.

or am I the one who is late to this game?

@kairin
Copy link
Author

kairin commented Oct 31, 2024

Dao-AILab/flash-attention#1308

let's see if updating windows and updating drivers cause any issues...

first... sudo apt update && sudo apt upgrade broke a lot of things...

@kairin
Copy link
Author

kairin commented Oct 31, 2024


ddd  insightface  ven
Thu Oct 31 20:24:45 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.57.01              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   39C    P0              9W /  320W |     724MiB /  16376MiB |     10%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A       438      G   /Xwayland                                   N/A      |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/ks/ven
VIRTUAL_ENV_PROMPT=(ven)
ddd  insightface  ven
CODEOWNERS       app               comfyui.prev2.log               fix_torch.py       models           pytest.ini        tests
CONTRIBUTING.md  comfy             comfyui_screenshot.png          folder_paths.py    new_updater.py   requirements.txt  tests-unit
LICENSE          comfy_execution   cuda_malloc.py                  input              node_helpers.py  script_examples   user
README.md        comfy_extras      custom_nodes                    latent_preview.py  nodes.py         server.py         utils
__pycache__      comfyui.log       execution.py                    main.py            notebooks        styles            web
api_server       comfyui.prev.log  extra_model_paths.yaml.example  model_filemanager  output           temp
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                            0ms ╭─ 20:24:45 |  31 Oct, Thursday |   in     ven  comf
╰─❯ python3 -m pip install --upgrade pip
Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Requirement already satisfied: pip in /home/ks/ven/lib/python3.12/site-packages (24.3.1)
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                       2s 403ms 
╭─ 20:25:06 |  31 Oct, Thursday |   in     ven  comf
╰─❯ pip check
ninja 1.11.1.1 is not supported on this platform
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                          518ms 
╭─ 20:25:14 |  31 Oct, Thursday |   in     ven  comf
╰─❯ pip install ninja
Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Requirement already satisfied: ninja in /home/ks/ven/lib/python3.12/site-packages (1.11.1.1)
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                       1s 186ms 
╭─ 20:25:26 |  31 Oct, Thursday |   in     ven  comf
╰─❯ pip check
ninja 1.11.1.1 is not supported on this platform
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                          477ms 
╭─ 20:25:32 |  31 Oct, Thursday |   in     ven  comf
╰─❯ sudo apt install ninja-build
[sudo] password for ks:
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
ninja-build is already the newest version (1.11.1-2).
0 upgraded, 0 newly installed, 0 to remove and 0 not upgraded.
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                       2s 996ms 
╭─ 20:25:52 |  31 Oct, Thursday |   in     ven  comf
╰─❯ ninja --version
1.12.1
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                           32ms 
╭─ 20:26:14 |  31 Oct, Thursday |   in     ven  comf
╰─❯ ninja
ninja: error: loading 'build.ninja': No such file or directory
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                            0ms 
╭─ 20:26:14 |  31 Oct, Thursday |   in     ven  comf
╰─❯ ls
CODEOWNERS       app               comfyui.prev2.log               fix_torch.py       models           pytest.ini        tests
CONTRIBUTING.md  comfy             comfyui_screenshot.png          folder_paths.py    new_updater.py   requirements.txt  tests-unit
LICENSE          comfy_execution   cuda_malloc.py                  input              node_helpers.py  script_examples   user
README.md        comfy_extras      custom_nodes                    latent_preview.py  nodes.py         server.py         utils
__pycache__      comfyui.log       execution.py                    main.py            notebooks        styles            web
api_server       comfyui.prev.log  extra_model_paths.yaml.example  model_filemanager  output           temp
  on    bash   ven 3.12.3    master ≡  ?1  1                                                                            5ms 
╭─ 20:26:21 |  31 Oct, Thursday |   in     ven  comf
╰─❯ python main.py
[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-31 20:26:27.471142
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/ven/bin/python
** ComfyUI Path: /home/ks/ven/comf
** Log path: /home/ks/ven/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.8)
### ComfyUI Revision: 2790 [9ee0a655] | Released on '2024-10-27'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.5.0+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 magnificent nodes. 🎉

WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ven/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Don't be afraid to give up the good to go for the great." - John D. Rockefeller

/home/ks/ven/comf/custom_nodes/blibla-comfyui-extensions/extensions
/home/ks/ven/comf/web/extensions/failfast-comfyui-extensions
FaceDetailer: Model directory already exists
FaceDetailer: Model already exists

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NodeAligner
   0.0 seconds: /home/ks/ven/comf/custom_nodes/cg-use-everywhere
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-ZeroShot-MTrans
   0.0 seconds: /home/ks/ven/comf/custom_nodes/Skimmed_CFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-AutomaticCFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ven/comf/custom_nodes/blibla-comfyui-extensions
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-browser
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
   0.0 seconds: /home/ks/ven/comf/custom_nodes/facerestore_cf
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_ExtraModels
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_cspnodes
   0.1 seconds: /home/ks/ven/comf/custom_nodes/comfyui-reactor-node
   0.3 seconds: /home/ks/ven/comf/custom_nodes/DZ-FaceDetailer
   0.3 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_omost
   0.3 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
   0.6 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Phi-3-mini
   0.6 seconds: /home/ks/ven/comf/custom_nodes/was-node-suite-comfyui
   0.8 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR
   1.9 seconds: /home/ks/ven/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
Update ComfyUI
Install: pip packages
[ComfyUI-Manager] skip black listed pip installation: 'torch'
[ComfyUI-Manager] skip black listed pip installation: 'torchsde'
[ComfyUI-Manager] skip black listed pip installation: 'torchvision'

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'torchaudio']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'einops']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'transformers>=4.28.1']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'tokenizers>=0.13.3']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'sentencepiece']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'safetensors>=0.4.2']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'aiohttp']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'pyyaml']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'Pillow']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'scipy']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'tqdm']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'psutil']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
 Line is comment...skipping

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'kornia>=0.7.1']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'spandrel']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'soundfile']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
   Update ComfyUI
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/.cache/1514988643_custom-node-list.json [DONE]
Updating: /home/ks/ven/comf/custom_nodes/Skimmed_CFGInstall: pip packages

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'GitPython']
Updating: /home/ks/ven/comf/custom_nodes/rgthree-comfy Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'PyGithub']
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updated: /home/ks/ven/comf/custom_nodes/cg-use-everywhere
Updating: /home/ks/ven/comf/custom_nodes/comfyui-browserInstall: pip packages

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'pillow>=10.3.0']
Updating: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'matrix-client==0.4.0']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
  Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'scipy']
Updating: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap[ComfyUI-Manager] skip black listed pip installation: 'transformers'

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'huggingface-hub>0.20']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
   Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI_cspnodesInstall: pip packages

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'piexif']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'color-matcher']
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI_omost Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updated: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
Updating: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'matplotlib']
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'typer']
Updating: /home/ks/ven/comf/custom_nodes/ComfyUI-NodeAligner Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
            Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'huggingface_hub']

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'rich']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
             Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'mss']

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'typing-extensions']
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
  Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com

## ComfyUI-Manager: EXECUTE => ['/home/ks/ven/bin/python', '-m', 'pip', 'install', 'opencv-python']
Updated: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
 Looking in indexes: https://pypi.org/simple, https://pypi.ngc.nvidia.com
Updated: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
Update done.

Restarting... [Legacy Mode]


[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-10-31 20:28:01.098550
** Platform: Linux
** Python version: 3.12.3 (main, Sep 11 2024, 14:17:37) [GCC 13.2.0]
** Python executable: /home/ks/ven/bin/python
** ComfyUI Path: /home/ks/ven/comf
** Log path: /home/ks/ven/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using xformers cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)
### Loading: ComfyUI-Manager (V2.51.9)
### ComfyUI Revision: 2801 [f2aaa0a4] | Released on '2024-10-31'
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
Torch version: 2.5.0+cu124
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.0+cu124
xformers version: 0.0.28.post2
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 epic nodes. 🎉

WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/ks/ven/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Art is not freedom from discipline, but disciplined freedom." - John F. Kennedy

/home/ks/ven/comf/custom_nodes/blibla-comfyui-extensions/extensions
/home/ks/ven/comf/web/extensions/failfast-comfyui-extensions
FaceDetailer: Model directory already exists
FaceDetailer: Model already exists

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NodeAligner
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-ZeroShot-MTrans
   0.0 seconds: /home/ks/ven/comf/custom_nodes/Skimmed_CFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/cg-use-everywhere
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-instantId-faceswap
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-AutomaticCFG
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/ks/ven/comf/custom_nodes/blibla-comfyui-extensions
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Custom-Scripts
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-browser
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/ks/ven/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui-faceless-node
   0.0 seconds: /home/ks/ven/comf/custom_nodes/facerestore_cf
   0.0 seconds: /home/ks/ven/comf/custom_nodes/comfyui_face_parsing
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-NuA-FlashFace
   0.0 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-KJNodes
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_ExtraModels
   0.1 seconds: /home/ks/ven/comf/custom_nodes/comfyui-reactor-node
   0.1 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_cspnodes
   0.2 seconds: /home/ks/ven/comf/custom_nodes/DZ-FaceDetailer
   0.2 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Facerestore-Tensorrt
   0.2 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI_omost
   0.3 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-Phi-3-mini
   0.5 seconds: /home/ks/ven/comf/custom_nodes/was-node-suite-comfyui
   0.5 seconds: /home/ks/ven/comf/custom_nodes/ComfyUI-SUPIR
   1.2 seconds: /home/ks/ven/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/ks/ven/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]
got prompt
model weight dtype torch.float16, manual cast: None
model_type EPS
Using xformers attention in VAE
Using xformers attention in VAE
[ReActor] 20:28:36 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512


0: 640x480 1 face, 39.1ms
0: 640x480 1 face, 39.1ms
Speed: 2.1ms preprocess, 39.1ms inference, 11.7ms postprocess per image at shape (1, 3, 640, 480)
Speed: 2.1ms preprocess, 39.1ms inference, 11.7ms postprocess per image at shape (1, 3, 640, 480)
Requested to load SDXLClipModel
Loading 1 new model
loaded completely 0.0 1560.802734375 True
Requested to load SDXL
Loading 1 new model
loaded completely 0.0 4897.0483474731445 True
100%|████████████████████████████████████████████████████████████████████████████████████████████████████████| 60/60 [00:32<00:00,  1.86it/s]
Requested to load AutoencoderKL
Loading 1 new model
loaded completely 0.0 159.55708122253418 True
[ReActor] 20:29:21 - STATUS - Restoring with GFPGANv1.4.pth | Face Size is set to 512
Warning torch.load doesn't support weights_only on this pytorch version, loading unsafely.


# redacted information 


[ReActor] 20:29:24 - STATUS - Working: source face index [0], target face index [0]
[ReActor] 20:29:24 - STATUS - Analyzing Source Image...
[ReActor] 20:29:25 - STATUS - Trying to halve 'det_size' parameter
[ReActor] 20:29:29 - STATUS - Analyzing Target Image...
[ReActor] 20:29:32 - STATUS - Source Face 0: Detected Gender -M-
[ReActor] 20:29:32 - STATUS - OK - Detected Gender matches Condition
[ReActor] 20:29:33 - STATUS - Target Face 0: Detected Gender -M-
[ReActor] 20:29:33 - STATUS - OK - Detected Gender matches Condition
[ReActor] 20:29:33 - STATUS - Swapping...
[ReActor] 20:29:33 - STATUS - Face Boost is enabled
[ReActor] 20:29:34 - STATUS - Boosting the Face with codeformer-v0.1.0.pth | Face Size is set to 512 with Scale Factor = 4.0 and 'Bilinear' interpolation
[ReActor] 20:29:35 - STATUS - --Done!--
[ReActor] 20:29:35 - STATUS - Restoring with GPEN-BFR-2048.onnx | Face Size is set to 2048
[ReActor] 20:29:54 - STATUS - Restoring with GPEN-BFR-512.onnx | Face Size is set to 512
VRAMdebug: free memory before:  2,378,621,714
VRAMdebug: free memory after:  2,387,591,954
VRAMdebug: freed memory:  8,970,240
Saved workflow to /home/ks/ven/comf/output/2024-10-31XXXXXXX.json
Prompt executed in 85.32 seconds

@kairin
Copy link
Author

kairin commented Nov 2, 2024

let's delete everything and start all over again to confirm that yes, there is a speed difference before installing flash_attn...

also... i needed conda to help install a software that is obscure... let's do this.

Screenshot 2024-11-02 155309

since there is no space for me to setup a conda environment... i think the next best step for me is to simply delete everything and start from scratch.

Screenshot 2024-11-02 170035

okay. so we are deleting the original virtual environment where i have installed ComfyUI.


Create an environment containing the package 'sqlite'::

    conda create -n myenv sqlite

Create an environment (env2) as a clone of an existing environment (env1)::

    conda create -n env2 --clone path/to/file/env1
  on    bash   base 3.12.7                                                                  225ms 
╭─ 15:49:57 |  2 Nov, Saturday |   in     ven
╰─❯ conda create -n ven2 --clone /home/ks/ven
Source:      /home/ks/ven
Destination: /home/ks/.conda/envs/ven2
Packages: 0
Files: 175465

NoSpaceLeftError: No space left on devices.

  on    bash   base 3.12.7                                                              11s 563ms 
╭─ 15:52:42 |  2 Nov, Saturday |   in     ven
╰─❯ cd ..
  on    bash                                                                                    3ms 
╭─ 17:04:32 |  2 Nov, Saturday |   in  
╰─❯ ls
comf  ddd  insightface  miniconda3  ven
  on    bash                                                                                    7ms 
╭─ 17:04:33 |  2 Nov, Saturday |   in  
╰─❯ sudo rm -r ven
[sudo] password for ks:
  on    bash                                                                               2s 251ms 
╭─ 17:04:46 |  2 Nov, Saturday |   in  
╰─❯ ls
comf  ddd  insightface  miniconda3
  on    bash                                                                                    5ms 
╭─ 17:04:48 |  2 Nov, Saturday |   in  
╰─❯ dir
comf  ddd  insightface  miniconda3
  on    bash                                                                                   10ms 
╭─ 17:04:49 |  2 Nov, Saturday |   in  
╰─❯ ls
comf  ddd  insightface  miniconda3
  on    bash                                                                                    5ms 
╭─ 17:04:51 |  2 Nov, Saturday |   in  
╰─❯

comf  ddd  insightface  miniconda3
Sat Nov  2 17:13:21 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.57.01              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   42C    P5             31W /  320W |     566MiB /  16376MiB |      2%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|  No running processes found                                                             |
+-----------------------------------------------------------------------------------------+
comf  ddd  insightface  miniconda3
-bash: cd: ven/comf/: No such file or directory
comf  ddd  insightface  miniconda3
  on    bash                                                                                                                   0ms 
╭─ 17:13:23 |2 Nov, Saturday |in  
╰─❯

image

@kairin
Copy link
Author

kairin commented Nov 3, 2024

first, uninstall everything and start from fresh.

 winget install wsl
Multiple packages found matching input criteria. Please refine the input.
Name                        Id            Source
-------------------------------------------------
Arch WSL                    9MZNMNKSM73X  msstore
Windows Subsystem for Linux Microsoft.WSL winget
 winget install Microsoft.WSL
Found Windows Subsystem for Linux [Microsoft.WSL] Version 2.1.5.0
This application is licensed to you by its owner.
Microsoft is not responsible for, nor does it grant any licenses to, third-party packages.
This package requires the following dependencies:
  - Windows Features
      Microsoft-Windows-Subsystem-Linux
      VirtualMachinePlatform
Enabling [Microsoft-Windows-Subsystem-Linux]...
Enabling [VirtualMachinePlatform]...
Successfully enabled Windows Features dependencies
Downloading https://github.com/microsoft/WSL/releases/download/2.1.5/wsl.2.1.5.0.x64.msi
  ██████████                      43.0 MB /  127 MB

@kairin
Copy link
Author

kairin commented Nov 3, 2024

image

i reset everything and reinstalled from scratch.

image

and to save space... i am moving everything i can elsewhere using system settings in windows 11.

PowerShell 7.5.0-preview.5
Loading personal and system profiles took 774ms.
 wsl --list --all
Windows Subsystem for Linux has no installed distributions.                                                             
Use 'wsl.exe --list --online' to list available distributions
and 'wsl.exe --install <Distro>' to install.

Distributions can also be installed by visiting the Microsoft Store:
https://aka.ms/wslstore
Error code: Wsl/WSL_E_DEFAULT_DISTRO_NOT_FOUND
 wsl --list --online
The following is a list of valid distributions that can be installed.
Install using 'wsl.exe --install <Distro>'.                                                                             
NAME                            FRIENDLY NAME
Ubuntu                          Ubuntu
Debian                          Debian GNU/Linux
kali-linux                      Kali Linux Rolling
Ubuntu-18.04                    Ubuntu 18.04 LTS
Ubuntu-20.04                    Ubuntu 20.04 LTS
Ubuntu-22.04                    Ubuntu 22.04 LTS
Ubuntu-24.04                    Ubuntu 24.04 LTS
OracleLinux_7_9                 Oracle Linux 7.9
OracleLinux_8_7                 Oracle Linux 8.7
OracleLinux_9_1                 Oracle Linux 9.1
openSUSE-Leap-15.6              openSUSE Leap 15.6
SUSE-Linux-Enterprise-15-SP5    SUSE Linux Enterprise 15 SP5
SUSE-Linux-Enterprise-15-SP6    SUSE Linux Enterprise 15 SP6
openSUSE-Tumbleweed             openSUSE Tumbleweed
 wsl --install Ubuntu-24.04
Installing: Ubuntu 24.04 LTS
[=====                      9.0%                           ]



image

you can move your files accordingly so do that if you want to save space or whatever.

and make sure you update accordingly... and set sparse if you need to...

PowerShell 7.5.0-preview.5
Loading personal and system profiles took 829ms.
 winget update --all
No installed package found matching input criteria.
 wsl --update
Checking for updates.
Updating Windows Subsystem for Linux to version: 2.3.24.

.
.
.
.
.

 wsl --manage Ubuntu-24.04 --set-sparse true
Conversion in progress, this may take a few minutes.
The operation completed successfully.

   pwsh MEM: 14% | 11/79GB   41ms 
╭─ ♥ 05:34 |   
╰─

and for windows users i strongly recommend installing xfe so that you don't have to worry too much about using the command line to get things done...

ks@4080SUPER:~$ curl -s https://ohmyposh.dev/install.sh | bash -s
unzip is required to install Oh My Posh. Please install unzip and try again.

ks@4080SUPER:~$ sudo apt install unzip
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
Suggested packages:
  zip
The following NEW packages will be installed:
  unzip
0 upgraded, 1 newly installed, 0 to remove and 0 not upgraded.
Need to get 174 kB of archives.
After this operation, 384 kB of additional disk space will be used.
Get:1 http://archive.ubuntu.com/ubuntu noble-updates/main amd64 unzip amd64 6.0-28ubuntu4.1 [174 kB]
Fetched 174 kB in 1s (182 kB/s)
Selecting previously unselected package unzip.
(Reading database ... 40787 files and directories currently installed.)
Preparing to unpack .../unzip_6.0-28ubuntu4.1_amd64.deb ...
Unpacking unzip (6.0-28ubuntu4.1) ...
Setting up unzip (6.0-28ubuntu4.1) ...
Processing triggers for man-db (2.12.0-4build2) ...
ks@4080SUPER:~$ curl -s https://ohmyposh.dev/install.sh | bash -s
⚠️  Installation directory /home/ks/.local/bin is not in your $PATH, add it using
export PATH=$PATH:/home/ks/.local/bin

ℹ️  Installing oh-my-posh for linux-amd64 in /home/ks/.local/bin
⬇️  Downloading oh-my-posh from https://github.com/JanDeDobbeleer/oh-my-posh/releases/latest/download/posh-linux-amd64
🎨 Installing oh-my-posh themes in /home/ks/.cache/oh-my-posh/themes

🚀 Installation complete.

You can follow the instructions at https://ohmyposh.dev/docs/installation/prompt
to setup your shell to use oh-my-posh.

If you want to use a built-in theme, you can find them in the /home/ks/.cache/oh-my-posh/themes directory:
  oh-my-posh init {shell} --config /home/ks/.cache/oh-my-posh/themes/{theme}.omp.json

ks@4080SUPER:~$ sudo apt install xfe
Reading package lists... Done
Building dependency tree... Done
Reading state information... Done
The following additional packages will be installed:

image

image

turns out, xfe does pretty much how a windows user expects files to be moved around without much issues or headaches. i like it!

also, use miniconda. turns out... it is a much better method to manage your environments... but you do you~!

https://docs.anaconda.com/miniconda/

┏[ ks from  4080SUPER][ 0s][ RAM: 0/39GB][ Monday at 5:51:27 AM]
┖[~]
└─Δ mkdir -p ~/miniconda3
wget https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh -O ~/miniconda3/miniconda.sh
bash ~/miniconda3/miniconda.sh -b -u -p ~/miniconda3
rm ~/miniconda3/miniconda.sh
--2024-11-04 05:52:49--  https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh
Resolving repo.anaconda.com (repo.anaconda.com)... 104.16.32.241, 104.16.191.158, 2606:4700::6810:bf9e, ...
Connecting to repo.anaconda.com (repo.anaconda.com)|104.16.32.241|:443... connected.
HTTP request sent, awaiting response... 200 OK
Length: 148337011 (141M) [application/octet-stream]
Saving to: /home/ks/miniconda3/miniconda.sh

/home/ks/miniconda3/miniconda.sh    100%[================================================================>] 141.46M  10.9MB/s    in 13s

2024-11-04 05:53:01 (10.9 MB/s) - /home/ks/miniconda3/miniconda.sh saved [148337011/148337011]

PREFIX=/home/ks/miniconda3
Unpacking payload ...

Installing base environment...

Preparing transaction: ...working... done
Executing transaction: ...working... done
installation finished.
┏[ ks from  4080SUPER][ 0.035s][ RAM: 1/39GB][ Monday at 5:53:07 AM]
┖[~]
└─Δ

and don't forget that conda should be the preferred method to install as much of the dependencies as possible. for eg, pytorch:

┏[ ks from  4080SUPER][ 0.004s][ RAM: 0/39GB][ Monday at 6:37:33 AM]
┖[~]
└─Δ conda install pytorch torchvision torchaudio pytorch-cuda=12.4 -c pytorch-nightly -c nvidia
Channels:
 - pytorch-nightly
 - nvidia
 - defaults
Platform: linux-64
Collecting package metadata (repodata.json): done
Solving environment: done

## Package Plan ##

  environment location: /home/ks/ven

  added / updated specs:
    - pytorch
    - pytorch-cuda=12.4
    - torchaudio
    - torchvision
.
.
.
.
.
.

after installing all the relevant softwares and dependencies

.   .bash_history  .bashrc  .conda    .config  .motd_shown  .profile                   .tcshrc   .zshrc      ven
..  .bash_logout   .cache   .condarc  .local   .nv          .sudo_as_admin_successful  .xonshrc  miniconda3
Mon Nov  4 07:24:24 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.51.01              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   40C    P0             13W /  320W |     685MiB /  16376MiB |      7%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|  No running processes found                                                             |
+-----------------------------------------------------------------------------------------+
.   .bash_history  .bashrc  .conda    .config  .motd_shown  .profile                   .tcshrc   .zshrc      ven
..  .bash_logout   .cache   .condarc  .local   .nv          .sudo_as_admin_successful  .xonshrc  miniconda3
.               .requirements.txt.swp  comfy                           fix_torch.py       node_helpers.py   temp
..              CODEOWNERS             comfy_execution                 folder_paths.py    nodes.py          tests
.ci             CONTRIBUTING.md        comfy_extras                    input              notebooks         tests-unit
.git            LICENSE                comfyui_screenshot.png          latent_preview.py  output            user
.gitattributes  README.md              cuda_malloc.py                  main.py            pytest.ini        utils
.github         __pycache__            custom_nodes                    model_filemanager  requirements.txt  web
.gitignore      api_server             execution.py                    models             script_examples
.pylintrc       app                    extra_model_paths.yaml.example  new_updater.py     server.py
Use the following command to check which environment you have activated: conda info --envs
Use the following command to check which environment you have activated: conda activate <FOLLOWED BY THE env NAME from above>
-bash: /home/ks/.cargo/env: No such file or directory
┏[ ks from  4080SUPER][ 0s][ RAM: 16/39GB][ Monday at 7:24:25 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ conda activate ~/ven
┏[ ks from  4080SUPER][ 0.215s][ RAM: 16/39GB][ Monday at 7:24:36 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ conda info --envs
# conda environments:
#
base                     /home/ks/miniconda3
                      *  /home/ks/ven

┏[ ks from  4080SUPER][ 0.626s][ RAM: 16/39GB][ Monday at 7:24:45 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ pip install torchaudio
Requirement already satisfied: torchaudio in /home/ks/ven/lib/python3.12/site-packages (2.5.0.dev20241103)
Requirement already satisfied: torch in /home/ks/ven/lib/python3.12/site-packages (from torchaudio) (2.5.1)
Requirement already satisfied: filelock in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (3.13.1)
Requirement already satisfied: typing-extensions>=4.8.0 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (4.11.0)
Requirement already satisfied: networkx in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (3.2.1)
Requirement already satisfied: jinja2 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (3.1.4)
Requirement already satisfied: fsspec in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (2024.3.1)
Requirement already satisfied: nvidia-cuda-nvrtc-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cuda-runtime-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cuda-cupti-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cudnn-cu12==9.1.0.70 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (9.1.0.70)
Requirement already satisfied: nvidia-cublas-cu12==12.4.5.8 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.5.8)
Requirement already satisfied: nvidia-cufft-cu12==11.2.1.3 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (11.2.1.3)
Requirement already satisfied: nvidia-curand-cu12==10.3.5.147 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (10.3.5.147)
Requirement already satisfied: nvidia-cusolver-cu12==11.6.1.9 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (11.6.1.9)
Requirement already satisfied: nvidia-cusparse-cu12==12.3.1.170 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.3.1.170)
Requirement already satisfied: nvidia-nccl-cu12==2.21.5 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (2.21.5)
Requirement already satisfied: nvidia-nvtx-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-nvjitlink-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (12.4.127)
Requirement already satisfied: triton==3.1.0 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (3.1.0)
Requirement already satisfied: setuptools in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (75.1.0)
Requirement already satisfied: sympy==1.13.1 in /home/ks/ven/lib/python3.12/site-packages (from torch->torchaudio) (1.13.1)
Requirement already satisfied: mpmath<1.4,>=1.1.0 in /home/ks/ven/lib/python3.12/site-packages (from sympy==1.13.1->torch->torchaudio) (1.3.0)
Requirement already satisfied: MarkupSafe>=2.0 in /home/ks/ven/lib/python3.12/site-packages (from jinja2->torch->torchaudio) (2.1.3)
┏[ ks from  4080SUPER][ 0.419s][ RAM: 16/39GB][ Monday at 7:24:59 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ pip uninstal torchaudio
ERROR: unknown command "uninstal" - maybe you meant "uninstall"
┏[ ks from  4080SUPER][ 0.144s][ RAM: 16/39GB][ Monday at 7:25:06 AM][ master ≡  ?1][ Error, check your command]
┖[~/ven/comf]
└─Δ pip uninstall torchaudio
Found existing installation: torchaudio 2.5.0.dev20241103
Uninstalling torchaudio-2.5.0.dev20241103:
  Would remove:
    /home/ks/ven/lib/python3.12/site-packages/torchaudio
    /home/ks/ven/lib/python3.12/site-packages/torchaudio-2.5.0.dev20241103-py3.12.egg-info
    /home/ks/ven/lib/python3.12/site-packages/torio
Proceed (Y/n)? y
  Successfully uninstalled torchaudio-2.5.0.dev20241103
┏[ ks from  4080SUPER][ 0.916s][ RAM: 16/39GB][ Monday at 7:25:12 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ pip install torchaudio
Collecting torchaudio
  Downloading torchaudio-2.5.1-cp312-cp312-manylinux1_x86_64.whl.metadata (6.4 kB)
Requirement already satisfied: torch==2.5.1 in /home/ks/ven/lib/python3.12/site-packages (from torchaudio) (2.5.1)
Requirement already satisfied: filelock in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (3.13.1)
Requirement already satisfied: typing-extensions>=4.8.0 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (4.11.0)
Requirement already satisfied: networkx in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (3.2.1)
Requirement already satisfied: jinja2 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (3.1.4)
Requirement already satisfied: fsspec in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (2024.3.1)
Requirement already satisfied: nvidia-cuda-nvrtc-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cuda-runtime-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cuda-cupti-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-cudnn-cu12==9.1.0.70 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (9.1.0.70)
Requirement already satisfied: nvidia-cublas-cu12==12.4.5.8 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.5.8)
Requirement already satisfied: nvidia-cufft-cu12==11.2.1.3 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (11.2.1.3)
Requirement already satisfied: nvidia-curand-cu12==10.3.5.147 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (10.3.5.147)
Requirement already satisfied: nvidia-cusolver-cu12==11.6.1.9 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (11.6.1.9)
Requirement already satisfied: nvidia-cusparse-cu12==12.3.1.170 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.3.1.170)
Requirement already satisfied: nvidia-nccl-cu12==2.21.5 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (2.21.5)
Requirement already satisfied: nvidia-nvtx-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.127)
Requirement already satisfied: nvidia-nvjitlink-cu12==12.4.127 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (12.4.127)
Requirement already satisfied: triton==3.1.0 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (3.1.0)
Requirement already satisfied: setuptools in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (75.1.0)
Requirement already satisfied: sympy==1.13.1 in /home/ks/ven/lib/python3.12/site-packages (from torch==2.5.1->torchaudio) (1.13.1)
Requirement already satisfied: mpmath<1.4,>=1.1.0 in /home/ks/ven/lib/python3.12/site-packages (from sympy==1.13.1->torch==2.5.1->torchaudio) (1.3.0)
Requirement already satisfied: MarkupSafe>=2.0 in /home/ks/ven/lib/python3.12/site-packages (from jinja2->torch==2.5.1->torchaudio) (2.1.3)
Downloading torchaudio-2.5.1-cp312-cp312-manylinux1_x86_64.whl (3.4 MB)
   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 3.4/3.4 MB 9.6 MB/s eta 0:00:00
Installing collected packages: torchaudio
Successfully installed torchaudio-2.5.1
┏[ ks from  4080SUPER][ 1.144s][ RAM: 16/39GB][ Monday at 7:25:17 AM][ master ≡  ?1]
┖[~/ven/comf]
└─Δ pyhon main.py
pyhon: command not found
┏[ ks from  4080SUPER][ 0.122s][ RAM: 16/39GB][ Monday at 7:25:21 AM][ master ≡  ?1][ Error, check your command]
┖[~/ven/comf]
└─Δ python main.py
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.5.1+cu124
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using pytorch cross attention
[Prompt Server] web root: /home/ks/ven/comf/web
/home/ks/ven/lib/python3.12/site-packages/kornia/feature/lightglue.py:44: FutureWarning: `torch.cuda.amp.custom_fwd(args...)` is deprecated. Please use `torch.amp.custom_fwd(args..., device_type='cuda')` instead.
  @torch.cuda.amp.custom_fwd(cast_inputs=torch.float32)

Import times for custom nodes:
   0.0 seconds: /home/ks/ven/comf/custom_nodes/websocket_image_save.py

Starting server

To see the GUI go to: http://127.0.0.1:8188

image

so installation works fine.

::editted:: i regret installing conda... so many broken dependencies issue... stick to pure python and create your environment in venv the only problem arising is when you need to install other tools in other languages and that's where conda come in. one of the modern attention methods requires a software that can be found in conda library but not pypi... so... will come to that bridge and solve it then. for now... back to python, create env using venv, and stick with pip.

.  ..  .bash_history  .bash_logout  .bashrc  .cache  .config  .local  .motd_shown  .profile  .pyenv  .sudo_as_admin_successful  comf
Running (pyenv virtualenvs) this Lists all Python virtualenvs found in (/home/k/.pyenv/versions/*) in the next line
Running now.
  3.12.7/envs/ven (created from /home/k/.pyenv/versions/3.12.7)
  ven (created from /home/k/.pyenv/versions/3.12.7)
We are running (pyenv activate) to activate the listed environment (ven) next.
Running now
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Fri_Jan__6_16:45:21_PST_2023
Cuda compilation tools, release 12.0, V12.0.140
Build cuda_12.0.r12.0/compiler.32267302_0
Tue Nov  5 23:55:31 2024
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 565.51.01              Driver Version: 565.90         CUDA Version: 12.7     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 4080 ...    On  |   00000000:0B:00.0  On |                  N/A |
|  0%   39C    P0             14W /  320W |     718MiB /  16376MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A       403      G   /Xwayland                                   N/A      |
+-----------------------------------------------------------------------------------------+
VIRTUAL_ENV=/home/k/.pyenv/versions/3.12.7/envs/ven
PYENV_VIRTUAL_ENV=/home/k/.pyenv/versions/3.12.7/envs/ven
  3.12.7/envs/ven (created from /home/k/.pyenv/versions/3.12.7)
* ven (created from /home/k/.pyenv/versions/3.12.7)
.  ..  .bash_history  .bash_logout  .bashrc  .cache  .config  .local  .motd_shown  .profile  .pyenv  .sudo_as_admin_successful  comf
.               .pylintrc        comfy                   extra_model_paths.yaml.example  models           requirements.txt
..              CODEOWNERS       comfy_execution         fix_torch.py                    new_updater.py   script_examples
.ci             CONTRIBUTING.md  comfy_extras            folder_paths.py                 node_helpers.py  server.py
.git            LICENSE          comfyui_screenshot.png  input                           nodes.py         tests
.gitattributes  README.md        cuda_malloc.py          latent_preview.py               notebooks        tests-unit
.github         api_server       custom_nodes            main.py                         output           utils
.gitignore      app              execution.py            model_filemanager               pytest.ini       web
(ven) k@4080SUPER:~/comf$

the next comment will be about installing comfyui and comparing a fresh install with and without flash_attn.

Stay tune~!

@kairin
Copy link
Author

kairin commented Nov 5, 2024

All extensions are already up-to-date.

Restarting... [Legacy Mode]


[START] Security scan
[DONE] Security scan
## ComfyUI-Manager: installing dependencies done.
** ComfyUI startup time: 2024-11-06 00:12:25.402474
** Platform: Linux
** Python version: 3.12.7 (main, Nov  5 2024, 23:19:51) [GCC 13.2.0]
** Python executable: /home/k/.pyenv/versions/ven/bin/python
** ComfyUI Path: /home/k/comf
** Log path: /home/k/comf/comfyui.log

Prestartup times for custom nodes:
   0.0 seconds: /home/k/comf/custom_nodes/rgthree-comfy
   0.6 seconds: /home/k/comf/custom_nodes/ComfyUI-Manager

Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.6.0.dev20241105+cu124
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync
Using pytorch cross attention
[Prompt Server] web root: /home/k/comf/web
WAS Node Suite: OpenCV Python FFMPEG support is enabled
WAS Node Suite Warning: `ffmpeg_bin_path` is not set in `/home/k/comf/custom_nodes/was-node-suite-comfyui/was_suite_config.json` config file. Will attempt to use system ffmpeg binaries if available.
WAS Node Suite: Finished. Loaded 218 nodes successfully.

        "Art is the mirror that reflects the beauty within us." - Unknown

[comfy_mtb] | INFO -> loaded 86 nodes successfuly
[comfy_mtb] | INFO -> Some nodes (2) could not be loaded. This can be ignored, but go to http://127.0.0.1:8188/mtb if you want more information.
Total VRAM 16376 MB, total RAM 40071 MB
pytorch version: 2.6.0.dev20241105+cu124
Set vram state to: NORMAL_VRAM
Device: cuda:0 NVIDIA GeForce RTX 4080 SUPER : cudaMallocAsync

[rgthree-comfy] Loaded 42 epic nodes. 🎉

### Loading: ComfyUI-Manager (V2.51.9)
### ComfyUI Revision: 2811 [8afb97cd] | Released on '2024-11-05'
[ReActor] - STATUS - Running v0.5.1-b2 in ComfyUI
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json
Torch version: 2.6.0.dev20241105+cu124
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json
[ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json

Import times for custom nodes:
   0.0 seconds: /home/k/comf/custom_nodes/websocket_image_save.py
   0.0 seconds: /home/k/comf/custom_nodes/ComfyUI_JPS-Nodes
   0.0 seconds: /home/k/comf/custom_nodes/ComfyUI-Image-Saver
   0.0 seconds: /home/k/comf/custom_nodes/rgthree-comfy
   0.0 seconds: /home/k/comf/custom_nodes/ComfyUI-KJNodes
   0.0 seconds: /home/k/comf/custom_nodes/ComfyUI-Manager
   0.0 seconds: /home/k/comf/custom_nodes/comfyui_face_parsing
   0.2 seconds: /home/k/comf/custom_nodes/comfyui-reactor-node
   1.0 seconds: /home/k/comf/custom_nodes/was-node-suite-comfyui
   1.3 seconds: /home/k/comf/custom_nodes/comfy_mtb

Starting server

To see the GUI go to: http://127.0.0.1:8188
FETCH DATA from: /home/k/comf/custom_nodes/ComfyUI-Manager/extension-node-map.json [DONE]

image

okay. so i will use a freshly installed WSL2 with Ubuntu and a clean fresh new install of ComfyUI with all dependencies for the above image generations first.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
User Support A user needs help with something, probably not a bug.
Projects
None yet
Development

No branches or pull requests

2 participants