Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DLL load failed while importing cuda_utils #19

Open
Nightnightlight opened this issue Dec 4, 2024 · 16 comments
Open

DLL load failed while importing cuda_utils #19

Nightnightlight opened this issue Dec 4, 2024 · 16 comments

Comments

@Nightnightlight
Copy link

I have triton and sageattention installed without errors yet I still just get this error when trying to inference.

ComfyUI Error Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: ImportError
  • Exception Message: DLL load failed while importing cuda_utils: The specified module could not be found.

Stack Trace

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 74, in compile_module_from_src
    mod = importlib.util.module_from_spec(spec)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "<frozen importlib._bootstrap>", line 813, in module_from_spec

  File "<frozen importlib._bootstrap_external>", line 1288, in create_module

  File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed

System Information

  • ComfyUI Version: v0.3.6-13-g3bed56b
  • Arguments: main.py
  • OS: nt
  • Python Version: 3.12.0 (tags/v3.12.0:0fb18b0, Oct 2 2023, 13:03:39) [MSC v.1935 64 bit (AMD64)]
  • Embedded Python: false
  • PyTorch Version: 2.5.1+cu121

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 25769148416
    • VRAM Free: 17778420304
    • Torch VRAM Total: 6643777536
    • Torch VRAM Free: 66923088

Logs

2024-12-04T16:18:24.060971 - [START] Security scan2024-12-04T16:18:24.060971 - 
2024-12-04T16:18:25.327014 - [DONE] Security scan2024-12-04T16:18:25.327014 - 
2024-12-04T16:18:25.444980 - ## ComfyUI-Manager: installing dependencies done.2024-12-04T16:18:25.444980 - 
2024-12-04T16:18:25.445976 - ** ComfyUI startup time:2024-12-04T16:18:25.445976 -  2024-12-04T16:18:25.445976 - 2024-12-04 16:18:25.4459762024-12-04T16:18:25.445976 - 
2024-12-04T16:18:25.475265 - ** Platform:2024-12-04T16:18:25.475265 -  2024-12-04T16:18:25.475265 - Windows2024-12-04T16:18:25.475265 - 
2024-12-04T16:18:25.475265 - ** Python version:2024-12-04T16:18:25.476066 -  2024-12-04T16:18:25.476066 - 3.12.0 (tags/v3.12.0:0fb18b0, Oct  2 2023, 13:03:39) [MSC v.1935 64 bit (AMD64)]2024-12-04T16:18:25.476066 - 
2024-12-04T16:18:25.476066 - ** Python executable:2024-12-04T16:18:25.476066 -  2024-12-04T16:18:25.476066 - E:\SD\ComfyUI\NewComfy\venv\Scripts\python.exe2024-12-04T16:18:25.476066 - 
2024-12-04T16:18:25.476066 - ** ComfyUI Path:2024-12-04T16:18:25.477065 -  2024-12-04T16:18:25.477065 - E:\SD\ComfyUI\NewComfy\ComfyUI2024-12-04T16:18:25.477065 - 
2024-12-04T16:18:25.477065 - ** Log path:2024-12-04T16:18:25.477065 -  2024-12-04T16:18:25.477065 - E:\SD\ComfyUI\NewComfy\ComfyUI\comfyui.log2024-12-04T16:18:25.478063 - 
2024-12-04T16:18:27.100987 - 
Prestartup times for custom nodes:2024-12-04T16:18:27.100987 - 
2024-12-04T16:18:27.100987 -    3.0 seconds:2024-12-04T16:18:27.101990 -  2024-12-04T16:18:27.101990 - E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-Manager2024-12-04T16:18:27.101990 - 
2024-12-04T16:18:27.101990 - 
2024-12-04T16:18:28.968994 - Total VRAM 24575 MB, total RAM 16335 MB
2024-12-04T16:18:28.968994 - pytorch version: 2.5.1+cu121
2024-12-04T16:18:28.969998 - Set vram state to: NORMAL_VRAM
2024-12-04T16:18:28.970995 - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-12-04T16:18:30.198688 - Using pytorch cross attention
2024-12-04T16:18:31.539013 - [Prompt Server] web root: E:\SD\ComfyUI\NewComfy\ComfyUI\web
2024-12-04T16:18:32.351646 - ### Loading: ComfyUI-Manager (V2.55)2024-12-04T16:18:32.351646 - 
2024-12-04T16:18:32.527020 - ### ComfyUI Version: v0.3.6-13-g3bed56b | Released on '2024-12-04'2024-12-04T16:18:32.527020 - 
2024-12-04T16:18:32.858229 - 
Import times for custom nodes:
2024-12-04T16:18:32.858229 -    0.0 seconds: E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\websocket_image_save.py
2024-12-04T16:18:32.859240 -    0.1 seconds: E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
2024-12-04T16:18:32.859240 -    0.4 seconds: E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper
2024-12-04T16:18:32.860240 -    0.4 seconds: E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-Manager
2024-12-04T16:18:32.860240 - 
2024-12-04T16:18:32.866250 - Starting server

2024-12-04T16:18:32.866250 - To see the GUI go to: http://127.0.0.1:8188
2024-12-04T16:18:32.931107 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json2024-12-04T16:18:32.931107 - 
2024-12-04T16:18:33.006027 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json2024-12-04T16:18:33.006027 - 
2024-12-04T16:18:33.033224 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json2024-12-04T16:18:33.033224 - 
2024-12-04T16:18:33.088223 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json2024-12-04T16:18:33.088223 - 
2024-12-04T16:18:33.118241 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2024-12-04T16:18:33.118241 - 
2024-12-04T16:18:38.255717 - FETCH DATA from: E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2024-12-04T16:18:38.255717 - 2024-12-04T16:18:38.259718 -  [DONE]2024-12-04T16:18:38.259718 - 
2024-12-04T16:18:49.348622 - got prompt
2024-12-04T16:18:49.440161 - The config attributes {'mid_block_causal_attn': True} were passed to AutoencoderKLCausal3D, but are not expected and will be ignored. Please verify your config.json configuration file.
2024-12-04T16:18:52.577161 - Loading text encoder model (clipL) from: E:\SD\ComfyUI\NewComfy\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T16:18:54.607171 - Text encoder to dtype: torch.float16
2024-12-04T16:18:54.782171 - Loading tokenizer (clipL) from: E:\SD\ComfyUI\NewComfy\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T16:18:54.921173 - Loading text encoder model (llm) from: E:\SD\ComfyUI\NewComfy\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T16:20:10.350111 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [01:15<00:00, 16.23s/it]2024-12-04T16:20:10.352111 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [01:15<00:00, 18.80s/it]2024-12-04T16:20:10.352111 - 
2024-12-04T16:21:48.884729 - Text encoder to dtype: torch.float16
2024-12-04T16:22:08.687450 - Loading tokenizer (llm) from: E:\SD\ComfyUI\NewComfy\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T16:22:28.564073 - Using accelerate to load and assign model weights to device...
2024-12-04T16:22:29.168557 - Scheduler config:2024-12-04T16:22:29.174557 -  2024-12-04T16:22:29.174557 - FrozenDict({'num_train_timesteps': 1000, 'shift': 9.0, 'reverse': True, 'solver': 'euler', 'n_tokens': None, '_use_default_values': ['num_train_timesteps', 'n_tokens']})2024-12-04T16:22:29.195562 - 
2024-12-04T16:22:29.305084 - Input (height, width, video_length) = (512, 512, 33)
2024-12-04T16:22:31.776529 - Swapping 20 double blocks and 0 single blocks2024-12-04T16:22:31.776529 - 
2024-12-04T16:22:49.660683 - Sampling 33 frames in 9 latents at 512x512 with 20 inference steps
2024-12-04T16:22:49.708686 - 
  0%|                                                                                           | 0/20 [00:00<?, ?it/s]2024-12-04T16:22:53.954282 - 
  0%|                                                                                           | 0/20 [00:04<?, ?it/s]2024-12-04T16:22:53.954282 - 
2024-12-04T16:22:54.245606 - !!! Exception during processing !!! DLL load failed while importing cuda_utils: The specified module could not be found.
2024-12-04T16:22:54.319706 - Traceback (most recent call last):
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\NewComfy\venv\Lib\site-packages\triton\backends\nvidia\driver.py", line 74, in compile_module_from_src
    mod = importlib.util.module_from_spec(spec)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<frozen importlib._bootstrap>", line 813, in module_from_spec
  File "<frozen importlib._bootstrap_external>", line 1288, in create_module
  File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
ImportError: DLL load failed while importing cuda_utils: The specified module could not be found.

2024-12-04T16:22:54.382658 - Prompt executed in 245.00 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

{"last_node_id":35,"last_link_id":43,"nodes":[{"id":16,"type":"DownloadAndLoadHyVideoTextEncoder","pos":[-310,248],"size":[441,154],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"hyvid_text_encoder","type":"HYVIDTEXTENCODER","links":[35]}],"properties":{"Node name for S&R":"DownloadAndLoadHyVideoTextEncoder"},"widgets_values":["Kijai/llava-llama-3-8b-text-encoder-tokenizer","openai/clip-vit-large-patch14","fp16",false,2]},{"id":35,"type":"HyVideoBlockSwap","pos":[-351,-44],"size":[315,82],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"block_swap_args","type":"BLOCKSWAPARGS","links":[43]}],"properties":{"Node name for S&R":"HyVideoBlockSwap"},"widgets_values":[20,0]},{"id":34,"type":"VHS_VideoCombine","pos":[1367,-275],"size":[371.7926940917969,310],"flags":{},"order":7,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":42},{"name":"audio","type":"AUDIO","link":null,"shape":7},{"name":"meta_batch","type":"VHS_BatchManager","link":null,"shape":7},{"name":"vae","type":"VAE","link":null,"shape":7}],"outputs":[{"name":"Filenames","type":"VHS_FILENAMES","links":null}],"properties":{"Node name for S&R":"VHS_VideoCombine"},"widgets_values":{"frame_rate":16,"loop_count":0,"filename_prefix":"HunyuanVideo","format":"video/h264-mp4","pix_fmt":"yuv420p","crf":19,"save_metadata":true,"pingpong":false,"save_output":false,"videopreview":{"hidden":false,"paused":false,"params":{"filename":"HunyuanVideo_00059.mp4","subfolder":"","type":"temp","format":"video/h264-mp4","frame_rate":16},"muted":false}}},{"id":5,"type":"HyVideoDecode","pos":[920,-279],"size":[345.4285888671875,102],"flags":{},"order":6,"mode":0,"inputs":[{"name":"vae","type":"VAE","link":6},{"name":"samples","type":"LATENT","link":4}],"outputs":[{"name":"images","type":"IMAGE","links":[42],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoDecode"},"widgets_values":[true,8]},{"id":3,"type":"HyVideoSampler","pos":[668,-62],"size":[315,314],"flags":{},"order":5,"mode":0,"inputs":[{"name":"model","type":"HYVIDEOMODEL","link":2},{"name":"hyvid_embeds","type":"HYVIDEMBEDS","link":36},{"name":"samples","type":"LATENT","link":null,"shape":7}],"outputs":[{"name":"samples","type":"LATENT","links":[4],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoSampler"},"widgets_values":[512,512,33,20,6,9,2,"fixed",1,1]},{"id":30,"type":"HyVideoTextEncode","pos":[203,247],"size":[400,200],"flags":{},"order":3,"mode":0,"inputs":[{"name":"text_encoders","type":"HYVIDTEXTENCODER","link":35}],"outputs":[{"name":"hyvid_embeds","type":"HYVIDEMBEDS","links":[36]}],"properties":{"Node name for S&R":"HyVideoTextEncode"},"widgets_values":["high quality anime style movie featuring a wolf in a forest","bad quality video"]},{"id":7,"type":"HyVideoVAELoader","pos":[442,-282],"size":[379.166748046875,82],"flags":{},"order":2,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7}],"outputs":[{"name":"vae","type":"VAE","links":[6],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoVAELoader"},"widgets_values":["hunyuan_video_vae_bf16.safetensors","fp16"]},{"id":1,"type":"HyVideoModelLoader","pos":[24,-63],"size":[509.7506103515625,178],"flags":{},"order":4,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7},{"name":"block_swap_args","type":"BLOCKSWAPARGS","link":43,"shape":7}],"outputs":[{"name":"model","type":"HYVIDEOMODEL","links":[2],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoModelLoader"},"widgets_values":["hunyuan_video_720_fp8_e4m3fn.safetensors","bf16","fp8_e4m3fn","offload_device","sageattn_varlen"]}],"links":[[2,1,0,3,0,"HYVIDEOMODEL"],[4,3,0,5,1,"LATENT"],[6,7,0,5,0,"VAE"],[35,16,0,30,0,"HYVIDTEXTENCODER"],[36,30,0,3,1,"HYVIDEMBEDS"],[42,5,0,34,0,"IMAGE"],[43,35,0,1,1,"BLOCKSWAPARGS"]],"groups":[],"config":{},"extra":{"ds":{"scale":1.0152559799477276,"offset":[995.1294871495446,363.57691812862925]}},"version":0.4}
@Nightnightlight
Copy link
Author

I even tried it on a separate clean Comfyui build, same error.

ComfyUI Error Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: ImportError
  • Exception Message: DLL load failed while importing cuda_utils: The specified module could not be found.

Stack Trace

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 74, in compile_module_from_src
    mod = importlib.util.module_from_spec(spec)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "<frozen importlib._bootstrap>", line 813, in module_from_spec

  File "<frozen importlib._bootstrap_external>", line 1289, in create_module

  File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed

System Information

  • ComfyUI Version: v0.3.6-13-g3bed56b
  • Arguments: ComfyUI\main.py --windows-standalone-build
  • OS: nt
  • Python Version: 3.12.7 (tags/v3.12.7:0b05ead, Oct 1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]
  • Embedded Python: true
  • PyTorch Version: 2.5.1+cu124

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 25769148416
    • VRAM Free: 17778420304
    • Torch VRAM Total: 6643777536
    • Torch VRAM Free: 66923088

Logs

2024-12-04T17:13:23.787525 - [START] Security scan2024-12-04T17:13:23.788527 - 
2024-12-04T17:13:24.652763 - [DONE] Security scan2024-12-04T17:13:24.652763 - 
2024-12-04T17:13:24.845434 - ## ComfyUI-Manager: installing dependencies done.2024-12-04T17:13:24.845434 - 
2024-12-04T17:13:24.845434 - ** ComfyUI startup time:2024-12-04T17:13:24.845434 -  2024-12-04T17:13:24.845434 - 2024-12-04 17:13:24.8454342024-12-04T17:13:24.845434 - 
2024-12-04T17:13:24.877632 - ** Platform:2024-12-04T17:13:24.877632 -  2024-12-04T17:13:24.877632 - Windows2024-12-04T17:13:24.877632 - 
2024-12-04T17:13:24.877632 - ** Python version:2024-12-04T17:13:24.877632 -  2024-12-04T17:13:24.877632 - 3.12.7 (tags/v3.12.7:0b05ead, Oct  1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]2024-12-04T17:13:24.877632 - 
2024-12-04T17:13:24.878693 - ** Python executable:2024-12-04T17:13:24.878693 -  2024-12-04T17:13:24.878693 - E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\python.exe2024-12-04T17:13:24.878693 - 
2024-12-04T17:13:24.878693 - ** ComfyUI Path:2024-12-04T17:13:24.878693 -  2024-12-04T17:13:24.878693 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI2024-12-04T17:13:24.878693 - 
2024-12-04T17:13:24.878693 - ** Log path:2024-12-04T17:13:24.878693 -  2024-12-04T17:13:24.878693 - E:\SD\ComfyUI\ComfyUI_windows_portable\comfyui.log2024-12-04T17:13:24.878693 - 
2024-12-04T17:13:25.822287 - 
Prestartup times for custom nodes:2024-12-04T17:13:25.822287 - 
2024-12-04T17:13:25.822287 -    2.0 seconds:2024-12-04T17:13:25.823287 -  2024-12-04T17:13:25.823287 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager2024-12-04T17:13:25.823287 - 
2024-12-04T17:13:25.823287 - 
2024-12-04T17:13:30.703345 - Total VRAM 24575 MB, total RAM 16335 MB
2024-12-04T17:13:30.703345 - pytorch version: 2.5.1+cu124
2024-12-04T17:13:30.703345 - Set vram state to: NORMAL_VRAM
2024-12-04T17:13:30.703345 - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-12-04T17:13:33.262358 - Using pytorch cross attention
2024-12-04T17:13:36.069654 - [Prompt Server] web root: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\web
2024-12-04T17:13:37.770188 - ### Loading: ComfyUI-Manager (V2.55)2024-12-04T17:13:37.770188 - 
2024-12-04T17:13:37.991696 - ### ComfyUI Version: v0.3.6-13-g3bed56b | Released on '2024-12-04'2024-12-04T17:13:37.991696 - 
2024-12-04T17:13:38.459641 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json2024-12-04T17:13:38.459641 - 
2024-12-04T17:13:38.470487 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json2024-12-04T17:13:38.470487 - 
2024-12-04T17:13:38.521147 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json2024-12-04T17:13:38.521147 - 
2024-12-04T17:13:38.560289 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json2024-12-04T17:13:38.560289 - 
2024-12-04T17:13:38.574289 - 
Import times for custom nodes:
2024-12-04T17:13:38.574289 -    0.0 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
2024-12-04T17:13:38.575289 -    0.2 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
2024-12-04T17:13:38.576292 -    0.6 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
2024-12-04T17:13:38.576292 -    0.6 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper
2024-12-04T17:13:38.576292 - 
2024-12-04T17:13:38.589290 - Starting server

2024-12-04T17:13:38.592292 - To see the GUI go to: http://127.0.0.1:8188
2024-12-04T17:13:38.600290 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2024-12-04T17:13:38.601290 - 
2024-12-04T17:13:39.671100 - FETCH DATA from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2024-12-04T17:13:39.671100 - 2024-12-04T17:13:39.678117 -  [DONE]2024-12-04T17:13:39.678117 - 
2024-12-04T17:14:01.364153 - got prompt
2024-12-04T17:14:01.450536 - The config attributes {'mid_block_causal_attn': True} were passed to AutoencoderKLCausal3D, but are not expected and will be ignored. Please verify your config.json configuration file.
2024-12-04T17:14:04.799653 - Loading text encoder model (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T17:14:21.225188 - Text encoder to dtype: torch.float16
2024-12-04T17:14:21.398156 - Loading tokenizer (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T17:14:21.527955 - Loading text encoder model (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T17:19:15.609834 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [04:53<00:00, 49.73s/it]2024-12-04T17:19:15.624845 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [04:53<00:00, 73.42s/it]2024-12-04T17:19:15.628426 - 
2024-12-04T17:21:23.178586 - Text encoder to dtype: torch.float16
2024-12-04T17:21:40.007754 - Loading tokenizer (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T17:22:00.322785 - Using accelerate to load and assign model weights to device...
2024-12-04T17:22:00.807891 - Scheduler config:2024-12-04T17:22:00.808892 -  2024-12-04T17:22:00.810891 - FrozenDict({'num_train_timesteps': 1000, 'shift': 9.0, 'reverse': True, 'solver': 'euler', 'n_tokens': None, '_use_default_values': ['num_train_timesteps', 'n_tokens']})2024-12-04T17:22:00.810891 - 
2024-12-04T17:22:01.018094 - Input (height, width, video_length) = (512, 512, 33)
2024-12-04T17:22:03.925027 - Swapping 20 double blocks and 0 single blocks2024-12-04T17:22:03.936566 - 
2024-12-04T17:23:10.765519 - Sampling 33 frames in 9 latents at 512x512 with 20 inference steps
2024-12-04T17:23:10.802403 - 
  0%|                                                                                           | 0/20 [00:00<?, ?it/s]2024-12-04T17:23:14.438655 - 
  0%|                                                                                           | 0/20 [00:03<?, ?it/s]2024-12-04T17:23:14.438655 - 
2024-12-04T17:23:14.732341 - !!! Exception during processing !!! DLL load failed while importing cuda_utils: The specified module could not be found.
2024-12-04T17:23:15.052903 - Traceback (most recent call last):
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 74, in compile_module_from_src
    mod = importlib.util.module_from_spec(spec)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "<frozen importlib._bootstrap>", line 813, in module_from_spec
  File "<frozen importlib._bootstrap_external>", line 1289, in create_module
  File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
ImportError: DLL load failed while importing cuda_utils: The specified module could not be found.

2024-12-04T17:23:15.079447 - Prompt executed in 553.68 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

{"last_node_id":35,"last_link_id":43,"nodes":[{"id":16,"type":"DownloadAndLoadHyVideoTextEncoder","pos":[-310,248],"size":[441,154],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"hyvid_text_encoder","type":"HYVIDTEXTENCODER","links":[35]}],"properties":{"Node name for S&R":"DownloadAndLoadHyVideoTextEncoder"},"widgets_values":["Kijai/llava-llama-3-8b-text-encoder-tokenizer","openai/clip-vit-large-patch14","fp16",false,2]},{"id":35,"type":"HyVideoBlockSwap","pos":[-351,-44],"size":[315,82],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"block_swap_args","type":"BLOCKSWAPARGS","links":[43]}],"properties":{"Node name for S&R":"HyVideoBlockSwap"},"widgets_values":[20,0]},{"id":34,"type":"VHS_VideoCombine","pos":[1367,-275],"size":[371.7926940917969,310],"flags":{},"order":7,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":42},{"name":"audio","type":"AUDIO","link":null,"shape":7},{"name":"meta_batch","type":"VHS_BatchManager","link":null,"shape":7},{"name":"vae","type":"VAE","link":null,"shape":7}],"outputs":[{"name":"Filenames","type":"VHS_FILENAMES","links":null}],"properties":{"Node name for S&R":"VHS_VideoCombine"},"widgets_values":{"frame_rate":16,"loop_count":0,"filename_prefix":"HunyuanVideo","format":"video/h264-mp4","pix_fmt":"yuv420p","crf":19,"save_metadata":true,"pingpong":false,"save_output":false,"videopreview":{"hidden":false,"paused":false,"params":{"filename":"HunyuanVideo_00059.mp4","subfolder":"","type":"temp","format":"video/h264-mp4","frame_rate":16},"muted":false}}},{"id":5,"type":"HyVideoDecode","pos":[920,-279],"size":[345.4285888671875,102],"flags":{},"order":6,"mode":0,"inputs":[{"name":"vae","type":"VAE","link":6},{"name":"samples","type":"LATENT","link":4}],"outputs":[{"name":"images","type":"IMAGE","links":[42],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoDecode"},"widgets_values":[true,8]},{"id":3,"type":"HyVideoSampler","pos":[668,-62],"size":[315,314],"flags":{},"order":5,"mode":0,"inputs":[{"name":"model","type":"HYVIDEOMODEL","link":2},{"name":"hyvid_embeds","type":"HYVIDEMBEDS","link":36},{"name":"samples","type":"LATENT","link":null,"shape":7}],"outputs":[{"name":"samples","type":"LATENT","links":[4],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoSampler"},"widgets_values":[512,512,33,20,6,9,2,"fixed",1,1]},{"id":30,"type":"HyVideoTextEncode","pos":[203,247],"size":[400,200],"flags":{},"order":3,"mode":0,"inputs":[{"name":"text_encoders","type":"HYVIDTEXTENCODER","link":35}],"outputs":[{"name":"hyvid_embeds","type":"HYVIDEMBEDS","links":[36]}],"properties":{"Node name for S&R":"HyVideoTextEncode"},"widgets_values":["high quality anime style movie featuring a wolf in a forest","bad quality video"]},{"id":7,"type":"HyVideoVAELoader","pos":[442,-282],"size":[379.166748046875,82],"flags":{},"order":2,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7}],"outputs":[{"name":"vae","type":"VAE","links":[6],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoVAELoader"},"widgets_values":["hunyuan_video_vae_bf16.safetensors","fp16"]},{"id":1,"type":"HyVideoModelLoader","pos":[24,-63],"size":[509.7506103515625,178],"flags":{},"order":4,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7},{"name":"block_swap_args","type":"BLOCKSWAPARGS","link":43,"shape":7}],"outputs":[{"name":"model","type":"HYVIDEOMODEL","links":[2],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoModelLoader"},"widgets_values":["hunyuan_video_720_fp8_e4m3fn.safetensors","bf16","fp8_e4m3fn","offload_device","sageattn_varlen"]}],"links":[[2,1,0,3,0,"HYVIDEOMODEL"],[4,3,0,5,1,"LATENT"],[6,7,0,5,0,"VAE"],[35,16,0,30,0,"HYVIDTEXTENCODER"],[36,30,0,3,1,"HYVIDEMBEDS"],[42,5,0,34,0,"IMAGE"],[43,35,0,1,1,"BLOCKSWAPARGS"]],"groups":[],"config":{},"extra":{"ds":{"scale":0.6934334949441483,"offset":[329.9757523237481,148.47837069800934]}},"version":0.4}

Additional Context

(Please add any additional context or steps to reproduce the error here)

@kijai
Copy link
Owner

kijai commented Dec 4, 2024

Triton/sageattention install is lot more involved than just installing the pip packages/wheels. You are missing some part of the required tools for compilation (visual studio + cuda-toolkit) or something in their configs are not correct.

You can refer to this guide https://purz.notion.site/Get-Windows-Triton-working-for-Mochi-6a0c055e21c84cfba7f1dd628e624e97

Alternatively you if you can install flash_attn you don't need triton/sageattention (but it will be slower).

@Nightnightlight
Copy link
Author

Nightnightlight commented Dec 4, 2024

I followed that guide to install triton / sageattention in both of these setups.

@kijai
Copy link
Owner

kijai commented Dec 4, 2024

I followed that guide to install triton / sageattention in both of these setups.

Something is wrong with that still though, I've seen that error fixed by clearing the . triton cache, which can be found at

C:\Users\username\.triton

@Nightnightlight
Copy link
Author

Clearing the triton cache resulted in a new error.

ComfyUI Error Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: subprocess.CalledProcessError
  • Exception Message: Command '['C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.40.33807\bin\Hostx64\x64\cl.EXE', 'C:\Users\B\AppData\Local\Temp\tmp4_2x_sqs\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\include', '/IC:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\include', '/IC:\Users\B\AppData\Local\Temp\tmp4_2x_sqs', '/IE:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Include', '/IC:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.40.33807\include', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\shared', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\ucrt', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\um', '/link', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\lib', '/LIBPATH:C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\lib\x64', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:C:\Python312\libs', '/LIBPATH:C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.40.33807\lib\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.22621.0\ucrt\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.22621.0\um\x64', 'cuda.lib', '/OUT:C:\Users\B\AppData\Local\Temp\tmp4_2x_sqs\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.

Stack Trace

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 69, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\build.py", line 71, in _build
    ret = subprocess.check_call(cc_cmd)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "subprocess.py", line 413, in check_call

System Information

  • ComfyUI Version: v0.3.6-13-g3bed56b
  • Arguments: ComfyUI\main.py --windows-standalone-build
  • OS: nt
  • Python Version: 3.12.7 (tags/v3.12.7:0b05ead, Oct 1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]
  • Embedded Python: true
  • PyTorch Version: 2.5.1+cu124

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 25769148416
    • VRAM Free: 17778420304
    • Torch VRAM Total: 6643777536
    • Torch VRAM Free: 66923088

Logs

2024-12-04T18:02:42.109835 - [START] Security scan2024-12-04T18:02:42.109835 - 
2024-12-04T18:02:42.999915 - [DONE] Security scan2024-12-04T18:02:42.999915 - 
2024-12-04T18:02:43.158927 - ## ComfyUI-Manager: installing dependencies done.2024-12-04T18:02:43.158927 - 
2024-12-04T18:02:43.159929 - ** ComfyUI startup time:2024-12-04T18:02:43.159929 -  2024-12-04T18:02:43.159929 - 2024-12-04 18:02:43.1599292024-12-04T18:02:43.159929 - 
2024-12-04T18:02:43.188995 - ** Platform:2024-12-04T18:02:43.188995 -  2024-12-04T18:02:43.188995 - Windows2024-12-04T18:02:43.188995 - 
2024-12-04T18:02:43.188995 - ** Python version:2024-12-04T18:02:43.188995 -  2024-12-04T18:02:43.188995 - 3.12.7 (tags/v3.12.7:0b05ead, Oct  1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]2024-12-04T18:02:43.188995 - 
2024-12-04T18:02:43.188995 - ** Python executable:2024-12-04T18:02:43.188995 -  2024-12-04T18:02:43.188995 - E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\python.exe2024-12-04T18:02:43.190005 - 
2024-12-04T18:02:43.190005 - ** ComfyUI Path:2024-12-04T18:02:43.190005 -  2024-12-04T18:02:43.190005 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI2024-12-04T18:02:43.190005 - 
2024-12-04T18:02:43.190005 - ** Log path:2024-12-04T18:02:43.190005 -  2024-12-04T18:02:43.190005 - E:\SD\ComfyUI\ComfyUI_windows_portable\comfyui.log2024-12-04T18:02:43.190005 - 
2024-12-04T18:02:44.127423 - 
Prestartup times for custom nodes:2024-12-04T18:02:44.127423 - 
2024-12-04T18:02:44.127423 -    2.0 seconds:2024-12-04T18:02:44.127423 -  2024-12-04T18:02:44.127423 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager2024-12-04T18:02:44.127423 - 
2024-12-04T18:02:44.127423 - 
2024-12-04T18:02:49.218128 - Total VRAM 24575 MB, total RAM 16335 MB
2024-12-04T18:02:49.219129 - pytorch version: 2.5.1+cu124
2024-12-04T18:02:49.219129 - Set vram state to: NORMAL_VRAM
2024-12-04T18:02:49.219129 - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-12-04T18:02:51.645546 - Using pytorch cross attention
2024-12-04T18:02:53.912644 - [Prompt Server] web root: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\web
2024-12-04T18:02:55.130718 - ### Loading: ComfyUI-Manager (V2.55)2024-12-04T18:02:55.130718 - 
2024-12-04T18:02:55.333721 - ### ComfyUI Version: v0.3.6-13-g3bed56b | Released on '2024-12-04'2024-12-04T18:02:55.333721 - 
2024-12-04T18:02:55.757015 - 
Import times for custom nodes:
2024-12-04T18:02:55.757015 -    0.0 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
2024-12-04T18:02:55.758017 -    0.2 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
2024-12-04T18:02:55.758017 -    0.4 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper
2024-12-04T18:02:55.758017 -    0.5 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
2024-12-04T18:02:55.758017 - 
2024-12-04T18:02:55.764015 - Starting server

2024-12-04T18:02:55.764015 - To see the GUI go to: http://127.0.0.1:8188
2024-12-04T18:02:55.795314 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json2024-12-04T18:02:55.795314 - 
2024-12-04T18:02:55.808233 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json2024-12-04T18:02:55.808233 - 
2024-12-04T18:02:55.810028 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json2024-12-04T18:02:55.810895 - 
2024-12-04T18:02:55.906029 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2024-12-04T18:02:55.906029 - 
2024-12-04T18:02:55.933026 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json2024-12-04T18:02:55.933026 - 
2024-12-04T18:02:56.610575 - FETCH DATA from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2024-12-04T18:02:56.610575 - 2024-12-04T18:02:56.617919 -  [DONE]2024-12-04T18:02:56.618921 - 
2024-12-04T18:03:01.366954 - got prompt
2024-12-04T18:03:01.449787 - The config attributes {'mid_block_causal_attn': True} were passed to AutoencoderKLCausal3D, but are not expected and will be ignored. Please verify your config.json configuration file.
2024-12-04T18:03:04.619041 - Loading text encoder model (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T18:03:06.614980 - Text encoder to dtype: torch.float16
2024-12-04T18:03:06.837006 - Loading tokenizer (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T18:03:06.951978 - Loading text encoder model (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T18:04:25.893512 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [01:18<00:00, 16.78s/it]2024-12-04T18:04:25.896511 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [01:18<00:00, 19.66s/it]2024-12-04T18:04:25.896511 - 
2024-12-04T18:06:06.276983 - Text encoder to dtype: torch.float16
2024-12-04T18:06:25.757912 - Loading tokenizer (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T18:06:46.377065 - Using accelerate to load and assign model weights to device...
2024-12-04T18:06:47.202499 - Scheduler config:2024-12-04T18:06:47.238040 -  2024-12-04T18:06:47.243040 - FrozenDict({'num_train_timesteps': 1000, 'shift': 9.0, 'reverse': True, 'solver': 'euler', 'n_tokens': None, '_use_default_values': ['n_tokens', 'num_train_timesteps']})2024-12-04T18:06:47.244039 - 
2024-12-04T18:06:47.524117 - Input (height, width, video_length) = (512, 512, 33)
2024-12-04T18:06:50.349965 - Swapping 20 double blocks and 0 single blocks2024-12-04T18:06:50.352960 - 
2024-12-04T18:07:08.649686 - Sampling 33 frames in 9 latents at 512x512 with 20 inference steps
2024-12-04T18:07:08.721518 - 
  0%|                                                                                           | 0/20 [00:00<?, ?it/s]2024-12-04T18:07:14.075529 - 
  0%|                                                                                           | 0/20 [00:05<?, ?it/s]2024-12-04T18:07:14.075529 - 
2024-12-04T18:07:14.439055 - !!! Exception during processing !!! Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\include', '/IC:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Include', '/IC:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\lib\\x64', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:C:\\Python312\\libs', '/LIBPATH:C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs\\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.
2024-12-04T18:07:14.524558 - Traceback (most recent call last):
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 69, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\build.py", line 71, in _build
    ret = subprocess.check_call(cc_cmd)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "subprocess.py", line 413, in check_call
subprocess.CalledProcessError: Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\include', '/IC:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Include', '/IC:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\lib\\x64', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:C:\\Python312\\libs', '/LIBPATH:C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.40.33807\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\B\\AppData\\Local\\Temp\\tmp4_2x_sqs\\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.

2024-12-04T18:07:14.555634 - Prompt executed in 253.16 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

{"last_node_id":35,"last_link_id":43,"nodes":[{"id":16,"type":"DownloadAndLoadHyVideoTextEncoder","pos":[-310,248],"size":[441,154],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"hyvid_text_encoder","type":"HYVIDTEXTENCODER","links":[35]}],"properties":{"Node name for S&R":"DownloadAndLoadHyVideoTextEncoder"},"widgets_values":["Kijai/llava-llama-3-8b-text-encoder-tokenizer","openai/clip-vit-large-patch14","fp16",false,2]},{"id":35,"type":"HyVideoBlockSwap","pos":[-351,-44],"size":[315,82],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"block_swap_args","type":"BLOCKSWAPARGS","links":[43]}],"properties":{"Node name for S&R":"HyVideoBlockSwap"},"widgets_values":[20,0]},{"id":34,"type":"VHS_VideoCombine","pos":[1367,-275],"size":[371.7926940917969,310],"flags":{},"order":7,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":42},{"name":"audio","type":"AUDIO","link":null,"shape":7},{"name":"meta_batch","type":"VHS_BatchManager","link":null,"shape":7},{"name":"vae","type":"VAE","link":null,"shape":7}],"outputs":[{"name":"Filenames","type":"VHS_FILENAMES","links":null}],"properties":{"Node name for S&R":"VHS_VideoCombine"},"widgets_values":{"frame_rate":16,"loop_count":0,"filename_prefix":"HunyuanVideo","format":"video/h264-mp4","pix_fmt":"yuv420p","crf":19,"save_metadata":true,"pingpong":false,"save_output":false,"videopreview":{"hidden":false,"paused":false,"params":{"filename":"HunyuanVideo_00059.mp4","subfolder":"","type":"temp","format":"video/h264-mp4","frame_rate":16},"muted":false}}},{"id":5,"type":"HyVideoDecode","pos":[920,-279],"size":[345.4285888671875,102],"flags":{},"order":6,"mode":0,"inputs":[{"name":"vae","type":"VAE","link":6},{"name":"samples","type":"LATENT","link":4}],"outputs":[{"name":"images","type":"IMAGE","links":[42],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoDecode"},"widgets_values":[true,8]},{"id":3,"type":"HyVideoSampler","pos":[668,-62],"size":[315,314],"flags":{},"order":5,"mode":0,"inputs":[{"name":"model","type":"HYVIDEOMODEL","link":2},{"name":"hyvid_embeds","type":"HYVIDEMBEDS","link":36},{"name":"samples","type":"LATENT","link":null,"shape":7}],"outputs":[{"name":"samples","type":"LATENT","links":[4],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoSampler"},"widgets_values":[512,512,33,20,6,9,2,"fixed",1,1]},{"id":30,"type":"HyVideoTextEncode","pos":[203,247],"size":[400,200],"flags":{},"order":3,"mode":0,"inputs":[{"name":"text_encoders","type":"HYVIDTEXTENCODER","link":35}],"outputs":[{"name":"hyvid_embeds","type":"HYVIDEMBEDS","links":[36]}],"properties":{"Node name for S&R":"HyVideoTextEncode"},"widgets_values":["high quality anime style movie featuring a wolf in a forest","bad quality video"]},{"id":7,"type":"HyVideoVAELoader","pos":[442,-282],"size":[379.166748046875,82],"flags":{},"order":2,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7}],"outputs":[{"name":"vae","type":"VAE","links":[6],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoVAELoader"},"widgets_values":["hunyuan_video_vae_bf16.safetensors","fp16"]},{"id":1,"type":"HyVideoModelLoader","pos":[24,-63],"size":[509.7506103515625,178],"flags":{},"order":4,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7},{"name":"block_swap_args","type":"BLOCKSWAPARGS","link":43,"shape":7}],"outputs":[{"name":"model","type":"HYVIDEOMODEL","links":[2],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoModelLoader"},"widgets_values":["hunyuan_video_720_fp8_e4m3fn.safetensors","bf16","fp8_e4m3fn","offload_device","sageattn_varlen"]}],"links":[[2,1,0,3,0,"HYVIDEOMODEL"],[4,3,0,5,1,"LATENT"],[6,7,0,5,0,"VAE"],[35,16,0,30,0,"HYVIDTEXTENCODER"],[36,30,0,3,1,"HYVIDEMBEDS"],[42,5,0,34,0,"IMAGE"],[43,35,0,1,1,"BLOCKSWAPARGS"]],"groups":[],"config":{},"extra":{"ds":{"scale":1.0152559799477276,"offset":[313.03549934988183,432.5250468592811]}},"version":0.4}

Additional Context

(Please add any additional context or steps to reproduce the error here)

@Nightnightlight
Copy link
Author

I even tested Mochi with sageattention. It works but this does not.

@Bocchi-Chan2023
Copy link

"new error" also occurred for me

@Bocchi-Chan2023
Copy link

https://github.com/woct0rdho/triton-windows/releases/tag/v3.0.0-windows.post1
extract zip depends on your python version.

@Nightnightlight
Copy link
Author

Nightnightlight commented Dec 5, 2024

Nope. I even uninstalled visual build tools, pip uninstalled all the requirements, then reinstalled visual build tools using what this shows to use reinstalled pytorch / cuda / cuda toolkit / all the requirements.txt / triton / sageattention, making sure all were for the same version of python / cuda that im using and that the PATHs were pointing to the right places. Still not working. Sage attention works for mochi though. But I get this error below when trying to use HunyuanVideoWrapper:

ComfyUI Error Report

Error Details

  • Node ID: 3
  • Node Type: HyVideoSampler
  • Exception Type: subprocess.CalledProcessError
  • Exception Message: Command '['C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.42.34433\bin\Hostx64\x64\cl.EXE', 'C:\Users\12105\AppData\Local\Temp\tmppaavr7ls\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\include', '/IC:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\include', '/IC:\Users\12105\AppData\Local\Temp\tmppaavr7ls', '/IE:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Include', '/IC:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.42.34433\include', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\shared', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\ucrt', '/IC:\Program Files (x86)\Windows Kits\10\Include\10.0.22621.0\um', '/link', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\lib', '/LIBPATH:C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v12.5\lib\x64', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\libs', '/LIBPATH:C:\Python312\libs', '/LIBPATH:C:\Program Files (x86)\Microsoft Visual Studio\2022\BuildTools\VC\Tools\MSVC\14.42.34433\lib\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.22621.0\ucrt\x64', '/LIBPATH:C:\Program Files (x86)\Windows Kits\10\Lib\10.0.22621.0\um\x64', 'cuda.lib', '/OUT:C:\Users\12105\AppData\Local\Temp\tmppaavr7ls\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.

Stack Trace

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 69, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\build.py", line 71, in _build
    ret = subprocess.check_call(cc_cmd)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^

  File "subprocess.py", line 413, in check_call

System Information

  • ComfyUI Version: v0.3.6-13-g3bed56b
  • Arguments: ComfyUI\main.py --windows-standalone-build
  • OS: nt
  • Python Version: 3.12.7 (tags/v3.12.7:0b05ead, Oct 1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]
  • Embedded Python: true
  • PyTorch Version: 2.5.1+cu124

Devices

  • Name: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
    • Type: cuda
    • VRAM Total: 25769148416
    • VRAM Free: 17778420304
    • Torch VRAM Total: 6643777536
    • Torch VRAM Free: 66923088

Logs

2024-12-04T23:13:47.312933 - [START] Security scan2024-12-04T23:13:47.312933 - 
2024-12-04T23:13:48.154586 - [DONE] Security scan2024-12-04T23:13:48.154586 - 
2024-12-04T23:13:48.277758 - ## ComfyUI-Manager: installing dependencies done.2024-12-04T23:13:48.277758 - 
2024-12-04T23:13:48.277758 - ** ComfyUI startup time:2024-12-04T23:13:48.277758 -  2024-12-04T23:13:48.277758 - 2024-12-04 23:13:48.2777582024-12-04T23:13:48.278758 - 
2024-12-04T23:13:48.308532 - ** Platform:2024-12-04T23:13:48.308532 -  2024-12-04T23:13:48.308532 - Windows2024-12-04T23:13:48.308532 - 
2024-12-04T23:13:48.308532 - ** Python version:2024-12-04T23:13:48.308532 -  2024-12-04T23:13:48.308532 - 3.12.7 (tags/v3.12.7:0b05ead, Oct  1 2024, 03:06:41) [MSC v.1941 64 bit (AMD64)]2024-12-04T23:13:48.308532 - 
2024-12-04T23:13:48.308532 - ** Python executable:2024-12-04T23:13:48.308532 -  2024-12-04T23:13:48.309530 - E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\python.exe2024-12-04T23:13:48.309530 - 
2024-12-04T23:13:48.309530 - ** ComfyUI Path:2024-12-04T23:13:48.309530 -  2024-12-04T23:13:48.309530 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI2024-12-04T23:13:48.309530 - 
2024-12-04T23:13:48.309530 - ** Log path:2024-12-04T23:13:48.309530 -  2024-12-04T23:13:48.309530 - E:\SD\ComfyUI\ComfyUI_windows_portable\comfyui.log2024-12-04T23:13:48.309530 - 
2024-12-04T23:13:49.227594 - 
Prestartup times for custom nodes:2024-12-04T23:13:49.227594 - 
2024-12-04T23:13:49.227594 -    1.9 seconds:2024-12-04T23:13:49.228594 -  2024-12-04T23:13:49.228594 - E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager2024-12-04T23:13:49.228594 - 
2024-12-04T23:13:49.228594 - 
2024-12-04T23:13:51.154608 - Total VRAM 24575 MB, total RAM 16335 MB
2024-12-04T23:13:51.154608 - pytorch version: 2.5.1+cu124
2024-12-04T23:13:51.155609 - Set vram state to: NORMAL_VRAM
2024-12-04T23:13:51.155609 - Device: cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync
2024-12-04T23:13:52.393611 - Using pytorch cross attention
2024-12-04T23:13:53.721555 - [Prompt Server] web root: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\web
2024-12-04T23:13:54.531666 - ### Loading: ComfyUI-Manager (V2.55)2024-12-04T23:13:54.531666 - 
2024-12-04T23:13:54.711892 - ### ComfyUI Version: v0.3.6-13-g3bed56b | Released on '2024-12-04'2024-12-04T23:13:54.711892 - 
2024-12-04T23:13:55.077711 - 
Import times for custom nodes:
2024-12-04T23:13:55.077711 -    0.0 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\websocket_image_save.py
2024-12-04T23:13:55.077711 -    0.1 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-VideoHelperSuite
2024-12-04T23:13:55.077711 -    0.4 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper
2024-12-04T23:13:55.078711 -    0.5 seconds: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager
2024-12-04T23:13:55.078711 - 
2024-12-04T23:13:55.083711 - Starting server

2024-12-04T23:13:55.083711 - To see the GUI go to: http://127.0.0.1:8188
2024-12-04T23:13:55.259631 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json2024-12-04T23:13:55.260635 - 
2024-12-04T23:13:55.274633 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json2024-12-04T23:13:55.274633 - 
2024-12-04T23:13:55.304819 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json2024-12-04T23:13:55.305819 - 
2024-12-04T23:13:55.367818 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json2024-12-04T23:13:55.367818 - 
2024-12-04T23:13:55.399818 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2024-12-04T23:13:55.399818 - 
2024-12-04T23:13:55.817631 - FETCH DATA from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-Manager\extension-node-map.json2024-12-04T23:13:55.818633 - 2024-12-04T23:13:55.822631 -  [DONE]2024-12-04T23:13:55.822631 - 
2024-12-04T23:14:07.368752 - got prompt
2024-12-04T23:14:07.459725 - The config attributes {'mid_block_causal_attn': True} were passed to AutoencoderKLCausal3D, but are not expected and will be ignored. Please verify your config.json configuration file.
2024-12-04T23:14:10.509739 - Loading text encoder model (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T23:14:12.474744 - Text encoder to dtype: torch.float16
2024-12-04T23:14:12.640750 - Loading tokenizer (clipL) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\clip\clip-vit-large-patch14
2024-12-04T23:14:12.743749 - Loading text encoder model (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T23:17:02.849151 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [02:49<00:00, 29.86s/it]2024-12-04T23:17:02.867951 - 
Loading checkpoint shards: 100%|█████████████████████████████████████████████████████████| 4/4 [02:49<00:00, 42.45s/it]2024-12-04T23:17:02.867951 - 
2024-12-04T23:18:42.556951 - Text encoder to dtype: torch.float16
2024-12-04T23:19:00.558619 - Loading tokenizer (llm) from: E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\models\LLM\llava-llama-3-8b-text-encoder-tokenizer
2024-12-04T23:19:20.462792 - Using accelerate to load and assign model weights to device...
2024-12-04T23:19:21.149761 - Scheduler config:2024-12-04T23:19:21.158171 -  2024-12-04T23:19:21.171730 - FrozenDict({'num_train_timesteps': 1000, 'shift': 9.0, 'reverse': True, 'solver': 'euler', 'n_tokens': None, '_use_default_values': ['n_tokens', 'num_train_timesteps']})2024-12-04T23:19:21.171730 - 
2024-12-04T23:19:21.363540 - Input (height, width, video_length) = (512, 512, 33)
2024-12-04T23:19:24.905569 - Swapping 20 double blocks and 0 single blocks2024-12-04T23:19:24.916881 - 
2024-12-04T23:19:42.437872 - Sampling 33 frames in 9 latents at 512x512 with 20 inference steps
2024-12-04T23:19:42.498370 - 
  0%|                                                                                           | 0/20 [00:00<?, ?it/s]2024-12-04T23:19:48.514723 - 
  0%|                                                                                           | 0/20 [00:06<?, ?it/s]2024-12-04T23:19:48.514723 - 
2024-12-04T23:19:48.960918 - !!! Exception during processing !!! Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\include', '/IC:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Include', '/IC:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\lib\\x64', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:C:\\Python312\\libs', '/LIBPATH:C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls\\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.
2024-12-04T23:19:49.051628 - Traceback (most recent call last):
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 324, in execute
    output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                                           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 199, in get_output_data
    return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb)
                    ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 170, in _map_node_over_list
    process_inputs(input_dict, i)
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\execution.py", line 159, in process_inputs
    results.append(getattr(obj, func)(**inputs))
                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\nodes.py", line 692, in process
    out_latents = model["pipe"](
                  ^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context
    return func(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\diffusion\pipelines\pipeline_hunyuan_video.py", line 537, in __call__
    noise_pred = self.transformer(  # For an input image (129, 192, 336) (1, 256, 256)
                 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 671, in forward
    img, txt = block(*double_block_args)
               ^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1736, in _wrapped_call_impl
    return self._call_impl(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1747, in _call_impl
    return forward_call(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\models.py", line 203, in forward
    attn = attention(
           ^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 126, in attention
    x = sageattn_varlen_func(
        ^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\torch\_dynamo\eval_frame.py", line 632, in _fn
    return fn(*args, **kwargs)
           ^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\ComfyUI\custom_nodes\ComfyUI-HunyuanVideoWrapper\hyvideo\modules\attention.py", line 24, in sageattn_varlen_func
    return sageattn_varlen(q, k, v, cu_seqlens_q, cu_seqlens_kv, max_seqlen_q, max_seqlen_kv)
           ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\core.py", line 198, in sageattn_varlen
    q_int8, q_scale, k_int8, k_scale, cu_seqlens_q_scale, cu_seqlens_k_scale = per_block_int8_varlen(q, k, cu_seqlens_q, cu_seqlens_k, max_seqlen_q, max_seqlen_k, sm_scale=sm_scale)
                                                                               ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\sageattention\quant_per_block_varlen.py", line 69, in per_block_int8
    quant_per_block_int8_kernel[grid](
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 345, in <lambda>
    return lambda *args, **kwargs: self.run(grid=grid, warmup=False, *args, **kwargs)
                                   ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\jit.py", line 607, in run
    device = driver.active.get_current_device()
             ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 23, in __getattr__
    self._initialize_obj()
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 20, in _initialize_obj
    self._obj = self._init_fn()
                ^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\driver.py", line 9, in _create_driver
    return actives[0]()
           ^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 414, in __init__
    self.utils = CudaUtils()  # TODO: make static
                 ^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 92, in __init__
    mod = compile_module_from_src(Path(os.path.join(dirname, "driver.c")).read_text(), "cuda_utils")
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\backends\nvidia\driver.py", line 69, in compile_module_from_src
    so = _build(name, src_path, tmpdir, library_dirs(), include_dir, libraries)
         ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "E:\SD\ComfyUI\ComfyUI_windows_portable\python_embeded\Lib\site-packages\triton\runtime\build.py", line 71, in _build
    ret = subprocess.check_call(cc_cmd)
          ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
  File "subprocess.py", line 413, in check_call
subprocess.CalledProcessError: Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64\\x64\\cl.EXE', 'C:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls\\main.c', '/nologo', '/O2', '/LD', '/wd4819', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\include', '/IC:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\include', '/IC:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls', '/IE:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Include', '/IC:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\include', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\shared', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\ucrt', '/IC:\\Program Files (x86)\\Windows Kits\\10\\Include\\10.0.22621.0\\um', '/link', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\Lib\\site-packages\\triton\\backends\\nvidia\\lib', '/LIBPATH:C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.5\\lib\\x64', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:E:\\SD\\ComfyUI\\ComfyUI_windows_portable\\python_embeded\\libs', '/LIBPATH:C:\\Python312\\libs', '/LIBPATH:C:\\Program Files (x86)\\Microsoft Visual Studio\\2022\\BuildTools\\VC\\Tools\\MSVC\\14.42.34433\\lib\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\ucrt\\x64', '/LIBPATH:C:\\Program Files (x86)\\Windows Kits\\10\\Lib\\10.0.22621.0\\um\\x64', 'cuda.lib', '/OUT:C:\\Users\\12105\\AppData\\Local\\Temp\\tmppaavr7ls\\cuda_utils.cp312-win_amd64.pyd']' returned non-zero exit status 2.

2024-12-04T23:19:49.138918 - Prompt executed in 341.74 seconds

Attached Workflow

Please make sure that workflow does not contain any sensitive information such as API keys or passwords.

{"last_node_id":35,"last_link_id":43,"nodes":[{"id":16,"type":"DownloadAndLoadHyVideoTextEncoder","pos":[-310,248],"size":[441,154],"flags":{},"order":0,"mode":0,"inputs":[],"outputs":[{"name":"hyvid_text_encoder","type":"HYVIDTEXTENCODER","links":[35]}],"properties":{"Node name for S&R":"DownloadAndLoadHyVideoTextEncoder"},"widgets_values":["Kijai/llava-llama-3-8b-text-encoder-tokenizer","openai/clip-vit-large-patch14","fp16",false,2]},{"id":35,"type":"HyVideoBlockSwap","pos":[-351,-44],"size":[315,82],"flags":{},"order":1,"mode":0,"inputs":[],"outputs":[{"name":"block_swap_args","type":"BLOCKSWAPARGS","links":[43]}],"properties":{"Node name for S&R":"HyVideoBlockSwap"},"widgets_values":[20,0]},{"id":34,"type":"VHS_VideoCombine","pos":[1367,-275],"size":[371.7926940917969,310],"flags":{},"order":7,"mode":0,"inputs":[{"name":"images","type":"IMAGE","link":42},{"name":"audio","type":"AUDIO","link":null,"shape":7},{"name":"meta_batch","type":"VHS_BatchManager","link":null,"shape":7},{"name":"vae","type":"VAE","link":null,"shape":7}],"outputs":[{"name":"Filenames","type":"VHS_FILENAMES","links":null}],"properties":{"Node name for S&R":"VHS_VideoCombine"},"widgets_values":{"frame_rate":16,"loop_count":0,"filename_prefix":"HunyuanVideo","format":"video/h264-mp4","pix_fmt":"yuv420p","crf":19,"save_metadata":true,"pingpong":false,"save_output":false,"videopreview":{"hidden":false,"paused":false,"params":{"filename":"HunyuanVideo_00059.mp4","subfolder":"","type":"temp","format":"video/h264-mp4","frame_rate":16},"muted":false}}},{"id":5,"type":"HyVideoDecode","pos":[920,-279],"size":[345.4285888671875,102],"flags":{},"order":6,"mode":0,"inputs":[{"name":"vae","type":"VAE","link":6},{"name":"samples","type":"LATENT","link":4}],"outputs":[{"name":"images","type":"IMAGE","links":[42],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoDecode"},"widgets_values":[true,8]},{"id":3,"type":"HyVideoSampler","pos":[668,-62],"size":[315,314],"flags":{},"order":5,"mode":0,"inputs":[{"name":"model","type":"HYVIDEOMODEL","link":2},{"name":"hyvid_embeds","type":"HYVIDEMBEDS","link":36},{"name":"samples","type":"LATENT","link":null,"shape":7}],"outputs":[{"name":"samples","type":"LATENT","links":[4],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoSampler"},"widgets_values":[512,512,33,20,6,9,2,"fixed",1,1]},{"id":30,"type":"HyVideoTextEncode","pos":[203,247],"size":[400,200],"flags":{},"order":3,"mode":0,"inputs":[{"name":"text_encoders","type":"HYVIDTEXTENCODER","link":35}],"outputs":[{"name":"hyvid_embeds","type":"HYVIDEMBEDS","links":[36]}],"properties":{"Node name for S&R":"HyVideoTextEncode"},"widgets_values":["high quality anime style movie featuring a wolf in a forest","bad quality video"]},{"id":7,"type":"HyVideoVAELoader","pos":[442,-282],"size":[379.166748046875,82],"flags":{},"order":2,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7}],"outputs":[{"name":"vae","type":"VAE","links":[6],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoVAELoader"},"widgets_values":["hunyuan_video_vae_bf16.safetensors","fp16"]},{"id":1,"type":"HyVideoModelLoader","pos":[24,-63],"size":[509.7506103515625,178],"flags":{},"order":4,"mode":0,"inputs":[{"name":"compile_args","type":"COMPILEARGS","link":null,"shape":7},{"name":"block_swap_args","type":"BLOCKSWAPARGS","link":43,"shape":7}],"outputs":[{"name":"model","type":"HYVIDEOMODEL","links":[2],"slot_index":0}],"properties":{"Node name for S&R":"HyVideoModelLoader"},"widgets_values":["hunyuan_video_720_fp8_e4m3fn.safetensors","bf16","fp8_e4m3fn","offload_device","sageattn_varlen"]}],"links":[[2,1,0,3,0,"HYVIDEOMODEL"],[4,3,0,5,1,"LATENT"],[6,7,0,5,0,"VAE"],[35,16,0,30,0,"HYVIDTEXTENCODER"],[36,30,0,3,1,"HYVIDEMBEDS"],[42,5,0,34,0,"IMAGE"],[43,35,0,1,1,"BLOCKSWAPARGS"]],"groups":[],"config":{},"extra":{"ds":{"scale":1.0152559799477276,"offset":[313.03549934988183,432.5250468592811]}},"version":0.4}

Additional Context

(Please add any additional context or steps to reproduce the error here)

@kijai
Copy link
Owner

kijai commented Dec 5, 2024

Someone else had this issue too, not sure what exactly finally fixed it but at one point they had Triton 3.0.0, so make sure it's Triton version 3.1.0 and sageattention version 1.0.6 (or the new 2.0.0 beta)

@al3dv2
Copy link

al3dv2 commented Dec 5, 2024

I fixed this issue by deleting the Triton cache in Users/Triton, but after that, I encountered an error saying that my GPU is too old. I have a Quadro RTX 6000 with 24GB of VRAM.

@esciron
Copy link

esciron commented Dec 5, 2024

Same problem here.
Deleting the Triton cache folder does nothing, reinstalling cuda or even multiple versions of cuda 12.x does not fix it.
Neither reinstalling sage or triton with the correct version and directory.

Edit: Seems like a combination of installing cuda 12.4, copying libs and include folder of python 3.12 to python_embeded (instead of python 3.11.9) and deleting the .triton cache folder again fixed it.

Edit: Nope, python 3.12 files don't work in the end, it throws a bunch of other errors.
Still the same now with 3.11.9 files: DLL load failed while importing cuda_utils

Last edit: Fixed it, this wrapper only works with python 3.11, at least for the 3090. Have to download a really old comfyui version with the 3.11 python_embeded files and drop the content in the embeded folder of my comfy install so comfy uses python 3.11.9 files, then install sage and triton again and that's it.

@u-madara
Copy link

u-madara commented Dec 9, 2024

Same problem here.

@u-madara
Copy link

u-madara commented Dec 9, 2024

image

@realstevewarner
Copy link

I tried for days to get this to work. It never did. Got the same Triton/Sageattention/CUDA DLL errors everyone has reported.

Comfy now has Hunyuan support baked in. I can confirm it works. Here's what I did:

  1. Go to the Comfy page detailing the Hunyuan support here:
    https://blog.comfy.org/p/hunyuanvideo-native-support-in-comfyui

  2. Use the Desktop version. I've only ever had portable installs. For this one, I went ahead with the desktop version as I figured it might resolve any DLL/dependency issues. You may be able to get it to work by pulling the latest comfy for your portable. I didn't want to screw around any longer. Did a fresh install. It works.

  3. Install the latest NVIDIA driver. I have a 4090. I used the Creative Studio driver, not the game driver.

  4. Load the workflow from the Comfy page in Command '['C:\\Program Files (x86)\\Microsoft Visual Studio\\2019\\BuildTools\\VC\\Tools\\MSVC\\14.29.30133\\bin\\Hostx64\\x64\\cl.EXE' #1 above. It referenced a different Hunyuan model and CLIP than I had previously downloaded. Make sure to grab the models needed.

That's pretty much it. It now runs without a hitch.

If you're still trying to get Hunyuan to run and are fighting the Triton/Sageattention/CUDA DLL errors, you can safely give up that merry go round and get it working.

@kijai
Copy link
Owner

kijai commented Dec 21, 2024

I tried for days to get this to work. It never did. Got the same Triton/Sageattention/CUDA DLL errors everyone has reported.

Comfy now has Hunyuan support baked in. I can confirm it works. Here's what I did:

  1. Go to the Comfy page detailing the Hunyuan support here:
    https://blog.comfy.org/p/hunyuanvideo-native-support-in-comfyui
  2. Use the Desktop version. I've only ever had portable installs. For this one, I went ahead with the desktop version as I figured it might resolve any DLL/dependency issues. You may be able to get it to work by pulling the latest comfy for your portable. I didn't want to screw around any longer. Did a fresh install. It works.
  3. Install the latest NVIDIA driver. I have a 4090. I used the Creative Studio driver, not the game driver.
  4. Load the workflow from the Comfy page in [Command ''C:\Program Files (x86)\Microsoft Visual Studio\2019\BuildTools\VC\Tools\MSVC\14.29.30133\bin\Hostx64\x64\cl.EXE' #1 above. It referenced a different Hunyuan model and CLIP than I had previously downloaded. Make sure to grab the models needed.

That's pretty much it. It now runs without a hitch.

If you're still trying to get Hunyuan to run and are fighting the Triton/Sageattention/CUDA DLL errors, you can safely give up that merry go round and get it working.

You know you could just not use sageattention or torch.compile...? Then you don't need to do any of that as it's only to make things faster, not a requirement for either the wrapper or Comfy native implementation. ComfyUI natively now finally also has support for sageattention, but to run it with that, it's the same hoops to jump through.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants