Traceback (most recent call last): File "F:\ComfyUI-aki-v1.6\ComfyUI\nodes.py", line 2141, in load_custom_node module_spec.loader.exec_module(module) File "<enhanced_experience utils.hook>", line 352, in exec_module File "F:\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-mixlab-nodes\__init__.py", line 1009, in <module> from .nodes.Audio import AudioPlayNode,SpeechRecognition,SpeechSynthesis,AnalyzeAudioNone File "F:\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-mixlab-nodes\nodes\Audio.py", line 4, in <module> import torchaudio File "F:\ComfyUI-aki-v1.6\python\Lib\site-packages\torchaudio\__init__.py", line 2, in <module> from . import _extension

时间: 2025-08-31 12:08:47 AIGC 浏览: 25
<think>好的,我现在需要解决用户在导入torchaudio时遇到的错误,特别是在ComfyUI自定义节点中出现的与_extension模块相关的错误。根据用户提供的引用信息,尤其是引用[1]、[2]、[4],我需要分析可能的原因并提出解决方案。 首先,引用[1]提到调用了torch.ops.torchaudio.sox_effects_apply_effects_file函数,这说明问题可能出现在torchaudio的扩展模块加载过程中。引用[2]中的错误信息显示在编译扩展时出现了RuntimeError,可能与C++扩展编译失败有关。而引用[4]提到AttributeError,指出torch.library缺少register_fake属性,这通常与PyTorch版本不兼容有关。 接下来,我需要考虑可能的解决方案。首先检查PyTorch和torchaudio的版本兼容性,因为版本不匹配是常见问题。用户可能需要升级或降级到兼容的版本组合。例如,引用[4]中建议安装torch==2.4.0和torchvision==0.19.0,这可能同样适用于torchaudio。此外,编译扩展时缺少依赖项(如libsox)也可能导致问题,需要确保系统安装了必要的依赖库。 另一个可能性是Python环境的问题,比如环境路径设置错误或存在多个Python环境导致冲突。用户应检查当前使用的Python环境是否正确,并尝试在干净的环境中重新安装依赖。 此外,如果错误涉及UnicodeDecodeError(如引用[3]中的情况),可能需要处理文件编码问题,但当前问题更可能与扩展模块编译或版本不兼容有关。 综合以上分析,解决方案可能包括:升级/降级PyTorch和torchaudio到兼容版本、安装系统依赖项、检查Python环境,以及重新安装torchaudio。需要逐步指导用户执行这些步骤,并验证每一步是否解决了问题。</think>在导入torchaudio时出现扩展模块相关的错误,通常与PyTorch版本兼容性、系统依赖或编译环境有关。以下是分步解决方案: 1. **验证PyTorch和torchaudio版本兼容性** - 运行以下命令查看当前版本: ```bash pip list | grep -E "torch|torchaudio" ``` - 根据[PyTorch官方版本矩阵](https://pytorch.org/get-started/previous-versions/),安装匹配的组合: ```bash # 例如安装1.13.0版本组合 pip install torch==1.13.0 torchaudio==0.13.0 ``` 2. **安装系统级依赖** - **Linux**:`sudo apt-get install sox libsox-dev libsox-fmt-all` - **Windows**:需通过[sox-binaries](https://sourceforge.net/projects/sox/)手动安装并添加环境变量 3. **强制重新编译扩展模块** ```bash pip uninstall torchaudio -y CXX=g++ pip install torchaudio --no-cache-dir --force-reinstall ``` 4. **检查CUDA工具链一致性** ```python import torch print(torch.__version__, torch.cuda.is_available()) # 应输出True ``` 如果CUDA不可用,需重新安装对应CUDA版本的PyTorch 5. **清理Python环境(针对多环境冲突)** ```bash python -m venv clean_env source clean_env/bin/activate # Linux/Mac # 或 clean_env\Scripts\activate # Windows pip install -U pip setuptools wheel ``` [^1]: 扩展模块加载涉及底层C++代码交互,版本不匹配会导致符号表错误 [^2]: 编译错误常因缺少开发依赖或编译器版本不兼容导致 [^4]: PyTorch 2.x版本部分API有破坏性变更,需严格匹配子模块版本
阅读全文

相关推荐

After restarting ComfyUI, please refresh the browser. translate\translate.go:25:2: github.com/abadojack/[email protected]: Get "https://proxy.golang.org/github.com/abadojack/whatlanggo/@v/v1.0.1.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. translate\translate.go:28:2: github.com/andybalholm/[email protected]: Get "https://proxy.golang.org/github.com/andybalholm/brotli/@v/v1.1.0.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. translate\translate.go:26:2: github.com/imroc/req/[email protected]: Get "https://proxy.golang.org/github.com/imroc/req/v3/@v/v3.48.0.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. translate\translate.go:29:2: github.com/tidwall/[email protected]: Get "https://proxy.golang.org/github.com/tidwall/gjson/@v/v1.14.3.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. service\service.go:23:2: github.com/gin-contrib/[email protected]: Get "https://proxy.golang.org/github.com/gin-contrib/cors/@v/v1.6.0.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. service\service.go:24:2: github.com/gin-gonic/[email protected]: Get "https://proxy.golang.org/github.com/gin-gonic/gin/@v/v1.9.1.zip": dial tcp 142.251.33.81:443: connectex: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond. [ComfyUI-Manager] Failed to perform initial fetching 'custom-node-list.json': Traceback (most recent call last): File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 347, in _wait await waiter asyncio.exceptions.CancelledError The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-manager\glob\manager_server.py", line 1729, in get_cache json_obj = await manager_util.get_data(uri, True) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-manager\glob\manager_util.py", line 146, in get_data json_text = await resp.text() ^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\client_reqrep.py", line 1260, in text await self.read() File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\client_reqrep.py", line 1218, in read self._body = await self.content.read() ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 418, in read block = await self.readany() ^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 440, in readany await self._wait("readany") File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 346, in _wait with self._timer: File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\helpers.py", line 685, in __exit__ raise asyncio.TimeoutError from exc_val TimeoutError [ComfyUI-Manager] Failed to perform initial fetching 'extension-node-map.json': Traceback (most recent call last): File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 347, in _wait await waiter asyncio.exceptions.CancelledError The above exception was the direct cause of the following exception: Traceback (most recent call last): File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-manager\glob\manager_server.py", line 1729, in get_cache json_obj = await manager_util.get_data(uri, True) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-manager\glob\manager_util.py", line 146, in get_data json_text = await resp.text() ^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\client_reqrep.py", line 1260, in text await self.read() File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\client_reqrep.py", line 1218, in read self._body = await self.content.read() ^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 418, in read block = await self.readany() ^^^^^^^^^^^^^^^^^^^^ File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 440, in readany await self._wait("readany") File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\streams.py", line 346, in _wait with self._timer: File "C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\aiohttp\helpers.py", line 685, in __exit__ raise asyncio.TimeoutError from exc_val TimeoutError FETCH ComfyRegistry Data: 5/85 FETCH ComfyRegistry Data: 10/85 FETCH ComfyRegistry Data: 15/85 FETCH ComfyRegistry Data: 20/85 FETCH ComfyRegistry Data: 25/85 FETCH ComfyRegistry Data: 30/85 FETCH ComfyRegistry Data: 35/85 FETCH ComfyRegistry Data: 40/85 FETCH ComfyRegistry Data: 45/85 FETCH ComfyRegistry Data: 50/85 FETCH ComfyRegistry Data: 55/85 FETCH ComfyRegistry Data: 60/85 FETCH ComfyRegistry Data: 65/85 FETCH ComfyRegistry Data: 70/85 FETCH ComfyRegistry Data: 75/85 FETCH ComfyRegistry Data: 80/85 FETCH ComfyRegistry Data: 85/85 FETCH ComfyRegistry Data [DONE] [ComfyUI-Manager] default cache updated: https://api.comfy.org/nodes FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json[ComfyUI-Manager] Due to a network error, switching to local mode. => custom-node-list.json => FETCH DATA from: C:\Documents\Ai\ComfyUI-aki\ComfyUI-aki-v1.6\ComfyUI\custom_nodes\comfyui-manager\custom-node-list.json [DONE] [ComfyUI-Manager] All startup tasks have been completed. [Crystools ERROR] For some reason, pynvml is not working in a laptop with only battery, try to connect and turn on the monitor [Crystools ERROR] Monitor of GPU is turning off (not on UI!) [BizyAir] Fail to get notification unread counts: Cannot connect to host api.bizyair.cn:443 ssl:default [getaddrinfo failed] 我在comfyUI中安装nunchaku有些问题 这个应该怎么修复呢

got prompt Failed to validate prompt for output 63: * UNETLoader 11: - Value not in list: unet_name: 'None' not in [] * LoraLoaderModelOnly 37: - Value not in list: lora_name: 'None' not in [] * DualCLIPLoader 14: - Value not in list: clip_name1: 'None' not in [] - Value not in list: clip_name2: 'None' not in [] Output will be ignored Failed to validate prompt for output 9: Output will be ignored !!! Exception during processing !!! Cannot find an appropriate cached snapshot folder for the specified revision on the local disk and outgoing traffic has been disabled. To enable repo look-ups and downloads online, set 'HF_HUB_OFFLINE=0' as environment variable. Traceback (most recent call last): File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\huggingface_hub\_snapshot_download.py", line 155, in snapshot_download repo_info = api.repo_info(repo_id=repo_id, repo_type=repo_type, revision=revision, token=token) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\huggingface_hub\hf_api.py", line 2748, in repo_info return method( ^^^^^^^ File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\huggingface_hub\utils\_validators.py", line 114, in _inner_fn return fn(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\huggingface_hub\hf_api.py", line 2532, in model_info r = get_session().get(path, headers=headers, timeout=timeout, params=params) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI-aki\ComfyUI-aki-v1.6\python\Lib\site-packages\requests\sessions.py", line 602, in get return self.request("GET", url, **

aria2: http://192.168.0.100:6800 [START] Security scan [DONE] Security scan ## ComfyUI-Manager: installing dependencies done. ** ComfyUI startup time: 2025-08-18 01:01:22.434 ** Platform: Windows ** Python version: 3.12.11 (main, Aug 8 2025, 17:05:04) [MSC v.1944 64 bit (AMD64)] ** Python executable: E:\ComfyUI\.venv\Scripts\python.exe ** ComfyUI Path: E:\ComfyUI\ComfyUI ** ComfyUI Base Folder Path: E:\ComfyUI\ComfyUI ** User directory: E:\ComfyUI\ComfyUI\user ** ComfyUI-Manager config path: E:\ComfyUI\ComfyUI\user\default\ComfyUI-Manager\config.ini ** Log path: E:\ComfyUI\ComfyUI\user\comfyui.log Prestartup times for custom nodes: 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-easy-use 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\rgthree-comfy 2.6 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-Manager Checkpoint files will always be loaded safely. Total VRAM 16303 MB, total RAM 130789 MB pytorch version: 2.8.0+cu128 Set vram state to: NORMAL_VRAM Device: cuda:0 NVIDIA GeForce RTX 5080 : cudaMallocAsync Using pytorch attention Python version: 3.12.11 (main, Aug 8 2025, 17:05:04) [MSC v.1944 64 bit (AMD64)] ComfyUI version: 0.3.50 ComfyUI frontend version: 1.24.4 [Prompt Server] web root: E:\ComfyUI\.venv\Lib\site-packages\comfyui_frontend_package\static Error: [WinError 1314] 客户端没有所需的特权。: 'E:\\ComfyUI\\ComfyUI\\custom_nodes\\ComfyLiterals\\js' -> 'E:\\ComfyUI\\ComfyUI\\web\\extensions\\ComfyLiterals' Failed to create symlink to E:\ComfyUI\ComfyUI\web\extensions\ComfyLiterals. Please copy the folder manually. Source: E:\ComfyUI\ComfyUI\custom_nodes\ComfyLiterals\js Target: E:\ComfyUI\ComfyUI\web\extensions\ComfyLiterals WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for: PyTorch 2.7.1+cu128 with CUDA 1208 (you have 2.8.0+cu128) Python 3.9.13 (you have 3.12.11) Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers) Memory-efficient attention, SwiGLU, sparse and more won't be available. Set XFORMERS_MORE_DETAILS=1 for more details xFormers not available xFormers not available Module 'diffusers' load failed. If you don't have it installed, do it: pip install diffusers [ComfyUI-Easy-Use] server: v1.3.2 Loaded [ComfyUI-Easy-Use] web root: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-easy-use\web_version/v2 Loaded theUpsiders Logic Nodes: Loaded ### Loading: ComfyUI-Manager (V3.36) [ComfyUI-Manager] network_mode: public ### ComfyUI Version: v0.3.50-6-ge4f7ea10 | Released on '2025-08-13' [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json -------------- ### Mixlab Nodes: Loaded json_repair## OK ChatGPT.available True edit_mask.available True ## clip_interrogator_model not found: E:\ComfyUI\ComfyUI\models\clip_interrogator\Salesforce\blip-image-captioning-base, pls download from https://huggingface.co/Salesforce/blip-image-captioning-base ClipInterrogator.available True ## text_generator_model not found: E:\ComfyUI\ComfyUI\models\prompt_generator\text2image-prompt-generator, pls download from https://huggingface.co/succinctly/text2image-prompt-generator/tree/main ## zh_en_model not found: E:\ComfyUI\ComfyUI\models\prompt_generator\opus-mt-zh-en, pls download from https://huggingface.co/Helsinki-NLP/opus-mt-zh-en/tree/main PromptGenerate.available True ChinesePrompt.available True RembgNode_.available True ffmpeg could not be found. Using ffmpeg from imageio-ffmpeg. Use Proxy: http://192.168.0.100:7897 TripoSR.available MiniCPMNode.available Scenedetect.available FishSpeech.available SenseVoice.available False Whisper.available False fal-client## OK FalVideo.available -------------- Use Proxy: http://192.168.0.100:7897 ======================================== ComfyUI-nunchaku Initialization ======================================== Nunchaku version: 0.3.1 ComfyUI-nunchaku version: 0.3.3 Traceback (most recent call last): File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\utils\import_utils.py", line 883, in _get_module return importlib.import_module("." + module_name, self.__name__) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\user\AppData\Roaming\uv\python\cpython-3.12.11-windows-x86_64-none\Lib\importlib\__init__.py", line 90, in import_module return _bootstrap._gcd_import(name[level:], package, level) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "<frozen importlib._bootstrap>", line 1387, in _gcd_import File "<frozen importlib._bootstrap>", line 1360, in _find_and_load File "<frozen importlib._bootstrap>", line 1310, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed File "<frozen importlib._bootstrap>", line 1387, in _gcd_import File "<frozen importlib._bootstrap>", line 1360, in _find_and_load File "<frozen importlib._bootstrap>", line 1331, in _find_and_load_unlocked File "<frozen importlib._bootstrap>", line 935, in _load_unlocked File "<frozen importlib._bootstrap_external>", line 999, in exec_module File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\models\transformers\__init__.py", line 5, in <module> from .auraflow_transformer_2d import AuraFlowTransformer2DModel File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\models\transformers\auraflow_transformer_2d.py", line 26, in <module> from ..attention_processor import ( File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\models\attention_processor.py", line 35, in <module> import xformers.ops File "E:\ComfyUI\.venv\Lib\site-packages\xformers\ops\__init__.py", line 9, in <module> from .fmha import ( File "E:\ComfyUI\.venv\Lib\site-packages\xformers\ops\fmha\__init__.py", line 10, in <module> from . import ( File "E:\ComfyUI\.venv\Lib\site-packages\xformers\ops\fmha\flash3.py", line 109, in <module> from ...flash_attn_3 import _C # type: ignore[attr-defined] # noqa: F401 ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ ImportError: DLL load failed while importing _C: 找不到指定的程序。 The above exception was the direct cause of the following exception: Traceback (most recent call last): File "E:\ComfyUI\ComfyUI\nodes.py", line 2129, in load_custom_node module_spec.loader.exec_module(module) File "<frozen importlib._bootstrap_external>", line 999, in exec_module File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed File "E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-nunchaku\__init__.py", line 31, in <module> from .nodes.models.flux import NunchakuFluxDiTLoader File "E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-nunchaku\nodes\models\flux.py", line 11, in <module> from diffusers import FluxTransformer2DModel File "<frozen importlib._bootstrap>", line 1412, in _handle_fromlist File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\utils\import_utils.py", line 874, in __getattr__ value = getattr(module, name) ^^^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\utils\import_utils.py", line 873, in __getattr__ module = self._get_module(self._class_to_module[name]) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "E:\ComfyUI\.venv\Lib\site-packages\diffusers\utils\import_utils.py", line 885, in _get_module raise RuntimeError( RuntimeError: Failed to import diffusers.models.transformers.transformer_flux because of the following error (look up to see its traceback): DLL load failed while importing _C: 找不到指定的程序。 Cannot import E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-nunchaku module for custom nodes: Failed to import diffusers.models.transformers.transformer_flux because of the following error (look up to see its traceback): DLL load failed while importing _C: 找不到指定的程序。 Error loading AILab_SegmentV2.py: No module named 'groundingdino' [ComfyUI-RMBG] v2.5.0 | 26 nodes Loaded ------------------------------------------ Comfyroll Studio v1.76 : 175 Nodes Loaded ------------------------------------------ ** For changes, please see patch notes at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/blob/main/Patch_Notes.md ** For help, please see the wiki at https://github.com/Suzie1/ComfyUI_Comfyroll_CustomNodes/wiki ------------------------------------------ [E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux] | INFO -> Using ckpts path: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux\ckpts [E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux] | INFO -> Using symlinks: False [E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux] | INFO -> Using ort providers: ['CUDAExecutionProvider', 'DirectMLExecutionProvider', 'OpenVINOExecutionProvider', 'ROCMExecutionProvider', 'CPUExecutionProvider', 'CoreMLExecutionProvider'] E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux\node_wrappers\dwpose.py:26: UserWarning: DWPose: Onnxruntime not found or doesn't come with acceleration providers, switch to OpenCV with CPU device. DWPose might run very slowly warnings.warn("DWPose: Onnxruntime not found or doesn't come with acceleration providers, switch to OpenCV with CPU device. DWPose might run very slowly") Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 5/94 ### [START] ComfyUI AlekPet Nodes v1.0.77 ### Node -> ArgosTranslateNode: ArgosTranslateCLIPTextEncodeNode, ArgosTranslateTextNode [Loading] Node -> ChatGLMNode[Loading] Node -> DeepLXTranslateNode[Loading] Node -> DeepTranslatorNode: DeepTranslatorCLIPTextEncodeNode, DeepTranslatorTextNode [Loading] Node -> ExtrasNode: PreviewTextNode, HexToHueNode, ColorsCorrectNode [Loading] Node -> GoogleTranslateNode: GoogleTranslateCLIPTextEncodeNode, GoogleTranslateTextNode [Loading] Node -> IDENode: IDENode [Loading] Node -> PainterNode: PainterNode [Loading] Node -> PoseNode: PoseNode [Loading] ### [END] ComfyUI AlekPet Nodes ### # 😺dzNodes: LayerStyle -> Cannot import name 'guidedFilter' from 'cv2.ximgproc' A few nodes cannot works properly, while most nodes are not affected. Please REINSTALL package 'opencv-contrib-python'. For detail refer to https://github.com/chflame163/ComfyUI_LayerStyle/issues/5 Initializing ControlAltAI Nodes Use Proxy: http://192.168.0.100:7897 [rgthree-comfy] Loaded 47 magnificent nodes. 🎉 Import times for custom nodes: 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\websocket_image_save.py 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_ttp_toolset 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-inpainteasy 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyLiterals 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_slk_joy_caption_two 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-logic 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI_IPAdapter_plus 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-custom-scripts 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-inpaint-nodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_essentials 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ControlAltAI-Nodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-advanced-controlnet 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-KJNodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\derfuu_comfyui_moddednodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-RMBG 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-tooling-nodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI_Mira 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI_Comfyroll_CustomNodes 0.0 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_controlnet_aux 0.1 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-model-manager 0.1 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-Dwpose-Tensorrt 0.1 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI_LayerStyle 0.2 seconds (IMPORT FAILED): E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-nunchaku 0.3 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-florence2 0.4 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-Depth-Anything-Tensorrt 0.5 seconds: E:\ComfyUI\ComfyUI\custom_nodes\ComfyUI-Manager 0.7 seconds: E:\ComfyUI\ComfyUI\custom_nodes\rgthree-comfy 1.7 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-depthanythingv2 1.8 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui_custom_nodes_alekpet 1.8 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-easy-use 4.5 seconds: E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes Context impl SQLiteImpl. Will assume non-transactional DDL. No target revision found. Starting server To see the GUI go to: http://0.0.0.0:8188 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/photoswipe-lightbox.esm.min.js E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/pickr.min.js E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/photoswipe.min.css E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/classic.min.css E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/model-viewer.min.js E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/juxtapose.css E:\ComfyUI\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/juxtapose.min.js Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 10/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 15/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 20/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 25/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 30/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 35/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 40/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 45/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 50/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 55/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 60/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 65/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 70/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 75/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 80/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 85/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data: 90/94 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 Use Proxy: http://192.168.0.100:7897 FETCH ComfyRegistry Data [DONE] [ComfyUI-Manager] default cache updated: https://api.comfy.org/nodes mode: remote filename: custom-node-list.json channel_url: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.jsonUse Proxy: http://192.168.0.100:7897 [DONE] Merge json from \\192.168.0.100\mcs\ComfyUI\custom-list;\\192.168.0.100\mcs\ComfyUI\civitai-list Merge json from \\192.168.0.100\mcs\ComfyUI\custom-list\custom-node-list.json FETCH DATA from: \\192.168.0.100\mcs\ComfyUI\custom-list\custom-node-list.json [DONE] merge custom_nodes [ComfyUI-Manager] broken item:{'title': 'ComfyUI Dwpose TensorRT', 'reference': 'https://github.com/yuvraj108c/ComfyUI-Dwpose-Tensorrt', 'req_models': [{'name': 'DWPose-l ucoco 384', 'reference': 'https://huggingface.co/yzd-v/DWPose'}, {'name': 'DWPose yolox l', 'reference': 'https://huggingface.co/yzd-v/DWPose'}]} [ComfyUI-Manager] broken item:{'title': 'ComfyUI Depth Anything TensorRT', 'reference': 'https://github.com/yuvraj108c/ComfyUI-Depth-Anything-Tensorrt', 'req_models': [{'name': 'Depth Anything 2 Onnx Large', 'reference': 'https://huggingface.co/yuvraj108c/Depth-Anything-2-Onnx'}]} [ComfyUI-Manager] All startup tasks have been completed.

# ComfyUI Error Report ## Error Details - **Node ID:** 42 - **Node Type:** WanVideoSampler - **Exception Type:** AttributeError - **Exception Message:** 'WanModel' object has no attribute 'vace_patch_embedding' ## Stack Trace File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 361, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 236, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 208, in _map_node_over_list process_inputs(input_dict, i) File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 197, in process_inputs results.append(getattr(obj, func)(**inputs)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3325, in process noise_pred, self.cache_state = predict_with_cfg( ^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 2617, in predict_with_cfg noise_pred_cond, cache_state_cond = transformer( ^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1617, in forward vace_hints = self.forward_vace(x, data["context"], data["seq_len"], kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1196, in forward_vace c = [self.vace_patch_embedding(u.unsqueeze(0).float()).to(x.dtype) for u in vace_context] ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\diffusers\models\modeling_utils.py", line 291, in __getattr__ return super().__getattr__(name) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1940, in __getattr__ raise AttributeError( ## System Information - **ComfyUI Version:** 0.3.43 - **Arguments:** ComfyUI\main.py --windows-standalone-build --listen 0.0.0.0 --enable-cors-header * - **OS:** nt - **Python Version:** 3.12.10 (tags/v3.12.10:0cc8128, Apr 8 2025, 12:21:36) [MSC v.1943 64 bit (AMD64)] - **Embedded Python:** true - **PyTorch Version:** 2.7.1+cu128 ## Devices - **Name:** cuda:0 NVIDIA GeForce RTX 3090 : cudaMallocAsync - **Type:** cuda - **VRAM Total:** 25769279488 - **VRAM Free:** 24347803648 - **Torch VRAM Total:** 67108864 - **Torch VRAM Free:** 50200576 ## Logs 2025-07-24T17:17:54.612419 - 2025-07-24T17:17:54.663419 - INFO: IPAdapter model loaded from F:\ComfyUI_Mie_V6.0\ComfyUI\models\ipadapter\ip-adapter-plus_sd15.safetensors2025-07-24T17:17:54.663419 - 2025-07-24T17:17:54.663419 - end_vram - start_vram: 2537542306 - 2537542306 = 02025-07-24T17:17:54.663419 - 2025-07-24T17:17:54.664419 - #11 [IPAdapterUnifiedLoader]: 0.05s - vram 0b2025-07-24T17:17:54.664419 - 2025-07-24T17:17:54.924153 - end_vram - start_vram: 2639727650 - 2537542306 = 1021853442025-07-24T17:17:54.924153 - 2025-07-24T17:17:54.924670 - #10 [IPAdapterAdvanced]: 0.26s - vram 102185344b2025-07-24T17:17:54.924670 - 2025-07-24T17:17:54.931880 - end_vram - start_vram: 2733990564 - 2635766434 = 982241302025-07-24T17:17:54.931880 - 2025-07-24T17:17:54.931880 - #14 [RescaleCFG]: 0.01s - vram 98224130b2025-07-24T17:17:54.931880 - 2025-07-24T17:17:54.933882 - Requested to load BaseModel 2025-07-24T17:17:55.233473 - loaded completely 19197.46322517395 1639.406135559082 True 2025-07-24T17:18:23.076833 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.07it/s]2025-07-24T17:18:23.076833 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.08it/s]2025-07-24T17:18:23.076833 - 2025-07-24T17:18:23.078833 - end_vram - start_vram: 4711053742 - 2733990562 = 19770631802025-07-24T17:18:23.078833 - 2025-07-24T17:18:23.078833 - #3 [KSampler]: 28.15s - vram 1977063180b2025-07-24T17:18:23.078833 - 2025-07-24T17:18:23.711202 - end_vram - start_vram: 8100460554 - 3605411690 = 44950488642025-07-24T17:18:23.711239 - 2025-07-24T17:18:23.711239 - #8 [VAEDecode]: 0.63s - vram 4495048864b2025-07-24T17:18:23.711754 - 2025-07-24T17:18:23.864142 - end_vram - start_vram: 3605411690 - 3605411690 = 02025-07-24T17:18:23.864142 - 2025-07-24T17:18:23.864142 - #15 [SaveImage]: 0.15s - vram 0b2025-07-24T17:18:23.865143 - 2025-07-24T17:18:23.865143 - comfyui lumi batcher overwrite task done2025-07-24T17:18:23.865143 - 2025-07-24T17:18:23.866143 - Prompt executed in 31.05 seconds 2025-07-24T17:18:26.480125 - got prompt 2025-07-24T17:18:53.685302 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.14it/s]2025-07-24T17:18:53.686303 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.11it/s]2025-07-24T17:18:53.686303 - 2025-07-24T17:18:53.686303 - end_vram - start_vram: 4689507218 - 3605411690 = 10840955282025-07-24T17:18:53.686303 - 2025-07-24T17:18:53.687302 - #3 [KSampler]: 27.17s - vram 1084095528b2025-07-24T17:18:53.687302 - 2025-07-24T17:18:54.232028 - end_vram - start_vram: 8100460554 - 3605411690 = 44950488642025-07-24T17:18:54.232028 - 2025-07-24T17:18:54.233027 - #8 [VAEDecode]: 0.54s - vram 4495048864b2025-07-24T17:18:54.233027 - 2025-07-24T17:18:54.380027 - end_vram - start_vram: 3605411690 - 3605411690 = 02025-07-24T17:18:54.380027 - 2025-07-24T17:18:54.380027 - #15 [SaveImage]: 0.15s - vram 0b2025-07-24T17:18:54.380027 - 2025-07-24T17:18:54.381027 - comfyui lumi batcher overwrite task done2025-07-24T17:18:54.382027 - 2025-07-24T17:18:54.385027 - Prompt executed in 27.90 seconds 2025-07-24T17:24:25.352704 - got prompt 2025-07-24T17:24:25.445704 - end_vram - start_vram: 3507187562 - 3507187562 = 02025-07-24T17:24:25.446704 - 2025-07-24T17:24:25.446704 - #20 [LoraLoader]: 0.05s - vram 0b2025-07-24T17:24:25.446704 - 2025-07-24T17:24:25.447704 - Requested to load SD1ClipModel 2025-07-24T17:24:25.792290 - loaded completely 18695.929541397094 235.84423828125 True 2025-07-24T17:24:25.800290 - end_vram - start_vram: 3521486046 - 3507187562 = 142984842025-07-24T17:24:25.800290 - 2025-07-24T17:24:25.800290 - #7 [CLIPTextEncode]: 0.35s - vram 14298484b2025-07-24T17:24:25.800290 - 2025-07-24T17:24:25.808290 - end_vram - start_vram: 3519593078 - 3507187562 = 124055162025-07-24T17:24:25.808290 - 2025-07-24T17:24:25.809290 - #6 [CLIPTextEncode]: 0.01s - vram 12405516b2025-07-24T17:24:25.809290 - 2025-07-24T17:24:25.810290 - end_vram - start_vram: 3507187562 - 3507187562 = 02025-07-24T17:24:25.810290 - 2025-07-24T17:24:25.810290 - #11 [IPAdapterUnifiedLoader]: 0.00s - vram 0b2025-07-24T17:24:25.810290 - 2025-07-24T17:24:26.053290 - end_vram - start_vram: 3609372906 - 3507187562 = 1021853442025-07-24T17:24:26.053290 - 2025-07-24T17:24:26.054290 - #10 [IPAdapterAdvanced]: 0.24s - vram 102185344b2025-07-24T17:24:26.054290 - 2025-07-24T17:24:26.059290 - end_vram - start_vram: 3703635820 - 3605411690 = 982241302025-07-24T17:24:26.059290 - 2025-07-24T17:24:26.060290 - #14 [RescaleCFG]: 0.01s - vram 98224130b2025-07-24T17:24:26.060290 - 2025-07-24T17:24:26.061290 - Requested to load BaseModel 2025-07-24T17:24:26.644908 - loaded completely 19912.1512134552 1639.406135559082 True 2025-07-24T17:24:53.885378 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.08it/s]2025-07-24T17:24:53.885378 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:27<00:00, 1.10it/s]2025-07-24T17:24:53.885378 - 2025-07-24T17:24:53.886378 - end_vram - start_vram: 4689507218 - 3703635818 = 9858714002025-07-24T17:24:53.886378 - 2025-07-24T17:24:53.886378 - #3 [KSampler]: 27.83s - vram 985871400b2025-07-24T17:24:53.887378 - 2025-07-24T17:24:54.482797 - end_vram - start_vram: 8100460554 - 3605411690 = 44950488642025-07-24T17:24:54.482797 - 2025-07-24T17:24:54.482797 - #8 [VAEDecode]: 0.60s - vram 4495048864b2025-07-24T17:24:54.482797 - 2025-07-24T17:24:54.640169 - end_vram - start_vram: 3605411690 - 3605411690 = 02025-07-24T17:24:54.640169 - 2025-07-24T17:24:54.640169 - #15 [SaveImage]: 0.16s - vram 0b2025-07-24T17:24:54.640169 - 2025-07-24T17:24:54.641169 - comfyui lumi batcher overwrite task done2025-07-24T17:24:54.641169 - 2025-07-24T17:24:54.643410 - Prompt executed in 29.28 seconds 2025-07-24T17:25:09.991282 - got prompt 2025-07-24T17:25:10.085282 - end_vram - start_vram: 3507187562 - 3507187562 = 02025-07-24T17:25:10.085282 - 2025-07-24T17:25:10.085282 - #20 [LoraLoader]: 0.05s - vram 0b2025-07-24T17:25:10.085282 - 2025-07-24T17:25:10.086282 - Requested to load SD1ClipModel 2025-07-24T17:25:10.369134 - loaded completely 18695.929541397094 235.84423828125 True 2025-07-24T17:25:10.377134 - end_vram - start_vram: 3521486046 - 3507187562 = 142984842025-07-24T17:25:10.377134 - 2025-07-24T17:25:10.377134 - #7 [CLIPTextEncode]: 0.29s - vram 14298484b2025-07-24T17:25:10.377134 - 2025-07-24T17:25:10.385134 - end_vram - start_vram: 3519593078 - 3507187562 = 124055162025-07-24T17:25:10.385134 - 2025-07-24T17:25:10.385134 - #6 [CLIPTextEncode]: 0.01s - vram 12405516b2025-07-24T17:25:10.385134 - 2025-07-24T17:25:10.386135 - end_vram - start_vram: 3507187562 - 3507187562 = 02025-07-24T17:25:10.386135 - 2025-07-24T17:25:10.386135 - #11 [IPAdapterUnifiedLoader]: 0.00s - vram 0b2025-07-24T17:25:10.386135 - 2025-07-24T17:25:10.631046 - end_vram - start_vram: 3609372906 - 3507187562 = 1021853442025-07-24T17:25:10.631046 - 2025-07-24T17:25:10.631046 - #10 [IPAdapterAdvanced]: 0.24s - vram 102185344b2025-07-24T17:25:10.631046 - 2025-07-24T17:25:10.637046 - end_vram - start_vram: 3703635820 - 3605411690 = 982241302025-07-24T17:25:10.637046 - 2025-07-24T17:25:10.637046 - #14 [RescaleCFG]: 0.01s - vram 98224130b2025-07-24T17:25:10.637046 - 2025-07-24T17:25:10.639047 - Requested to load BaseModel 2025-07-24T17:25:11.180881 - loaded completely 19912.1512134552 1639.406135559082 True 2025-07-24T17:25:39.272121 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:28<00:00, 1.08it/s]2025-07-24T17:25:39.272121 - 100%|██████████████████████████████████████████████████████████████████████████████████| 30/30 [00:28<00:00, 1.07it/s]2025-07-24T17:25:39.272121 - 2025-07-24T17:25:39.273121 - end_vram - start_vram: 4689507218 - 3703635818 = 9858714002025-07-24T17:25:39.273121 - 2025-07-24T17:25:39.274121 - #3 [KSampler]: 28.64s - vram 985871400b2025-07-24T17:25:39.274121 - 2025-07-24T17:25:39.888702 - end_vram - start_vram: 8100460554 - 3605411690 = 44950488642025-07-24T17:25:39.888702 - 2025-07-24T17:25:39.888702 - #8 [VAEDecode]: 0.61s - vram 4495048864b2025-07-24T17:25:39.888702 - 2025-07-24T17:25:40.051703 - end_vram - start_vram: 3605411690 - 3605411690 = 02025-07-24T17:25:40.051703 - 2025-07-24T17:25:40.051703 - #15 [SaveImage]: 0.16s - vram 0b2025-07-24T17:25:40.051703 - 2025-07-24T17:25:40.052703 - comfyui lumi batcher overwrite task done2025-07-24T17:25:40.052703 - 2025-07-24T17:25:40.060703 - Prompt executed in 30.06 seconds 2025-07-24T17:43:56.476860 - got prompt 2025-07-24T17:43:57.039587 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:43:57.039587 - 2025-07-24T17:43:57.039587 - #7 [WanVideoVAELoader]: 0.48s - vram 0b2025-07-24T17:43:57.040588 - 2025-07-24T17:43:57.040588 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:43:57.040588 - 2025-07-24T17:43:57.040588 - #12 [WanVideoExperimentalArgs]: 0.00s - vram 0b2025-07-24T17:43:57.040588 - 2025-07-24T17:43:57.041588 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:43:57.041588 - 2025-07-24T17:43:57.041588 - #10 [WanVideoEnhanceAVideo]: 0.00s - vram 0b2025-07-24T17:43:57.041588 - 2025-07-24T17:43:57.042588 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:43:57.042588 - 2025-07-24T17:43:57.042588 - #21 [WanVideoVACEModelSelect]: 0.00s - vram 0b2025-07-24T17:43:57.042588 - 2025-07-24T17:44:00.282807 - Detected model in_channels: 36 2025-07-24T17:44:00.282807 - Model type: fl2v, num_heads: 40, num_layers: 40 2025-07-24T17:44:00.283807 - Model variant detected: i2v_720 2025-07-24T17:44:00.596667 - model_type FLOW 2025-07-24T17:44:00.597701 - Using accelerate to load and assign model weights to device... 2025-07-24T17:44:00.801188 - Loading transformer parameters to cpu: 83%|██████████████████████████████▋ | 1080/1304 [00:00<00:00, 5458.86it/s]2025-07-24T17:44:00.839187 - Loading transformer parameters to cpu: 100%|█████████████████████████████████████| 1304/1304 [00:00<00:00, 5446.33it/s]2025-07-24T17:44:00.839187 - 2025-07-24T17:44:00.844187 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:44:00.844187 - 2025-07-24T17:44:00.844187 - #22 [WanVideoModelLoader]: 3.80s - vram 0b2025-07-24T17:44:00.844187 - 2025-07-24T17:44:02.947231 - end_vram - start_vram: 106744016 - 106744016 = 02025-07-24T17:44:02.948230 - 2025-07-24T17:44:02.948230 - #6 [LoadWanVideoT5TextEncoder]: 2.10s - vram 0b2025-07-24T17:44:02.948230 - 2025-07-24T17:44:02.948230 - Moving video model to cpu 2025-07-24T17:45:21.569551 - end_vram - start_vram: 11728619736 - 106744016 = 116218757202025-07-24T17:45:21.569558 - 2025-07-24T17:45:21.569558 - #8 [WanVideoTextEncode]: 78.62s - vram 11621875720b2025-07-24T17:45:21.569558 - 2025-07-24T17:45:21.624655 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.625215 - 2025-07-24T17:45:21.625215 - #34 [LoadImage]: 0.05s - vram 0b2025-07-24T17:45:21.625215 - 2025-07-24T17:45:21.626235 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.626242 - 2025-07-24T17:45:21.626242 - #19 [INTConstant]: 0.00s - vram 0b2025-07-24T17:45:21.626242 - 2025-07-24T17:45:21.671234 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.671234 - 2025-07-24T17:45:21.671234 - #24 [ImageResizeKJ]: 0.04s - vram 0b2025-07-24T17:45:21.671234 - 2025-07-24T17:45:21.738079 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.738592 - 2025-07-24T17:45:21.739108 - #30 [ImageResizeKJ]: 0.07s - vram 0b2025-07-24T17:45:21.739108 - 2025-07-24T17:45:21.782786 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.782786 - 2025-07-24T17:45:21.782786 - #35 [ImageResizeKJ]: 0.04s - vram 0b2025-07-24T17:45:21.782786 - 2025-07-24T17:45:21.915384 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.915902 - 2025-07-24T17:45:21.915902 - #29 [WanVideoVACEStartToEndFrame]: 0.13s - vram 0b2025-07-24T17:45:21.915902 - 2025-07-24T17:45:21.916414 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.916918 - 2025-07-24T17:45:21.917002 - #40 [GetImageRangeFromBatch]: 0.00s - vram 0b2025-07-24T17:45:21.917002 - 2025-07-24T17:45:21.981944 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.981944 - 2025-07-24T17:45:21.981944 - #25 [WanVideoVACEStartToEndFrame]: 0.06s - vram 0b2025-07-24T17:45:21.981944 - 2025-07-24T17:45:21.995235 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:21.995235 - 2025-07-24T17:45:21.995235 - #41 [MaskBatchMulti]: 0.01s - vram 0b2025-07-24T17:45:21.995235 - 2025-07-24T17:45:22.039203 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:22.039203 - 2025-07-24T17:45:22.039203 - #37 [ImageBatchMulti]: 0.04s - vram 0b2025-07-24T17:45:22.039708 - 2025-07-24T17:45:22.039713 - end_vram - start_vram: 115132624 - 115132624 = 02025-07-24T17:45:22.040227 - 2025-07-24T17:45:22.040227 - #26 [GetImageSizeAndCount]: 0.00s - vram 0b2025-07-24T17:45:22.040227 - 2025-07-24T17:45:22.883828 - input_masks shape2025-07-24T17:45:22.883828 - 2025-07-24T17:45:22.884341 - torch.Size([81, 496, 896])2025-07-24T17:45:22.884341 - 2025-07-24T17:45:34.863567 - VAE encoding: 100%|██████████████████████████████████████████████████████████████████████| 9/9 [00:11<00:00, 1.30s/it]2025-07-24T17:45:34.863567 - VAE encoding: 100%|██████████████████████████████████████████████████████████████████████| 9/9 [00:11<00:00, 1.32s/it]2025-07-24T17:45:34.864077 - 2025-07-24T17:45:46.589102 - VAE encoding: 100%|██████████████████████████████████████████████████████████████████████| 9/9 [00:11<00:00, 1.30s/it]2025-07-24T17:45:46.589612 - VAE encoding: 100%|██████████████████████████████████████████████████████████████████████| 9/9 [00:11<00:00, 1.30s/it]2025-07-24T17:45:46.589612 - 2025-07-24T17:45:46.600743 - refs shape2025-07-24T17:45:46.600743 - 2025-07-24T17:45:46.600743 - torch.Size([1, 3, 1, 496, 896])2025-07-24T17:45:46.600743 - 2025-07-24T17:45:46.715123 - VAE encoding: 56%|██████████████████████████████████████▉ | 5/9 [00:00<00:00, 44.31it/s]2025-07-24T17:45:46.805051 - VAE encoding: 100%|██████████████████████████████████████████████████████████████████████| 9/9 [00:00<00:00, 44.38it/s]2025-07-24T17:45:46.805583 - 2025-07-24T17:45:46.870662 - end_vram - start_vram: 2252473238 - 115132624 = 21373406142025-07-24T17:45:46.870662 - 2025-07-24T17:45:46.871171 - #16 [WanVideoVACEEncode]: 24.83s - vram 2137340614b2025-07-24T17:45:46.871171 - 2025-07-24T17:45:46.873216 - timesteps: tensor([1000.0000, 983.7124, 964.5367, 941.6301, 913.7864, 879.2166, 835.1485, 777.0445, 696.9283, 579.3684, 390.1099, 34.7826], device='cuda:0') 2025-07-24T17:45:46.887141 - Seq len: 38192 2025-07-24T17:49:31.510530 - Sampling 85 frames at 896x496 with 12 steps 2025-07-24T17:49:37.849674 - 0%| | 0/12 [00:00<?, ?it/s]2025-07-24T17:49:44.296375 - 0%| | 0/12 [00:06<?, ?it/s]2025-07-24T17:49:44.297376 - 2025-07-24T17:49:44.747159 - !!! Exception during processing !!! 'WanModel' object has no attribute 'vace_patch_embedding' 2025-07-24T17:49:44.783076 - Traceback (most recent call last): File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 361, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 236, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 208, in _map_node_over_list process_inputs(input_dict, i) File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 197, in process_inputs results.append(getattr(obj, func)(**inputs)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3325, in process noise_pred, self.cache_state = predict_with_cfg( ^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 2617, in predict_with_cfg noise_pred_cond, cache_state_cond = transformer( ^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1617, in forward vace_hints = self.forward_vace(x, data["context"], data["seq_len"], kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1196, in forward_vace c = [self.vace_patch_embedding(u.unsqueeze(0).float()).to(x.dtype) for u in vace_context] ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\diffusers\models\modeling_utils.py", line 291, in __getattr__ return super().__getattr__(name) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1940, in __getattr__ raise AttributeError( AttributeError: 'WanModel' object has no attribute 'vace_patch_embedding'. Did you mean: 'patch_embedding'? 2025-07-24T17:49:44.785980 - end_vram - start_vram: 34183609272 - 173795536 = 340098137362025-07-24T17:49:44.785980 - 2025-07-24T17:49:44.785980 - #42 [WanVideoSampler]: 237.91s - vram 34009813736b2025-07-24T17:49:44.785980 - 2025-07-24T17:49:44.798561 - comfyui lumi batcher overwrite task done2025-07-24T17:49:44.798968 - 2025-07-24T17:49:44.868694 - Prompt executed in 348.38 seconds 2025-07-24T18:17:32.104770 - [ComfyUI-Manager] The ComfyRegistry cache update is still in progress, so an outdated cache is being used.2025-07-24T18:17:32.105296 - 2025-07-24T18:17:33.147891 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-07-24T18:17:33.147891 - 2025-07-24T18:17:36.036752 - [DONE]2025-07-24T18:17:36.036752 - 2025-07-24T18:17:36.120191 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json2025-07-24T18:17:36.120191 - 2025-07-24T18:17:37.292226 - [DONE]2025-07-24T18:17:37.292226 - 2025-07-24T18:17:37.313785 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extras.json2025-07-24T18:17:37.313785 - 2025-07-24T18:17:38.420206 - [DONE]2025-07-24T18:17:38.420897 - 2025-07-24T18:17:38.602146 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json2025-07-24T18:17:38.602146 - 2025-07-24T18:17:39.726647 - [DONE]2025-07-24T18:17:39.726647 - 2025-07-24T18:21:32.720524 - got prompt 2025-07-24T18:21:36.403235 - timesteps: tensor([1000.0000, 983.7124, 964.5367, 941.6301, 913.7864, 879.2166, 835.1485, 777.0445, 696.9283, 579.3684, 390.1099, 34.7826], device='cuda:0') 2025-07-24T18:21:36.571239 - Seq len: 38192 2025-07-24T18:21:36.578239 - Sampling 85 frames at 896x496 with 12 steps 2025-07-24T18:21:36.834814 - 0%| | 0/12 [00:00<?, ?it/s]2025-07-24T18:21:40.522225 - 0%| | 0/12 [00:03<?, ?it/s]2025-07-24T18:21:40.522225 - 2025-07-24T18:21:40.567079 - !!! Exception during processing !!! 'WanModel' object has no attribute 'vace_patch_embedding' 2025-07-24T18:21:40.572236 - Traceback (most recent call last): File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 361, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 236, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 208, in _map_node_over_list process_inputs(input_dict, i) File "F:\ComfyUI_Mie_V6.0\ComfyUI\execution.py", line 197, in process_inputs results.append(getattr(obj, func)(**inputs)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3325, in process noise_pred, self.cache_state = predict_with_cfg( ^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 2617, in predict_with_cfg noise_pred_cond, cache_state_cond = transformer( ^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1617, in forward vace_hints = self.forward_vace(x, data["context"], data["seq_len"], kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\wanvideo\modules\model.py", line 1196, in forward_vace c = [self.vace_patch_embedding(u.unsqueeze(0).float()).to(x.dtype) for u in vace_context] ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\diffusers\models\modeling_utils.py", line 291, in __getattr__ return super().__getattr__(name) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "F:\ComfyUI_Mie_V6.0\python_embeded\Lib\site-packages\torch\nn\modules\module.py", line 1940, in __getattr__ raise AttributeError( AttributeError: 'WanModel' object has no attribute 'vace_patch_embedding'. Did you mean: 'patch_embedding'? 2025-07-24T18:21:40.574784 - end_vram - start_vram: 34183609272 - 32966353744 = 12172555282025-07-24T18:21:40.574784 - 2025-07-24T18:21:40.575297 - #42 [WanVideoSampler]: 7.68s - vram 1217255528b2025-07-24T18:21:40.575297 - 2025-07-24T18:21:40.575811 - comfyui lumi batcher overwrite task done2025-07-24T18:21:40.575811 - 2025-07-24T18:21:40.577866 - Prompt executed in 7.84 seconds ## Attached Workflow Please make sure that workflow does not contain any sensitive information such as API keys or passwords. Workflow too large. Please manually upload the workflow from local file system. ## Additional Context (Please add any additional context or steps to reproduce the error here) 什么问题,该怎么解决

# ComfyUI Error Report ## Error Details - **Node ID:** 5 - **Node Type:** Joy_caption_two - **Exception Type:** RuntimeError - **Exception Message:** All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] ## Stack Trace File "C:\comfyui2\ComfyUI\execution.py", line 349, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 224, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 196, in _map_node_over_list process_inputs(input_dict, i) File "C:\comfyui2\ComfyUI\execution.py", line 185, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two\joy_caption_two_node.py", line 407, in generate generate_ids = text_model.generate(input_ids, inputs_embeds=input_embeds, attention_mask=attention_mask, File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 2215, in generate result = self._sample( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 3206, in _sample outputs = self(**model_inputs, return_dict=True) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 1190, in forward outputs = self.model( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 945, in forward layer_outputs = decoder_layer( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 676, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 559, in forward query_states = self.q_proj(hidden_states) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\peft\tuners\lora\bnb.py", line 467, in forward result = self.base_layer(x, *args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\nn\modules.py", line 484, in forward return bnb.matmul_4bit(x, self.weight.t(), bias=bias, quant_state=self.weight.quant_state).to(inp_dtype) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 533, in matmul_4bit return MatMul4Bit.apply(A, B, out, bias, quant_state) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 462, in forward output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 1363, in dequantize_4bit is_on_gpu([A, absmax, out]) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 464, in is_on_gpu raise RuntimeError( ## System Information - **ComfyUI Version:** 0.3.33 - **Arguments:** main.py - **OS:** nt - **Python Version:** 3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)] - **Embedded Python:** false - **PyTorch Version:** 2.7.0+cu126 ## Devices - **Name:** cuda:0 NVIDIA GeForce GTX 1660 Ti : cudaMallocAsync - **Type:** cuda - **VRAM Total:** 6442123264 - **VRAM Free:** 5379194880 - **Torch VRAM Total:** 0 - **Torch VRAM Free:** 0 ## Logs 2025-05-12T12:33:00.245025 - Adding extra search path checkpoints C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\Stable-diffusion 2025-05-12T12:33:00.245025 - Adding extra search path configs C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\Stable-diffusion 2025-05-12T12:33:00.246022 - Adding extra search path vae C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\VAE 2025-05-12T12:33:00.246022 - Adding extra search path loras C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\Lora 2025-05-12T12:33:00.246022 - Adding extra search path loras C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\LyCORIS 2025-05-12T12:33:00.247019 - Adding extra search path upscale_models C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\ESRGAN 2025-05-12T12:33:00.247019 - Adding extra search path upscale_models C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\RealESRGAN 2025-05-12T12:33:00.247019 - Adding extra search path upscale_models C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\SwinIR 2025-05-12T12:33:00.247019 - Adding extra search path embeddings C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\embeddings 2025-05-12T12:33:00.247019 - Adding extra search path hypernetworks C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\models\hypernetworks 2025-05-12T12:33:00.248017 - Adding extra search path controlnet C:\baidunetdiskdownload\sd-webui-aki\sd-webui-aki-v4.10\extensions\sd-webui-controlnet 2025-05-12T12:33:00.823554 - [START] Security scan2025-05-12T12:33:00.823554 - 2025-05-12T12:33:02.367066 - [DONE] Security scan2025-05-12T12:33:02.367066 - 2025-05-12T12:33:02.643316 - ## ComfyUI-Manager: installing dependencies done.2025-05-12T12:33:02.643316 - 2025-05-12T12:33:02.643316 - ** ComfyUI startup time:2025-05-12T12:33:02.644221 - 2025-05-12T12:33:02.644221 - 2025-05-12 12:33:02.6432025-05-12T12:33:02.644221 - 2025-05-12T12:33:02.644221 - ** Platform:2025-05-12T12:33:02.644221 - 2025-05-12T12:33:02.644221 - Windows2025-05-12T12:33:02.645181 - 2025-05-12T12:33:02.645181 - ** Python version:2025-05-12T12:33:02.645181 - 2025-05-12T12:33:02.645181 - 3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)]2025-05-12T12:33:02.645181 - 2025-05-12T12:33:02.645181 - ** Python executable:2025-05-12T12:33:02.645181 - 2025-05-12T12:33:02.646180 - C:\Users\张三\AppData\Local\Programs\Python\Python310\python.exe2025-05-12T12:33:02.646180 - 2025-05-12T12:33:02.646180 - ** ComfyUI Path:2025-05-12T12:33:02.646180 - 2025-05-12T12:33:02.646180 - C:\comfyui2\ComfyUI2025-05-12T12:33:02.646180 - 2025-05-12T12:33:02.646180 - ** ComfyUI Base Folder Path:2025-05-12T12:33:02.647178 - 2025-05-12T12:33:02.647178 - C:\comfyui2\ComfyUI2025-05-12T12:33:02.647178 - 2025-05-12T12:33:02.647178 - ** User directory:2025-05-12T12:33:02.647178 - 2025-05-12T12:33:02.648176 - C:\comfyui2\ComfyUI\user2025-05-12T12:33:02.648176 - 2025-05-12T12:33:02.648176 - ** ComfyUI-Manager config path:2025-05-12T12:33:02.648176 - 2025-05-12T12:33:02.648176 - C:\comfyui2\ComfyUI\user\default\ComfyUI-Manager\config.ini2025-05-12T12:33:02.648176 - 2025-05-12T12:33:02.648176 - ** Log path:2025-05-12T12:33:02.648176 - 2025-05-12T12:33:02.648176 - C:\comfyui2\ComfyUI\user\comfyui.log2025-05-12T12:33:02.648176 - 2025-05-12T12:33:04.301068 - Prestartup times for custom nodes: 2025-05-12T12:33:04.302066 - 0.0 seconds: C:\comfyui2\ComfyUI\custom_nodes\comfyui-easy-use 2025-05-12T12:33:04.302066 - 4.1 seconds: C:\comfyui2\ComfyUI\custom_nodes\ComfyUI-Manager 2025-05-12T12:33:04.302066 - 2025-05-12T12:33:06.807275 - Checkpoint files will always be loaded safely. 2025-05-12T12:33:07.017820 - Total VRAM 6144 MB, total RAM 16294 MB 2025-05-12T12:33:07.018365 - pytorch version: 2.7.0+cu126 2025-05-12T12:33:09.352975 - xformers version: 0.0.30 2025-05-12T12:33:09.352975 - Set vram state to: NORMAL_VRAM 2025-05-12T12:33:09.352975 - Device: cuda:0 NVIDIA GeForce GTX 1660 Ti : cudaMallocAsync 2025-05-12T12:33:09.686612 - Using xformers attention 2025-05-12T12:33:11.295329 - Python version: 3.10.11 (tags/v3.10.11:7d4cc5a, Apr 5 2023, 00:38:17) [MSC v.1929 64 bit (AMD64)] 2025-05-12T12:33:11.295743 - ComfyUI version: 0.3.33 2025-05-12T12:33:11.358578 - ComfyUI frontend version: 1.18.10 2025-05-12T12:33:11.361570 - [Prompt Server] web root: C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\comfyui_frontend_package\static 2025-05-12T12:33:13.337748 - Module 'diffusers' load failed. If you don't have it installed, do it:2025-05-12T12:33:13.338688 - 2025-05-12T12:33:13.339179 - pip install diffusers2025-05-12T12:33:13.339179 - 2025-05-12T12:33:13.688119 - [ComfyUI-Easy-Use] server: v1.3.0 Loaded2025-05-12T12:33:13.688119 - 2025-05-12T12:33:13.688119 - [ComfyUI-Easy-Use] web root: C:\comfyui2\ComfyUI\custom_nodes\comfyui-easy-use\web_version/v2 Loaded2025-05-12T12:33:13.689074 - 2025-05-12T12:33:13.717123 - ### Loading: ComfyUI-Manager (V3.31.13) 2025-05-12T12:33:13.717123 - [ComfyUI-Manager] network_mode: public 2025-05-12T12:33:14.072667 - ### ComfyUI Version: v0.3.33-10-g577de83c | Released on '2025-05-11' 2025-05-12T12:33:14.311459 - Import times for custom nodes: 2025-05-12T12:33:14.311459 - 0.0 seconds: C:\comfyui2\ComfyUI\custom_nodes\websocket_image_save.py 2025-05-12T12:33:14.313453 - 0.0 seconds: C:\comfyui2\ComfyUI\custom_nodes\aigodlike-comfyui-translation 2025-05-12T12:33:14.313453 - 0.0 seconds: C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two 2025-05-12T12:33:14.314451 - 0.0 seconds: C:\comfyui2\ComfyUI\custom_nodes\comfyui-custom-scripts 2025-05-12T12:33:14.314451 - 0.6 seconds: C:\comfyui2\ComfyUI\custom_nodes\ComfyUI-Manager 2025-05-12T12:33:14.315449 - 0.7 seconds: C:\comfyui2\ComfyUI\custom_nodes\comfyui-easy-use 2025-05-12T12:33:14.315449 - 2025-05-12T12:33:14.362371 - Starting server 2025-05-12T12:33:14.363371 - To see the GUI go to: http://127.0.0.1:8188 2025-05-12T12:33:19.620936 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/model-list.json 2025-05-12T12:33:19.678088 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/extension-node-map.json 2025-05-12T12:33:19.845706 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/alter-list.json 2025-05-12T12:33:19.955392 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/github-stats.json 2025-05-12T12:33:20.436522 - [ComfyUI-Manager] default cache updated: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json 2025-05-12T12:34:14.934119 - Cannot connect to comfyregistry.2025-05-12T12:34:14.934119 - 2025-05-12T12:34:14.940242 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-05-12T12:34:14.940242 - 2025-05-12T12:34:19.785536 - got prompt 2025-05-12T12:34:19.982016 - C:\comfyui2\ComfyUI\models\clip\siglip-so400m-patch14-3842025-05-12T12:34:19.982016 - 2025-05-12T12:34:20.277243 - [DONE]2025-05-12T12:34:20.277243 - 2025-05-12T12:34:20.377933 - [ComfyUI-Manager] All startup tasks have been completed. 2025-05-12T12:34:26.095833 - Loading VLM's custom vision model2025-05-12T12:34:26.096832 - 2025-05-12T12:34:28.260562 - Requested to load SiglipVisionTransformer 2025-05-12T12:34:28.272048 - loaded completely 9.5367431640625e+25 809.1729736328125 True 2025-05-12T12:34:53.793004 - Requested to load ImageAdapter 2025-05-12T12:34:53.793004 - loaded completely 9.5367431640625e+25 41.0390625 True 2025-05-12T12:35:10.488040 - Loading tokenizer2025-05-12T12:35:10.488040 - 2025-05-12T12:35:11.002348 - Loading LLM: unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit2025-05-12T12:35:11.003346 - 2025-05-12T12:35:11.003346 - C:\comfyui2\ComfyUI\models\LLM\Meta-Llama-3.1-8B-Instruct-bnb-4bit2025-05-12T12:35:11.003346 - 2025-05-12T12:35:11.005339 - Successfully modified 'base_model_name_or_path' value in 'C:\comfyui2\ComfyUI\models\Joy_caption_two\text_model\adapter_config.json'.2025-05-12T12:35:11.005339 - 2025-05-12T12:35:24.237554 - Unused kwargs: ['_load_in_4bit', '_load_in_8bit', 'quant_method']. These kwargs are not used in <class 'transformers.utils.quantization_config.BitsAndBytesConfig'>. 2025-05-12T12:35:30.940677 - !!! Exception during processing !!! All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T12:35:30.962005 - Traceback (most recent call last): File "C:\comfyui2\ComfyUI\execution.py", line 349, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 224, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 196, in _map_node_over_list process_inputs(input_dict, i) File "C:\comfyui2\ComfyUI\execution.py", line 185, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two\joy_caption_two_node.py", line 407, in generate generate_ids = text_model.generate(input_ids, inputs_embeds=input_embeds, attention_mask=attention_mask, File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 2215, in generate result = self._sample( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 3206, in _sample outputs = self(**model_inputs, return_dict=True) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 1190, in forward outputs = self.model( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 945, in forward layer_outputs = decoder_layer( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 676, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 559, in forward query_states = self.q_proj(hidden_states) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\peft\tuners\lora\bnb.py", line 467, in forward result = self.base_layer(x, *args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\nn\modules.py", line 484, in forward return bnb.matmul_4bit(x, self.weight.t(), bias=bias, quant_state=self.weight.quant_state).to(inp_dtype) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 533, in matmul_4bit return MatMul4Bit.apply(A, B, out, bias, quant_state) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 462, in forward output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 1363, in dequantize_4bit is_on_gpu([A, absmax, out]) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 464, in is_on_gpu raise RuntimeError( RuntimeError: All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T12:35:30.967286 - Prompt executed in 71.18 seconds 2025-05-12T12:51:12.715731 - got prompt 2025-05-12T12:51:12.907212 - loaded completely 9.5367431640625e+25 809.1729736328125 True 2025-05-12T12:51:36.518268 - loaded completely 9.5367431640625e+25 41.0390625 True 2025-05-12T12:51:53.412534 - !!! Exception during processing !!! All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T12:51:53.419550 - Traceback (most recent call last): File "C:\comfyui2\ComfyUI\execution.py", line 349, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 224, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 196, in _map_node_over_list process_inputs(input_dict, i) File "C:\comfyui2\ComfyUI\execution.py", line 185, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two\joy_caption_two_node.py", line 407, in generate generate_ids = text_model.generate(input_ids, inputs_embeds=input_embeds, attention_mask=attention_mask, File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 2215, in generate result = self._sample( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 3206, in _sample outputs = self(**model_inputs, return_dict=True) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 1190, in forward outputs = self.model( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 945, in forward layer_outputs = decoder_layer( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 676, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 559, in forward query_states = self.q_proj(hidden_states) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\peft\tuners\lora\bnb.py", line 467, in forward result = self.base_layer(x, *args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\nn\modules.py", line 484, in forward return bnb.matmul_4bit(x, self.weight.t(), bias=bias, quant_state=self.weight.quant_state).to(inp_dtype) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 533, in matmul_4bit return MatMul4Bit.apply(A, B, out, bias, quant_state) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 462, in forward output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 1363, in dequantize_4bit is_on_gpu([A, absmax, out]) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 464, in is_on_gpu raise RuntimeError( RuntimeError: All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T12:51:53.426451 - Prompt executed in 40.69 seconds 2025-05-12T13:27:37.964284 - got prompt 2025-05-12T13:27:38.213613 - loaded completely 9.5367431640625e+25 809.1729736328125 True 2025-05-12T13:28:03.041961 - loaded completely 9.5367431640625e+25 41.0390625 True 2025-05-12T13:28:19.913094 - !!! Exception during processing !!! All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T13:28:19.919269 - Traceback (most recent call last): File "C:\comfyui2\ComfyUI\execution.py", line 349, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 224, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 196, in _map_node_over_list process_inputs(input_dict, i) File "C:\comfyui2\ComfyUI\execution.py", line 185, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two\joy_caption_two_node.py", line 407, in generate generate_ids = text_model.generate(input_ids, inputs_embeds=input_embeds, attention_mask=attention_mask, File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 2215, in generate result = self._sample( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 3206, in _sample outputs = self(**model_inputs, return_dict=True) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 1190, in forward outputs = self.model( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 945, in forward layer_outputs = decoder_layer( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 676, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 559, in forward query_states = self.q_proj(hidden_states) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\peft\tuners\lora\bnb.py", line 467, in forward result = self.base_layer(x, *args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\nn\modules.py", line 484, in forward return bnb.matmul_4bit(x, self.weight.t(), bias=bias, quant_state=self.weight.quant_state).to(inp_dtype) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 533, in matmul_4bit return MatMul4Bit.apply(A, B, out, bias, quant_state) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 462, in forward output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 1363, in dequantize_4bit is_on_gpu([A, absmax, out]) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 464, in is_on_gpu raise RuntimeError( RuntimeError: All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T13:28:19.925246 - Prompt executed in 41.93 seconds 2025-05-12T14:01:51.529873 - got prompt 2025-05-12T14:01:51.690142 - loaded completely 9.5367431640625e+25 809.1729736328125 True 2025-05-12T14:01:56.251833 - [ComfyUI-Manager] The ComfyRegistry cache update is still in progress, so an outdated cache is being used.2025-05-12T14:01:56.251833 - 2025-05-12T14:01:56.355173 - FETCH DATA from: C:\comfyui2\ComfyUI\user\default\ComfyUI-Manager\cache\1514988643_custom-node-list.json2025-05-12T14:01:56.355173 - 2025-05-12T14:01:56.372727 - [DONE]2025-05-12T14:01:56.372727 - 2025-05-12T14:01:56.465008 - FETCH DATA from: C:\comfyui2\ComfyUI\user\default\ComfyUI-Manager\cache\746607195_github-stats.json2025-05-12T14:01:56.465008 - 2025-05-12T14:01:56.472987 - [DONE]2025-05-12T14:01:56.473984 - 2025-05-12T14:01:56.476976 - FETCH DATA from: C:\comfyui2\ComfyUI\user\default\ComfyUI-Manager\cache\832903789_extras.json2025-05-12T14:01:56.476976 - 2025-05-12T14:01:56.477973 - [DONE]2025-05-12T14:01:56.477973 - 2025-05-12T14:01:56.646551 - FETCH DATA from: C:\comfyui2\ComfyUI\user\default\ComfyUI-Manager\cache\1742899825_extension-node-map.json2025-05-12T14:01:56.646551 - 2025-05-12T14:01:56.656524 - [DONE]2025-05-12T14:01:56.656524 - 2025-05-12T14:02:15.505597 - loaded completely 9.5367431640625e+25 41.0390625 True 2025-05-12T14:02:31.244068 - !!! Exception during processing !!! All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T14:02:31.248613 - Traceback (most recent call last): File "C:\comfyui2\ComfyUI\execution.py", line 349, in execute output_data, output_ui, has_subgraph = get_output_data(obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 224, in get_output_data return_values = _map_node_over_list(obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb) File "C:\comfyui2\ComfyUI\execution.py", line 196, in _map_node_over_list process_inputs(input_dict, i) File "C:\comfyui2\ComfyUI\execution.py", line 185, in process_inputs results.append(getattr(obj, func)(**inputs)) File "C:\comfyui2\ComfyUI\custom_nodes\ComfyUI_SLK_joy_caption_two\joy_caption_two_node.py", line 407, in generate generate_ids = text_model.generate(input_ids, inputs_embeds=input_embeds, attention_mask=attention_mask, File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\utils\_contextlib.py", line 116, in decorate_context return func(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 2215, in generate result = self._sample( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\generation\utils.py", line 3206, in _sample outputs = self(**model_inputs, return_dict=True) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 1190, in forward outputs = self.model( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 945, in forward layer_outputs = decoder_layer( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 676, in forward hidden_states, self_attn_weights, present_key_value = self.self_attn( File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\transformers\models\llama\modeling_llama.py", line 559, in forward query_states = self.q_proj(hidden_states) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\peft\tuners\lora\bnb.py", line 467, in forward result = self.base_layer(x, *args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1751, in _wrapped_call_impl return self._call_impl(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\nn\modules\module.py", line 1762, in _call_impl return forward_call(*args, **kwargs) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\nn\modules.py", line 484, in forward return bnb.matmul_4bit(x, self.weight.t(), bias=bias, quant_state=self.weight.quant_state).to(inp_dtype) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 533, in matmul_4bit return MatMul4Bit.apply(A, B, out, bias, quant_state) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\torch\autograd\function.py", line 575, in apply return super().apply(*args, **kwargs) # type: ignore[misc] File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\autograd\_functions.py", line 462, in forward output = torch.nn.functional.linear(A, F.dequantize_4bit(B, quant_state).to(A.dtype).t(), bias) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 1363, in dequantize_4bit is_on_gpu([A, absmax, out]) File "C:\Users\张三\AppData\Local\Programs\Python\Python310\lib\site-packages\bitsandbytes\functional.py", line 464, in is_on_gpu raise RuntimeError( RuntimeError: All input tensors need to be on the same GPU, but found some tensors to not be on a GPU: [(torch.Size([1, 8388608]), device(type='cpu')), (torch.Size([262144]), device(type='cpu')), (torch.Size([4096, 4096]), device(type='cpu'))] 2025-05-12T14:02:31.253592 - Prompt executed in 39.72 seconds ## Attached Workflow Please make sure that workflow does not contain any sensitive information such as API keys or passwords. {"id":"00000000-0000-0000-0000-000000000000","revision":0,"last_node_id":6,"last_link_id":6,"nodes":[{"id":2,"type":"LoadImage","pos":[-536.7379150390625,-152.25942993164062],"size":[270,314],"flags":{},"order":0,"mode":0,"inputs":[{"label":"image","localized_name":"图像","name":"image","type":"COMBO","widget":{"name":"image"},"link":null},{"label":"upload","localized_name":"选择文件上传","name":"upload","type":"IMAGEUPLOAD","widget":{"name":"upload"},"link":null}],"outputs":[{"label":"IMAGE","localized_name":"图像","name":"IMAGE","type":"IMAGE","links":[5]},{"label":"MASK","localized_name":"遮罩","name":"MASK","type":"MASK","links":null}],"properties":{"cnr_id":"comfy-core","ver":"0.3.33","Node name for S&R":"LoadImage"},"widgets_values":["CN_6.jpg","image"]},{"id":4,"type":"ShowText|pysssss","pos":[247.84080505371094,74.16905975341797],"size":[366.1156921386719,159.55372619628906],"flags":{},"order":3,"mode":0,"inputs":[{"label":"text","localized_name":"text","name":"text","type":"STRING","link":4}],"outputs":[{"label":"STRING","localized_name":"字符串","name":"STRING","shape":6,"type":"STRING","links":null}],"properties":{"cnr_id":"comfyui-custom-scripts","ver":"1.2.5","Node name for S&R":"ShowText|pysssss"},"widgets_values":[]},{"id":6,"type":"Joy_caption_two_load","pos":[-148.00003051757812,-86.00003051757812],"size":[270,58],"flags":{},"order":1,"mode":0,"inputs":[{"label":"model","localized_name":"model","name":"model","type":"COMBO","widget":{"name":"model"},"link":null}],"outputs":[{"label":"JoyTwoPipeline","localized_name":"JoyTwoPipeline","name":"JoyTwoPipeline","type":"JoyTwoPipeline","links":[6]}],"properties":{"cnr_id":"comfyui_slk_joy_caption_two","ver":"667751cab945bd8e9fb0be4d557d47e36821350a","Node name for S&R":"Joy_caption_two_load"},"widgets_values":["unsloth/Meta-Llama-3.1-8B-Instruct-bnb-4bit"]},{"id":5,"type":"Joy_caption_two","pos":[-139.20001220703125,33.199951171875],"size":[270,126],"flags":{},"order":2,"mode":0,"inputs":[{"label":"joy_two_pipeline","localized_name":"joy_two_pipeline","name":"joy_two_pipeline","type":"JoyTwoPipeline","link":6},{"label":"image","localized_name":"image","name":"image","type":"IMAGE","link":5},{"label":"caption_type","localized_name":"caption_type","name":"caption_type","type":"COMBO","widget":{"name":"caption_type"},"link":null},{"label":"caption_length","localized_name":"caption_length","name":"caption_length","type":"COMBO","widget":{"name":"caption_length"},"link":null},{"label":"low_vram","localized_name":"low_vram","name":"low_vram","type":"BOOLEAN","widget":{"name":"low_vram"},"link":null}],"outputs":[{"label":"STRING","localized_name":"字符串","name":"STRING","type":"STRING","links":[4]}],"properties":{"cnr_id":"comfyui_slk_joy_caption_two","ver":"667751cab945bd8e9fb0be4d557d47e36821350a","Node name for S&R":"Joy_caption_two"},"widgets_values":["Descriptive","long",true]}],"links":[[4,5,0,4,0,"STRING"],[5,2,0,5,1,"IMAGE"],[6,6,0,5,0,"JoyTwoPipeline"]],"groups":[],"config":{},"extra":{"frontendVersion":"1.18.10"},"version":0.4} ## Additional Context (Please add any additional context or steps to reproduce the error here) 以上代码是什么意思

# ComfyUI Error Report ## Error Details - **Node ID:** 90 - **Node Type:** WanVideoSampler - **Exception Type:** UnboundLocalError - **Exception Message:** cannot access local variable 'callback_latent' where it is not associated with a value ## Stack Trace File "G:\comfyui\ComfyUI\execution.py", line 496, in execute output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 315, in get_output_data return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 289, in _async_map_node_over_list await process_inputs(input_dict, i) File "G:\comfyui\ComfyUI\execution.py", line 277, in process_inputs result = f(**inputs) ^^^^^^^^^^^ File "G:\comfyui\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3988, in process "samples": callback_latent.unsqueeze(0).cpu() if callback is not None else None, ^^^^^^^^^^^^^^^ ## System Information - **ComfyUI Version:** 0.3.56 - **Arguments:** main.py --auto-launch - **OS:** nt - **Python Version:** 3.12.10 (tags/v3.12.10:0cc8128, Apr 8 2025, 12:21:36) [MSC v.1943 64 bit (AMD64)] - **Embedded Python:** false - **PyTorch Version:** 2.7.1+cu128 ## Devices - **Name:** cuda:0 NVIDIA GeForce RTX 4070 Ti SUPER : cudaMallocAsync - **Type:** cuda - **VRAM Total:** 17170956288 - **VRAM Free:** 15821963264 - **Torch VRAM Total:** 0 - **Torch VRAM Free:** 0 ## Logs 2025-09-01T14:58:56.860057 - 0.9 seconds: G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes 2025-09-01T14:58:56.860057 - 1.1 seconds: G:\comfyui\ComfyUI\custom_nodes\ComfyUI_Custom_Nodes_AlekPet 2025-09-01T14:58:56.861054 - 1.7 seconds: G:\comfyui\ComfyUI\custom_nodes\ComfyUI-nunchaku 2025-09-01T14:58:56.861054 - 2.4 seconds: G:\comfyui\ComfyUI\custom_nodes\ComfyUI-Easy-Use 2025-09-01T14:58:56.861054 - 2025-09-01T14:58:57.088527 - Context impl SQLiteImpl. 2025-09-01T14:58:57.088527 - Will assume non-transactional DDL. 2025-09-01T14:58:57.089524 - No target revision found. 2025-09-01T14:58:57.111451 - Starting server 2025-09-01T14:58:57.112447 - To see the GUI go to: http://127.0.0.1:8188 2025-09-01T14:58:58.110145 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/photoswipe-lightbox.esm.min.js2025-09-01T14:58:58.110145 - 2025-09-01T14:58:58.154639 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/photoswipe.min.css2025-09-01T14:58:58.154639 - 2025-09-01T14:58:58.155636 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/pickr.min.js2025-09-01T14:58:58.155636 - 2025-09-01T14:58:58.192513 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/classic.min.css2025-09-01T14:58:58.192513 - 2025-09-01T14:58:58.193509 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/model-viewer.min.js2025-09-01T14:58:58.193509 - 2025-09-01T14:58:58.197497 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/juxtapose.css2025-09-01T14:58:58.198493 - 2025-09-01T14:58:58.209456 - G:\comfyui\ComfyUI\custom_nodes\comfyui-mixlab-nodes\webApp\lib/juxtapose.min.js2025-09-01T14:58:58.209456 - 2025-09-01T14:58:58.799579 - [G:\comfyui\ComfyUI\custom_nodes\comfy_mtb] | INFO -> Found multiple match, we will pick the last D:\AI-sd\sd-webui-aki-v4\models\SwinIR ['G:\\comfyui\\ComfyUI\\models\\upscale_models', 'D:\\AI-sd\\sd-webui-aki-v4\\models\\ESRGAN', 'D:\\AI-sd\\sd-webui-aki-v4\\models\\RealESRGAN', 'D:\\AI-sd\\sd-webui-aki-v4\\models\\SwinIR'] 2025-09-01T14:58:58.807556 - []2025-09-01T14:58:58.807556 - 2025-09-01T14:58:58.807556 - []2025-09-01T14:58:58.807556 - 2025-09-01T14:59:00.532965 - got prompt 2025-09-01T14:59:00.569842 - Using xformers attention in VAE 2025-09-01T14:59:00.570844 - Using xformers attention in VAE 2025-09-01T14:59:00.726582 - VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16 2025-09-01T14:59:00.882662 - Requested to load FluxClipModel_ 2025-09-01T14:59:00.889639 - loaded completely 9.5367431640625e+25 4777.53759765625 True 2025-09-01T14:59:00.998583 - CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cuda:0, dtype: torch.float16 2025-09-01T14:59:01.063367 - clip missing: ['text_projection.weight'] 2025-09-01T14:59:01.722971 - FETCH ComfyRegistry Data: 5/962025-09-01T14:59:01.722971 - 2025-09-01T14:59:03.042519 - model weight dtype torch.float8_e4m3fn, manual cast: torch.bfloat16 2025-09-01T14:59:03.043516 - model_type FLUX 2025-09-01T14:59:11.471386 - FETCH ComfyRegistry Data: 10/962025-09-01T14:59:11.473376 - 2025-09-01T14:59:11.998346 - Requested to load Flux 2025-09-01T14:59:18.503848 - loaded completely 0.0 11350.067443847656 True 2025-09-01T14:59:18.514805 - Patching comfy attention to use sageattn2025-09-01T14:59:18.514805 - 2025-09-01T14:59:18.688225 - 0%| | 0/20 [00:00<?, ?it/s]2025-09-01T14:59:20.150041 - FETCH ComfyRegistry Data: 15/962025-09-01T14:59:20.150041 - 2025-09-01T14:59:28.628200 - 40%|█████████████████████████████████▏ | 8/20 [00:09<00:14, 1.17s/it]2025-09-01T14:59:29.656666 - FETCH ComfyRegistry Data: 20/962025-09-01T14:59:29.657170 - 2025-09-01T14:59:37.969589 - 80%|█████████████████████████████████████████████████████████████████▌ | 16/20 [00:19<00:04, 1.16s/it]2025-09-01T14:59:38.276116 - FETCH ComfyRegistry Data: 25/962025-09-01T14:59:38.276619 - 2025-09-01T14:59:42.606388 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:23<00:00, 1.16s/it]2025-09-01T14:59:42.609378 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:23<00:00, 1.20s/it]2025-09-01T14:59:42.610375 - 2025-09-01T14:59:42.612369 - Restoring initial comfy attention2025-09-01T14:59:42.613364 - 2025-09-01T14:59:43.566091 - Requested to load AutoencodingEngine 2025-09-01T14:59:44.164328 - loaded completely 0.0 159.87335777282715 True 2025-09-01T14:59:44.683600 - comfyui lumi batcher overwrite task done2025-09-01T14:59:44.684596 - 2025-09-01T14:59:44.695560 - Prompt executed in 44.16 seconds 2025-09-01T14:59:47.355547 - FETCH ComfyRegistry Data: 30/962025-09-01T14:59:47.355547 - 2025-09-01T14:59:50.013416 - got prompt 2025-09-01T14:59:50.431031 - loaded completely 0.0 11350.067443847656 True 2025-09-01T14:59:50.435018 - Patching comfy attention to use sageattn2025-09-01T14:59:50.435018 - 2025-09-01T14:59:55.538670 - 25%|████████████████████▊ | 5/20 [00:05<00:16, 1.11s/it]2025-09-01T14:59:56.060186 - FETCH ComfyRegistry Data: 35/962025-09-01T14:59:56.060689 - 2025-09-01T15:00:03.927199 - 60%|█████████████████████████████████████████████████▏ | 12/20 [00:13<00:09, 1.19s/it]2025-09-01T15:00:04.665127 - FETCH ComfyRegistry Data: 40/962025-09-01T15:00:04.666123 - 2025-09-01T15:00:13.251827 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.14s/it]2025-09-01T15:00:13.252823 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.14s/it]2025-09-01T15:00:13.252823 - 2025-09-01T15:00:13.253821 - Restoring initial comfy attention2025-09-01T15:00:13.253821 - 2025-09-01T15:00:13.397095 - FETCH ComfyRegistry Data: 45/962025-09-01T15:00:13.397604 - 2025-09-01T15:00:14.837381 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:00:15.176809 - comfyui lumi batcher overwrite task done2025-09-01T15:00:15.177805 - 2025-09-01T15:00:15.179799 - Prompt executed in 25.16 seconds 2025-09-01T15:00:22.087367 - FETCH ComfyRegistry Data: 50/962025-09-01T15:00:22.088368 - 2025-09-01T15:00:26.342724 - got prompt 2025-09-01T15:00:27.137497 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:00:27.148464 - Patching comfy attention to use sageattn2025-09-01T15:00:27.148464 - 2025-09-01T15:00:29.963897 - 15%|████████████▍ | 3/20 [00:02<00:17, 1.03s/it]2025-09-01T15:00:30.838878 - FETCH ComfyRegistry Data: 55/962025-09-01T15:00:30.839380 - 2025-09-01T15:00:39.316242 - 55%|█████████████████████████████████████████████ | 11/20 [00:12<00:10, 1.17s/it]2025-09-01T15:00:39.731658 - FETCH ComfyRegistry Data: 60/962025-09-01T15:00:39.737147 - 2025-09-01T15:00:47.708181 - 90%|█████████████████████████████████████████████████████████████████████████▊ | 18/20 [00:20<00:02, 1.20s/it]2025-09-01T15:00:48.592029 - FETCH ComfyRegistry Data: 65/962025-09-01T15:00:48.592029 - 2025-09-01T15:00:50.028377 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.18s/it]2025-09-01T15:00:50.029376 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.14s/it]2025-09-01T15:00:50.029376 - 2025-09-01T15:00:50.030370 - Restoring initial comfy attention2025-09-01T15:00:50.030370 - 2025-09-01T15:00:51.666354 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:00:52.027193 - comfyui lumi batcher overwrite task done2025-09-01T15:00:52.027193 - 2025-09-01T15:00:52.033721 - Prompt executed in 25.69 seconds 2025-09-01T15:00:57.732591 - FETCH ComfyRegistry Data: 70/962025-09-01T15:00:57.732591 - 2025-09-01T15:01:02.463536 - got prompt 2025-09-01T15:01:03.089034 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:01:03.098004 - Patching comfy attention to use sageattn2025-09-01T15:01:03.098004 - 2025-09-01T15:01:05.773837 - 15%|████████████▍ | 3/20 [00:02<00:17, 1.01s/it]2025-09-01T15:01:06.618083 - FETCH ComfyRegistry Data: 75/962025-09-01T15:01:06.618083 - 2025-09-01T15:01:15.225214 - 55%|█████████████████████████████████████████████ | 11/20 [00:12<00:10, 1.17s/it]2025-09-01T15:01:15.559178 - FETCH ComfyRegistry Data: 80/962025-09-01T15:01:15.559178 - 2025-09-01T15:01:23.462415 - 90%|█████████████████████████████████████████████████████████████████████████▊ | 18/20 [00:20<00:02, 1.18s/it]2025-09-01T15:01:24.532134 - FETCH ComfyRegistry Data: 85/962025-09-01T15:01:24.532134 - 2025-09-01T15:01:25.823332 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.18s/it]2025-09-01T15:01:25.824332 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.14s/it]2025-09-01T15:01:25.824332 - 2025-09-01T15:01:25.824332 - Restoring initial comfy attention2025-09-01T15:01:25.825360 - 2025-09-01T15:01:27.465611 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:01:27.822201 - comfyui lumi batcher overwrite task done2025-09-01T15:01:27.822201 - 2025-09-01T15:01:27.825191 - Prompt executed in 25.36 seconds 2025-09-01T15:01:33.620743 - FETCH ComfyRegistry Data: 90/962025-09-01T15:01:33.620743 - 2025-09-01T15:01:43.439162 - FETCH ComfyRegistry Data: 95/962025-09-01T15:01:43.439162 - 2025-09-01T15:01:45.474384 - FETCH ComfyRegistry Data [DONE]2025-09-01T15:01:45.474384 - 2025-09-01T15:01:45.625519 - [ComfyUI-Manager] default cache updated: https://api.comfy.org/nodes 2025-09-01T15:01:45.715218 - FETCH DATA from: https://raw.githubusercontent.com/ltdrdata/ComfyUI-Manager/main/custom-node-list.json2025-09-01T15:01:45.715218 - 2025-09-01T15:01:47.345156 - got prompt 2025-09-01T15:01:47.361993 - Requested to load Flux 2025-09-01T15:01:47.516137 - [DONE]2025-09-01T15:01:47.516137 - 2025-09-01T15:01:47.591887 - [ComfyUI-Manager] All startup tasks have been completed. 2025-09-01T15:01:48.028249 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:02:12.473821 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:24<00:00, 1.26s/it]2025-09-01T15:02:12.474817 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:24<00:00, 1.22s/it]2025-09-01T15:02:12.474817 - 2025-09-01T15:02:14.322793 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:02:14.683971 - comfyui lumi batcher overwrite task done2025-09-01T15:02:14.683971 - 2025-09-01T15:02:14.685964 - Prompt executed in 27.34 seconds 2025-09-01T15:02:19.663561 - got prompt 2025-09-01T15:02:20.178226 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:02:44.645309 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:24<00:00, 1.25s/it]2025-09-01T15:02:44.646305 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:24<00:00, 1.22s/it]2025-09-01T15:02:44.646305 - 2025-09-01T15:02:46.461739 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:02:46.818816 - comfyui lumi batcher overwrite task done2025-09-01T15:02:46.818816 - 2025-09-01T15:02:46.820813 - Prompt executed in 27.16 seconds 2025-09-01T15:02:53.822858 - got prompt 2025-09-01T15:02:53.845781 - Requested to load Flux 2025-09-01T15:02:54.625791 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:02:54.635757 - Patching comfy attention to use sageattn2025-09-01T15:02:54.635757 - 2025-09-01T15:03:17.948689 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:23<00:00, 1.19s/it]2025-09-01T15:03:17.950687 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:23<00:00, 1.17s/it]2025-09-01T15:03:17.951683 - 2025-09-01T15:03:17.953675 - Restoring initial comfy attention2025-09-01T15:03:17.953675 - 2025-09-01T15:03:19.664335 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:03:20.021412 - comfyui lumi batcher overwrite task done2025-09-01T15:03:20.021412 - 2025-09-01T15:03:20.027392 - Prompt executed in 26.20 seconds 2025-09-01T15:03:29.398450 - got prompt 2025-09-01T15:03:30.062602 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:03:30.071572 - Patching comfy attention to use sageattn2025-09-01T15:03:30.071572 - 2025-09-01T15:03:52.914438 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.17s/it]2025-09-01T15:03:52.915434 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.14s/it]2025-09-01T15:03:52.915434 - 2025-09-01T15:03:52.916431 - Restoring initial comfy attention2025-09-01T15:03:52.916431 - 2025-09-01T15:03:54.593879 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:03:54.936053 - comfyui lumi batcher overwrite task done2025-09-01T15:03:54.937051 - 2025-09-01T15:03:54.938046 - Prompt executed in 25.54 seconds 2025-09-01T15:05:17.191259 - got prompt 2025-09-01T15:05:17.823083 - loaded completely 0.0 11350.067443847656 True 2025-09-01T15:05:17.832052 - Patching comfy attention to use sageattn2025-09-01T15:05:17.833050 - 2025-09-01T15:05:40.406129 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.16s/it]2025-09-01T15:05:40.406129 - 100%|██████████████████████████████████████████████████████████████████████████████████| 20/20 [00:22<00:00, 1.13s/it]2025-09-01T15:05:40.406999 - 2025-09-01T15:05:40.407512 - Restoring initial comfy attention2025-09-01T15:05:40.407512 - 2025-09-01T15:05:42.007023 - loaded completely 0.0 159.87335777282715 True 2025-09-01T15:05:42.368928 - comfyui lumi batcher overwrite task done2025-09-01T15:05:42.368928 - 2025-09-01T15:05:42.370921 - Prompt executed in 25.18 seconds 2025-09-01T15:18:47.644508 - got prompt 2025-09-01T15:18:59.759942 - T5Encoder: 29%|███████████████████████████████████████████████████████▏ | 7/24 [00:00<00:00, 69.54it/s]2025-09-01T15:18:59.794825 - T5Encoder: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 24/24 [00:00<00:00, 177.06it/s]2025-09-01T15:18:59.795821 - 2025-09-01T15:19:00.091094 - T5Encoder: 0%| | 0/24 [00:00<?, ?it/s]2025-09-01T15:19:00.139935 - T5Encoder: 100%|███████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 24/24 [00:00<00:00, 501.63it/s]2025-09-01T15:19:00.140927 - 2025-09-01T15:19:16.616558 - CUDA Compute Capability: 8.9 2025-09-01T15:19:16.616558 - Detected model in_channels: 16 2025-09-01T15:19:16.617551 - Model cross attention type: t2v, num_heads: 40, num_layers: 40 2025-09-01T15:19:16.618547 - Model variant detected: 14B 2025-09-01T15:19:17.079385 - model_type FLOW 2025-09-01T15:19:17.592839 - Loading LoRA: Wan2.2加速\Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank32 with strength: 3.0 2025-09-01T15:19:17.807323 - Using accelerate to load and assign model weights to device... 2025-09-01T15:19:24.630193 - Loading transformer parameters to cuda:0: 83%|██████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▌ | 906/1095 [00:06<00:00, 966.70it/s]2025-09-01T15:19:24.655110 - Loading transformer parameters to cuda:0: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1095/1095 [00:06<00:00, 159.93it/s]2025-09-01T15:19:24.656107 - 2025-09-01T15:19:24.657103 - Using 1052 LoRA weight patches for WanVideo model 2025-09-01T15:19:25.598506 - sigmas: tensor([1.0000, 0.9600, 0.8889, 0.7272, 0.0000]) 2025-09-01T15:19:25.599502 - timesteps: tensor([999, 959, 888, 727], device='cuda:0') 2025-09-01T15:19:25.599502 - Using per-step cfg list: [2.0, 1.0, 1.0, 1.0] 2025-09-01T15:19:25.973765 - Input sequence length: 31050 2025-09-01T15:19:25.973765 - Sampling 89 frames at 720x480 with 4 steps 2025-09-01T15:21:45.420019 - 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 4/4 [02:19<00:00, 26.10s/it]2025-09-01T15:21:45.421016 - 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 4/4 [02:19<00:00, 34.77s/it]2025-09-01T15:21:45.421016 - 2025-09-01T15:21:47.906432 - Allocated memory: memory=0.097 GB 2025-09-01T15:21:47.906432 - Max allocated memory: max_memory=12.714 GB 2025-09-01T15:21:47.907430 - Max reserved memory: max_reserved=12.938 GB 2025-09-01T15:21:48.211045 - CUDA Compute Capability: 8.9 2025-09-01T15:21:48.211045 - Detected model in_channels: 16 2025-09-01T15:21:48.212042 - Model cross attention type: t2v, num_heads: 40, num_layers: 40 2025-09-01T15:21:48.213039 - Model variant detected: 14B 2025-09-01T15:21:48.689720 - model_type FLOW 2025-09-01T15:21:49.195154 - Loading LoRA: Wan2.2加速\Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank32 with strength: 1.0 2025-09-01T15:21:49.294817 - Using accelerate to load and assign model weights to device... 2025-09-01T15:21:58.279127 - Loading transformer parameters to cuda:0: 51%|█████████████████████████████████████████████████████████████████████████████▊ | 553/1095 [00:08<00:18, 29.40it/s]2025-09-01T15:21:58.373810 - Loading transformer parameters to cuda:0: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1095/1095 [00:09<00:00, 120.63it/s]2025-09-01T15:21:58.373810 - 2025-09-01T15:21:58.386767 - Using 1052 LoRA weight patches for WanVideo model 2025-09-01T15:21:58.555691 - sigmas: tensor([0.]) 2025-09-01T15:21:58.556688 - timesteps: tensor([], device='cuda:0', dtype=torch.int64) 2025-09-01T15:22:00.877498 - Input sequence length: 31050 2025-09-01T15:22:00.878495 - Sampling 89 frames at 720x480 with 0 steps 2025-09-01T15:22:01.295938 - 0it [00:00, ?it/s]2025-09-01T15:22:01.297932 - 0it [00:00, ?it/s]2025-09-01T15:22:01.297932 - 2025-09-01T15:22:01.847336 - Allocated memory: memory=0.053 GB 2025-09-01T15:22:01.847336 - Max allocated memory: max_memory=7.034 GB 2025-09-01T15:22:01.847336 - Max reserved memory: max_reserved=7.094 GB 2025-09-01T15:22:02.003449 - !!! Exception during processing !!! cannot access local variable 'callback_latent' where it is not associated with a value 2025-09-01T15:22:02.009429 - Traceback (most recent call last): File "G:\comfyui\ComfyUI\execution.py", line 496, in execute output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 315, in get_output_data return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 289, in _async_map_node_over_list await process_inputs(input_dict, i) File "G:\comfyui\ComfyUI\execution.py", line 277, in process_inputs result = f(**inputs) ^^^^^^^^^^^ File "G:\comfyui\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3988, in process "samples": callback_latent.unsqueeze(0).cpu() if callback is not None else None, ^^^^^^^^^^^^^^^ UnboundLocalError: cannot access local variable 'callback_latent' where it is not associated with a value 2025-09-01T15:22:02.017402 - comfyui lumi batcher overwrite task done2025-09-01T15:22:02.017402 - 2025-09-01T15:22:02.024379 - Prompt executed in 194.37 seconds 2025-09-01T15:26:18.533291 - got prompt 2025-09-01T15:26:18.565355 - Using accelerate to load and assign model weights to device... 2025-09-01T15:26:25.545938 - Loading transformer parameters to cuda:0: 64%|█████████████████████████████████████████████████████████████████████████████████████████████████▌ | 698/1095 [00:06<00:00, 569.60it/s]2025-09-01T15:26:25.588794 - Loading transformer parameters to cuda:0: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1095/1095 [00:07<00:00, 155.95it/s]2025-09-01T15:26:25.588794 - 2025-09-01T15:26:25.589791 - Using 1052 LoRA weight patches for WanVideo model 2025-09-01T15:26:25.652581 - sigmas: tensor([1.0000, 0.9600, 0.8889, 0.7272, 0.0000]) 2025-09-01T15:26:25.652581 - timesteps: tensor([999, 959, 888, 727], device='cuda:0') 2025-09-01T15:26:25.654574 - Using per-step cfg list: [2.0, 1.0, 1.0, 1.0] 2025-09-01T15:26:26.099813 - Input sequence length: 31050 2025-09-01T15:26:26.099813 - Sampling 89 frames at 720x480 with 4 steps 2025-09-01T15:28:34.761959 - 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 4/4 [02:08<00:00, 24.73s/it]2025-09-01T15:28:34.763951 - 100%|█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 4/4 [02:08<00:00, 32.06s/it]2025-09-01T15:28:34.764948 - 2025-09-01T15:28:37.388334 - Allocated memory: memory=0.097 GB 2025-09-01T15:28:37.389330 - Max allocated memory: max_memory=11.671 GB 2025-09-01T15:28:37.390934 - Max reserved memory: max_reserved=12.312 GB 2025-09-01T15:28:37.402353 - Using accelerate to load and assign model weights to device... 2025-09-01T15:28:42.447834 - Loading transformer parameters to cuda:0: 47%|████████████████████████████████████████████████████████████████████████▍ | 518/1095 [00:05<00:02, 266.27it/s]2025-09-01T15:28:42.541592 - Loading transformer parameters to cuda:0: 100%|████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████| 1095/1095 [00:05<00:00, 213.15it/s]2025-09-01T15:28:42.541592 - 2025-09-01T15:28:42.542513 - Using 1052 LoRA weight patches for WanVideo model 2025-09-01T15:28:42.585369 - sigmas: tensor([0.]) 2025-09-01T15:28:42.585369 - timesteps: tensor([], device='cuda:0', dtype=torch.int64) 2025-09-01T15:28:42.834138 - Input sequence length: 31050 2025-09-01T15:28:42.835132 - Sampling 89 frames at 720x480 with 0 steps 2025-09-01T15:28:43.084302 - 0it [00:00, ?it/s]2025-09-01T15:28:43.089282 - 0it [00:00, ?it/s]2025-09-01T15:28:43.089282 - 2025-09-01T15:28:43.383958 - Allocated memory: memory=0.053 GB 2025-09-01T15:28:43.383958 - Max allocated memory: max_memory=7.034 GB 2025-09-01T15:28:43.384950 - Max reserved memory: max_reserved=7.094 GB 2025-09-01T15:28:43.414091 - !!! Exception during processing !!! cannot access local variable 'callback_latent' where it is not associated with a value 2025-09-01T15:28:43.418074 - Traceback (most recent call last): File "G:\comfyui\ComfyUI\execution.py", line 496, in execute output_data, output_ui, has_subgraph, has_pending_tasks = await get_output_data(prompt_id, unique_id, obj, input_data_all, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 315, in get_output_data return_values = await _async_map_node_over_list(prompt_id, unique_id, obj, input_data_all, obj.FUNCTION, allow_interrupt=True, execution_block_cb=execution_block_cb, pre_execute_cb=pre_execute_cb, hidden_inputs=hidden_inputs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "G:\comfyui\ComfyUI\execution.py", line 289, in _async_map_node_over_list await process_inputs(input_dict, i) File "G:\comfyui\ComfyUI\execution.py", line 277, in process_inputs result = f(**inputs) ^^^^^^^^^^^ File "G:\comfyui\ComfyUI\custom_nodes\ComfyUI-WanVideoWrapper\nodes.py", line 3988, in process "samples": callback_latent.unsqueeze(0).cpu() if callback is not None else None, ^^^^^^^^^^^^^^^ UnboundLocalError: cannot access local variable 'callback_latent' where it is not associated with a value 2025-09-01T15:28:43.423058 - comfyui lumi batcher overwrite task done2025-09-01T15:28:43.423058 - 2025-09-01T15:28:43.428044 - Prompt executed in 144.88 seconds ## Attached Workflow Please make sure that workflow does not contain any sensitive information such as API keys or passwords. Workflow too large. Please manually upload the workflow from local file system. ## Additional Context (Please add any additional context or steps to reproduce the error here) 报错的原因,怎么解决

最新推荐

recommend-type

二维码工具(1).zip

二维码工具(1).zip
recommend-type

基于ZXing修改的二维码扫描器.zip

基于ZXing修改的二维码扫描器.zip
recommend-type

在网页上生成二维码.zip

在网页上生成二维码.zip
recommend-type

Node.js构建的运动咖啡馆RESTful API介绍

标题《sportscafeold:体育咖啡馆》指出了项目名称为“体育咖啡馆”,这个名字暗示了该项目可能是一个结合了运动和休闲主题的咖啡馆相关的网络服务平台。该项目运用了多种技术栈,核心的开发语言为JavaScript,这从标签中可以得到明确的信息。 从描述中可以提取以下知识点: 1. **Node.js**:体育咖啡馆项目使用了Node.js作为服务器端运行环境。Node.js是一个基于Chrome V8引擎的JavaScript运行环境,它能够使得JavaScript应用于服务器端开发。Node.js的事件驱动、非阻塞I/O模型使其适合处理大量并发连接,这对于RESTFUL API的构建尤为重要。 2. **Express Framework**:项目中使用了Express框架来创建RESTFUL API。Express是基于Node.js平台,快速、灵活且极简的Web应用开发框架。它提供了构建Web和移动应用的强大功能,是目前最流行的Node.js Web应用框架之一。RESTFUL API是一组遵循REST原则的应用架构,其设计宗旨是让Web服务通过HTTP协议进行通信,并且可以使用各种语言和技术实现。 3. **Mongoose ORM**:这个项目利用了Mongoose作为操作MongoDB数据库的接口。Mongoose是一个对象文档映射器(ODM),它为Node.js提供了MongoDB数据库的驱动。通过Mongoose可以定义数据模型,进行数据库操作和查询,从而简化了对MongoDB数据库的操作。 4. **Passport.js**:项目中采用了Passport.js库来实现身份验证系统。Passport是一个灵活的Node.js身份验证中间件,它支持多种验证策略,例如用户名和密码、OAuth等。它提供了标准化的方法来为用户登录提供认证,是用户认证功能的常用解决方案。 5. **版权信息**:项目的版权声明表明了Sportscafe 2015是版权所有者,这表明项目或其相关内容最早发布于2015年或之前。这可能表明该API背后有商业实体的支持或授权使用。 从【压缩包子文件的文件名称列表】中我们可以了解到,该文件的版本控制仓库使用的是“master”分支。在Git版本控制系统中,“master”分支通常用于存放当前可部署的稳定版本代码。在“master”分支上进行的更改通常都是经过测试且准备发布到生产环境的。 综上所述,我们可以知道体育咖啡馆项目是一个利用现代JavaScript技术栈搭建的后端服务。它包含了处理HTTP请求的Express框架、连接MongoDB数据库的Mongoose库和实现用户身份验证的Passport.js中间件。该项目可用于构建提供体育信息、咖啡馆菜单信息、预约服务等的Web应用或API服务,这为体育咖啡馆的营销、用户体验和数据管理提供了可能。 考虑到文档资料的提及,该项目的安装和API文档应该包含在项目资料中,可能在项目的README文件或其他说明文档中。对于项目的使用者或者开发者而言,这部分文档非常重要,因为它们可以提供详细的信息和指导,帮助用户快速部署和使用该API。 总结来说,这是一套针对体育咖啡馆相关业务的后端解决方案,它使用了流行的开源技术栈,可以灵活地应用于体育咖啡馆的网络服务中,如信息发布、在线预约、会员管理等。
recommend-type

【LNR优化与用户体验】:一文看透互操作优化如何提升用户感知

# 摘要 LNR(邻区关系优化)是提升移动通信网络性能与用户体验的关键技术。本文系统阐述了LNR优化的基本概念、理论基础与技术实现路径,构建了包括信道质量评估、干扰建模及关键性能指标分析在内的数学模型,并探讨了参数配置、邻区管理及AI与大数据在LNR优化中的应用。通过量化用户感知指标,分析了LNR优化对视频流畅度、页面加载速度及切换性能的实际改善效果,并结合典型场景验证了其优化成效。本文进一步展望了LNR技术在5G及未来网络中的发展趋势与技术挑战。
recommend-type

Java1.8 的编程语言、使用场景、版本号、厂商、是否开源、发行日期、终止日期、可替代产品、推荐产品是什么

以下是关于 **Java 1.8** 的全面信息汇总,涵盖其编程语言特性、使用场景、版本号、厂商、是否开源、发行日期、终止支持日期、可替代产品和推荐产品。 --- ## ✅ Java 1.8 综合信息表 | 项目 | 内容 | |------|------| | **编程语言** | Java | | **版本号** | Java 1.8(也称为 Java 8) | | **厂商** | Oracle、Adoptium、Amazon(Corretto)、Azul(Zulu)、Red Hat、IBM 等 | | **是否开源** | ✅ 是(OpenJDK 1.8 是开源的,Oracle
recommend-type

Java开发的教区牧民支持系统介绍

根据给定文件信息,下面将详细阐述相关知识点: ### 标题知识点 #### catecumenus-java: 教区牧民支持系统 - **Java技术栈应用**:标题提到的“catecumenus-java”表明这是一个使用Java语言开发的系统。Java是目前最流行的编程语言之一,广泛应用于企业级应用、Web开发、移动应用等,尤其是在需要跨平台运行的应用中。Java被设计为具有尽可能少的实现依赖,所以它可以在多种处理器上运行。 - **教区牧民支持系统**:从标题来看,这个系统可能面向的是教会管理或教区管理,用来支持牧民(教会领导者或牧师)的日常管理工作。具体功能可能包括教友信息管理、教区活动安排、宗教教育资料库、财务管理、教堂资源调配等。 ### 描述知识点 #### 儿茶类 - **儿茶素(Catechin)**:描述中提到的“儿茶类”可能与“catecumenus”(新信徒、教徒)有关联,暗示这个系统可能与教会或宗教教育相关。儿茶素是一类天然的多酚类化合物,常见于茶、巧克力等植物中,具有抗氧化、抗炎等多种生物活性,但在系统标题中可能并无直接关联。 - **系统版本号**:“0.0.1”表示这是一个非常初期的版本,意味着该系统可能刚刚开始开发,功能尚不完善。 ### 标签知识点 #### Java - **Java语言特点**:标签中明确提到了“Java”,这暗示了整个系统都是用Java编程语言开发的。Java的特点包括面向对象、跨平台(即一次编写,到处运行)、安全性、多线程处理能力等。系统使用Java进行开发,可能看重了这些特点,尤其是在构建可扩展、稳定的后台服务。 - **Java应用领域**:Java广泛应用于企业级应用开发中,包括Web应用程序、大型系统后台、桌面应用以及移动应用(Android)。所以,此系统可能也会涉及这些技术层面。 ### 压缩包子文件的文件名称列表知识点 #### catecumenus-java-master - **Git项目结构**:文件名称中的“master”表明了这是Git版本控制系统中的一个主分支。在Git中,“master”分支通常被用作项目的主干,是默认的开发分支,所有开发工作都是基于此分支进行的。 - **项目目录结构**:在Git项目中,“catecumenus-java”文件夹应该包含了系统的源代码、资源文件、构建脚本、文档等。文件夹可能包含各种子文件夹和文件,比如src目录存放Java源代码,lib目录存放相关依赖库,以及可能的build.xml文件用于构建过程(如Ant或Maven构建脚本)。 ### 结合以上信息的知识点整合 综合以上信息,我们可以推断“catecumenus-java: 教区牧民支持系统”是一个使用Java语言开发的系统,可能正处于初级开发阶段。这个系统可能是为了支持教会内部管理,提供信息管理、资源调度等功能。其使用Java语言的目的可能是希望利用Java的多线程处理能力、跨平台特性和强大的企业级应用支持能力,以实现一个稳定和可扩展的系统。项目结构遵循了Git版本控制的规范,并且可能采用了模块化的开发方式,各个功能模块的代码和资源文件都有序地组织在不同的子文件夹内。 该系统可能采取敏捷开发模式,随着版本号的增加,系统功能将逐步完善和丰富。由于是面向教会的内部支持系统,对系统的用户界面友好性、安全性和数据保护可能会有较高的要求。此外,考虑到宗教性质的敏感性,系统的开发和使用可能还需要遵守特定的隐私和法律法规。
recommend-type

LNR切换成功率提升秘籍:参数配置到网络策略的全面指南

# 摘要 LNR(LTE to NR)切换技术是5G网络部署中的关键环节,直接影
recommend-type

How to install watt toolkit in linux ?

安装 Watt Toolkit(原名 Steam++)在 Linux 系统上通常可以通过编译源码或者使用预编译的二进制文件来完成。Watt Toolkit 是一个开源工具,主要用于加速 Steam 平台的下载速度,支持跨平台运行,因此在 Linux 上也能够很好地工作。 ### 安装步骤 #### 方法一:使用预编译的二进制文件 1. 访问 [Watt Toolkit 的 GitHub 仓库](https://github.com/BeyondDimension/SteamTools) 并下载适用于 Linux 的最新版本。 2. 解压下载的压缩包。 3. 给予可执行权限: ```
recommend-type

PHP实现用户墙上帖子与评论的分享功能

根据给定文件信息,我们可以推导出与“userwallposts”相关的知识点。这里涉及的关键概念包括用户墙面墙(User Wall)、帖子(Posts)和评论(Comments),以及它们在编程语言PHP中的实现方式。用户墙是一种允许用户发布信息,并让他们的朋友或跟随者查看并参与讨论的功能,常见于社交网站。 ### 用户墙概念 用户墙类似于现实生活中的一面墙,用户可以在上面贴上“帖子”来分享自己的想法、照片、视频等信息。其他用户可以在这些帖子下面进行“评论”,类似于在墙上留言。这种互动方式构建了一个社区式的交流环境,增加了用户之间的互动性和参与感。 ### 用户墙的实现 在PHP中实现用户墙功能需要处理前端用户界面和后端服务器逻辑。前端负责展示用户墙、帖子和评论的界面,而后端则负责存储、检索和管理这些数据。 1. **前端实现**:前端可以使用HTML、CSS和JavaScript来构建用户墙的界面。使用AJAX技术可以让用户无需刷新页面即可提交和获取新的帖子和评论。此外,可能还会用到模板引擎(如Twig或Smarty)来动态生成页面内容。 2. **后端实现**:后端PHP代码将负责处理前端发送的请求,如帖子和评论的添加、删除和检索。数据库(如MySQL)将用于存储用户信息、帖子内容、评论以及它们之间的关联关系。 3. **数据库设计**: - 用户表(users):存储用户信息,例如用户名、密码(加密存储)、用户状态等。 - 帖子表(posts):存储帖子信息,例如帖子ID、帖子内容、发帖时间、所属用户ID等。 - 评论表(comments):存储评论信息,包括评论ID、评论内容、评论时间、所属帖子ID和用户ID等。 4. **PHP与数据库交互**:使用PDO(PHP Data Objects)或mysqli扩展来执行SQL语句与数据库进行交互。PDO提供了数据库访问的抽象层,可以连接多种数据库系统,而mysqli则针对MySQL进行了优化。 5. **安全性**: - 输入验证:为了防止SQL注入等安全问题,需要对用户输入进行验证和清理。 - 输出编码:在将数据输出到浏览器之前,应将特殊字符转换为HTML实体。 - 用户认证:用户登录系统时,应采取安全措施如使用会话管理(session management)和防止跨站请求伪造(CSRF)。 6. **功能实现细节**: - 发布帖子:用户输入帖子内容并提交,后端接收并存储到帖子表中。 - 显示帖子:从帖子表中检索所有帖子并展示在用户墙上,包括帖子标题、内容、发布时间和发帖人等信息。 - 发布评论:用户对特定帖子发表评论,后端接收评论信息并将其存储到评论表中。 - 显示评论:为每个帖子显示其下的所有评论,包括评论内容、时间、评论者等。 ### 开源项目实践 “userwallposts-master”暗示了可能存在一个与用户墙帖子和评论相关的开源项目或代码库。这个项目可能包含预设的代码和文件结构,允许开发者下载、安装和配置来创建他们自己的用户墙功能。开发人员可以使用这个项目作为起点,根据自己的需求进一步定制和扩展功能。 在实际开发过程中,还需要考虑系统的可扩展性和维护性。例如,可以设计RESTful API让前端和后端分离,或者使用现代的PHP框架(如Laravel、Symfony等)来简化开发流程和提高代码的组织性。 在总结上述内容后,我们可以了解到用户墙是社交平台中常见的一种功能,其核心功能包括发布帖子和评论。在PHP环境下实现这样的功能需要对前端界面和后端逻辑有深入的理解,以及对数据库设计和交互的安全性有一定的认识。开源项目如“userwallposts-master”可能提供了一个基础框架,帮助开发者快速构建和部署一个用户墙系统。