活动介绍

-- Process 0 terminated with the following error: Traceback (most recent call last): File "/home/chenjiayi/anaconda3/envs/pointcloud/lib/python3.9/site-packages/torch/multiprocessing/spawn.py", line 69, in _wrap fn(i, *args) File "/home/chenjiayi/code/surface_cls2/tools/train_jiayi.py", line 211, in distributed_train_process loss_total = loss_fn(pred_logits, cls_labels.squeeze()) File "/home/chenjiayi/anaconda3/envs/pointcloud/lib/python3.9/site-packages/torch/nn/modules/module.py", line 1501, in _call_impl return forward_call(*args, **kwargs) File "/home/chenjiayi/anaconda3/envs/pointcloud/lib/python3.9/site-packages/torch/nn/modules/loss.py", line 1174, in forward return F.cross_entropy(input, target, weight=self.weight, File "/home/chenjiayi/anaconda3/envs/pointcloud/lib/python3.9/site-packages/torch/nn/functional.py", line 3029, in cross_entropy return torch._C._nn.cross_entropy_loss(input, target, weight, _Reduction.get_enum(reduction), ignore_index,

时间: 2025-08-31 13:12:20 AIGC 浏览: 5
<think>好的,用户遇到了在PyTorch中使用交叉熵损失函数时导致的进程终止错误,特别是在分布式训练环境中。首先,我需要理解这个问题的背景和可能的原因。 根据用户提供的引用内容,错误出现在分布式训练的设置中。用户引用了两段代码,其中一个是错误的,另一个是正确的。在错误代码中,`train_params`中的参数是通过`DDPmodel`获取的,而正确代码则是直接使用原始的`model`。这可能是一个关键点。 交叉熵损失函数(CrossEntropyLoss)在多进程训练中可能会因为输入的数据不符合要求而报错。例如,标签(target)的维度是否正确,或者是否在正确的设备上。在分布式数据并行(DDP)中,每个进程处理不同的数据分片,如果数据没有正确分配到对应的GPU设备上,就会导致错误。 另一个可能的问题是,在使用DDP时,模型已经被包装在`DistributedDataParallel`中,但可能在获取模型参数时错误地引用了DDP包装后的模型,而不是原始模型。例如,在错误代码中,`VGG16_3D.get_1x_lr_params(DDPmodel)`可能无法正确获取原始模型的参数,导致优化器无法正确处理参数,进而引发损失函数计算时的错误。 还需要检查输入数据的形状是否符合CrossEntropyLoss的要求。CrossEntropyLoss期望的输入是未归一化的类别概率(logits),形状为$(N, C)$,其中N是批量大小,C是类别数。标签的形状应为$(N,)$,每个元素是类别索引。如果数据在分布式环境中没有正确分片或形状错误,损失函数会抛出错误,导致进程终止。 此外,进程间同步的问题也可能导致错误。例如,如果某些进程的输入数据为空(可能由于错误的数据划分),在计算损失时会引发异常。需要确保数据采样器(如DistributedSampler)正确配置,每个进程获得有效的数据批次。 接下来,我需要整理解决步骤,包括检查数据设备和形状、正确引用模型参数、验证数据采样器配置、处理可能的空数据情况,以及调试工具的使用。同时,用户提供的正确代码中,使用原始模型而非DDP模型来获取参数,这可能是解决问题的关键点之一,需要重点强调。</think>在PyTorch分布式训练中遇到`cross_entropy_loss`导致进程终止的问题,通常与以下原因相关: --- ### 一、错误原因分析 1. **输入数据设备不一致** - 在分布式训练中,输入数据必须与模型在同一设备上(如GPU)。若数据未通过`to(device)`迁移到对应GPU,或`DataLoader`未配合`DistributedSampler`使用,会导致设备不匹配。 - **验证方法**:检查输入张量是否包含`.device`属性并匹配模型设备。 2. **标签维度错误** - `CrossEntropyLoss`要求标签为**类别索引**(形状为$(N,)$),而非one-hot编码。若标签维度错误(如$(N,C)$),会触发`ValueError`。 - **数学验证**:损失函数计算式为$ \text{Loss} = -\sum_{i=1}^C y_i \log(p_i) $,其中$y_i \in \{0,1\}$且$\sum y_i=1$,但PyTorch内部会自动处理为索引形式。 3. **分布式参数同步问题** - 在错误代码中,`get_1x_lr_params(DDPmodel)`可能错误地从包装后的DDP模型提取参数。正确做法应直接引用原始模型(如正确代码中的`get_1x_lr_params(model)`),因为DDP模型的参数访问需通过`module`属性[^2]。 --- ### 二、解决方案 #### 步骤1:检查数据与设备一致性 ```python # 确保数据和标签在正确设备上 for inputs, labels in dataloader: inputs = inputs.to(device) labels = labels.to(device) # 标签形状应为 (batch_size,) outputs = DDPmodel(inputs) loss = criterion(outputs, labels) ``` #### 步骤2:修正模型参数引用方式 ```python # 正确方式:从原始模型获取参数,而非DDP包装后的模型 train_params = [ {'params': VGG16_3D.get_1x_lr_params(model), 'lr': lr}, # 直接使用model而非DDPmodel {'params': VGG16_3D.get_10x_lr_params(model), 'lr': lr * 10} ] ``` #### 步骤3:验证数据采样器配置 ```python # 使用DistributedSampler确保数据分片正确 from torch.utils.data.distributed import DistributedSampler sampler = DistributedSampler(dataset, shuffle=True) dataloader = DataLoader(dataset, batch_size=64, sampler=sampler) ``` #### 步骤4:处理空数据问题 - 若某个进程的批次数据为空(如数据集无法均匀分配),添加异常处理: ```python try: loss = criterion(outputs, labels) except RuntimeError as e: if "empty tensors" in str(e): print("跳过空批次") continue else: raise e ``` #### 步骤5:调试工具 - 使用`TORCH_DISTRIBUTED_DEBUG=DETAIL`环境变量输出详细日志: ```bash python -m torch.distributed.launch --nproc_per_node=4 train.py --env TORCH_DISTRIBUTED_DEBUG=DETAIL ``` --- ### 三、关键点总结 | 问题类型 | 检查项 | 解决方法 | |----------------|--------------------------------|----------------------------| | 设备不一致 | 数据/标签是否在GPU上 | 添加`.to(device)` | | 标签维度错误 | 标签形状是否为$(N,)$ | 使用`labels.squeeze()` | | 参数引用错误 | 是否从原始模型提取参数 | 改用`model`而非`DDPmodel` | | 数据分片异常 | 是否使用`DistributedSampler` | 配置正确的采样器 | ---
阅读全文

相关推荐

(VScode) zhaohy@workspace2:~/VSCode-main$ python train_test_eval.py --Training True --Testing True --Evaluation True /home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/timm/models/layers/__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning) /home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/timm/models/layers/__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning) /home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/timm/models/layers/__init__.py:48: FutureWarning: Importing from timm.models.layers is deprecated, please import via timm.layers warnings.warn(f"Importing from {__name__} is deprecated, please import via timm.layers", FutureWarning) W0311 15:54:21.788752 136232691705664 torch/multiprocessing/spawn.py:145] Terminating process 2342884 via signal SIGTERM Traceback (most recent call last): File "train_test_eval.py", line 67, in <module> Training.train_net(num_gpus=num_gpus, args=args) File "/home/zhaohy/VSCode-main/Training.py", line 64, in train_net mp.spawn(main, nprocs=num_gpus, args=(num_gpus, args)) File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/multiprocessing/spawn.py", line 281, in spawn return start_processes(fn, args, nprocs, join, daemon, start_method="spawn") File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/multiprocessing/spawn.py", line 237, in start_processes while not context.join(): File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/multiprocessing/spawn.py", line 188, in join raise ProcessRaisedException(msg, error_index, failed_process.pid) torch.multiprocessing.spawn.ProcessRaisedException: -- Process 1 terminated with the following error: Traceback (most recent call last): File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/multiprocessing/spawn.py", line 75, in _wrap fn(i, *args) File "/home/zhaohy/VSCode-main/Training.py", line 71, in main dist.init_process_group(backend='nccl', init_method='env://') # 自动从环境变量读取 RANK/WORLD_SIZE File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/distributed/c10d_logger.py", line 75, in wrapper return func(*args, **kwargs) File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/distributed/c10d_logger.py", line 89, in wrapper func_return = func(*args, **kwargs) File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/distributed/distributed_c10d.py", line 1305, in init_process_group store, rank, world_size = next(rendezvous_iterator) File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/distributed/rendezvous.py", line 234, in _env_rendezvous_handler rank = int(_get_env_or_raise("RANK")) File "/home/zhaohy/anaconda3/envs/VScode/lib/python3.8/site-packages/torch/distributed/rendezvous.py", line 219, in _get_env_or_raise raise _env_error(env_var) ValueError: Error initializing torch.distributed using env:// rendezvous: environment variable RANK expected, but not set

Traceback (most recent call last): File "tools/train.py", line 87, in <module> main() File "tools/train.py", line 76, in main train_model( File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/apis/train.py", line 126, in train_model runner.run(data_loaders, [("train", 1)]) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/epoch_based_runner.py", line 127, in run epoch_runner(data_loaders[i], **kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/runner/epoch_based_runner.py", line 14, in train super().train(data_loader, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/epoch_based_runner.py", line 47, in train for i, data_batch in enumerate(self.data_loader): File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 521, in __next__ data = self._next_data() File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1203, in _next_data return self._process_data(data) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 1229, in _process_data data.reraise() File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/_utils.py", line 425, in reraise raise self.exc_type(msg) AttributeError: Caught AttributeError in DataLoader worker process 0. Original Traceback (most recent call last): File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/_utils/worker.py", line 287, in _worker_loop data = fetcher.fetch(index) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 44, in fetch data = [self.dataset[idx] for idx in possibly_batched_index] File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 44, in data = [self.dataset[idx] for idx in possibly_batched_index] File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/datasets/dataset_wrappers.py", line 68, in __getitem__ return self.dataset[ori_idx] File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/datasets/custom_3d.py", line 293, in __getitem__ data = self.prepare_train_data(idx) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/datasets/custom_3d.py", line 164, in prepare_train_data example = self.pipeline(input_dict) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmdet/datasets/pipelines/compose.py", line 41, in __call__ data = t(data) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/datasets/pipelines/transforms_3d.py", line 218, in __call__ data["radar"].rotate(-theta) AttributeError: 'dict' object has no attribute 'rotate' -------------------------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code. Per user-direction, the job has been aborted. ------------------------------------------------------------------------

(li) lenovo@lenovo-ThinkStation-P920:/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan$ pip install -v --no-cache-dir --force-reinstall . 2>&1 | tee build.log Using pip 25.1 from /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/pip (python 3.10) Looking in indexes: https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple Processing /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan Preparing metadata (setup.py): started Running command python setup.py egg_info A module that was compiled using NumPy 1.x cannot be run in NumPy 2.2.6 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. If you are a user of the module, the easiest solution will be to downgrade to 'numpy<2' or try to upgrade the affected module. We expect that some modules will need time to support NumPy 2. Traceback (most recent call last): File "<string>", line 2, in <module> File "", line 35, in <module> File "/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/setup.py", line 17, in <module> import torch File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/__init__.py", line 1471, in <module> from .functional import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/functional.py", line 9, in <module> import torch.nn.functional as F File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/__init__.py", line 1, in <module> from .modules import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/__init__.py", line 35, in <module> from .transformer import TransformerEncoder, TransformerDecoder, \ File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py", line 20, in <module> device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: _ARRAY_API not found (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.) device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), torch.__version__ = 2.2.0+cu118 CUDA_HOME = /home/lenovo/anaconda3/envs/li CUDA version: 11.8 /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. !! ******************************************************************************** Please consider removing the following classifiers in favor of a SPDX license expression: License :: OSI Approved :: BSD License See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. ******************************************************************************** !! self._finalize_license_expression() running egg_info creating /tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info writing /tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/PKG-INFO writing dependency_links to /tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/dependency_links.txt writing requirements to /tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/requires.txt writing top-level names to /tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/top_level.txt writing manifest file '/tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/SOURCES.txt' reading manifest file '/tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/SOURCES.txt' writing manifest file '/tmp/pip-pip-egg-info-w2djvsss/selective_scan.egg-info/SOURCES.txt' Preparing metadata (setup.py): finished with status 'done' Collecting torch (from selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/0a/7c/0a5b3aee977596459ec45be2220370fde8e017f651fecc40522fd478cb1e/torch-2.7.1-cp310-cp310-manylinux_2_28_x86_64.whl (821.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 821.2/821.2 MB 1.1 MB/s eta 0:00:00 Collecting packaging (from selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/20/12/38679034af332785aac8774540895e234f4d07f7545804097de4b666afd8/packaging-25.0-py3-none-any.whl (66 kB) Collecting ninja (from selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/eb/7a/455d2877fe6cf99886849c7f9755d897df32eaf3a0fba47b56e615f880f7/ninja-1.11.1.4-py3-none-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (422 kB) Collecting einops (from selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/87/62/9773de14fe6c45c23649e98b83231fffd7b9892b6cf863251dc2afa73643/einops-0.8.1-py3-none-any.whl (64 kB) Collecting filelock (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/4d/36/2a115987e2d8c300a974597416d9de88f2444426de9571f4b59b2cca3acc/filelock-3.18.0-py3-none-any.whl (16 kB) Collecting typing-extensions>=4.10.0 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/69/e0/552843e0d356fbb5256d21449fa957fa4eff3bbc135a74a691ee70c7c5da/typing_extensions-4.14.0-py3-none-any.whl (43 kB) Collecting sympy>=1.13.3 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/a2/09/77d55d46fd61b4a135c444fc97158ef34a095e5681d0a6c10b75bf356191/sympy-1.14.0-py3-none-any.whl (6.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 6.3/6.3 MB 946.1 kB/s eta 0:00:00 Link requires a different Python (3.10.18 not in: '>=3.11'): https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/eb/8d/776adee7bbf76365fdd7f2552710282c79a4ead5d2a46408c9043a2b70ba/networkx-3.5-py3-none-any.whl (from https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple/networkx/) (requires-python:>=3.11) Link requires a different Python (3.10.18 not in: '>=3.11'): https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/6c/4f/ccdb8ad3a38e583f214547fd2f7ff1fc160c43a75af88e6aec213404b96a/networkx-3.5.tar.gz (from https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple/networkx/) (requires-python:>=3.11) Link requires a different Python (3.10.18 not in: '>=3.11'): https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/3f/a1/46c1b6e202e3109d2a035b21a7e5534c5bb233ee30752d7f16a0bd4c3989/networkx-3.5rc0-py3-none-any.whl (from https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple/networkx/) (requires-python:>=3.11) Link requires a different Python (3.10.18 not in: '>=3.11'): https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/90/7e/0319606a20ced20730806b9f7fe91d8a92f7da63d76a5c388f87d3f7d294/networkx-3.5rc0.tar.gz (from https://mirrors.tuna.tsinghua.edu.cn/pypi/web/simple/networkx/) (requires-python:>=3.11) Collecting networkx (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/b9/54/dd730b32ea14ea797530a4479b2ed46a6fb250f682a9cfb997e968bf0261/networkx-3.4.2-py3-none-any.whl (1.7 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.7/1.7 MB 883.3 kB/s eta 0:00:00 Collecting jinja2 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/62/a1/3d680cbfd5f4b8f15abc1d571870c5fc3e594bb582bc3b64ea099db13e56/jinja2-3.1.6-py3-none-any.whl (134 kB) Collecting fsspec (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/bb/61/78c7b3851add1481b048b5fdc29067397a1784e2910592bc81bb3f608635/fsspec-2025.5.1-py3-none-any.whl (199 kB) Collecting nvidia-cuda-nvrtc-cu12==12.6.77 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/75/2e/46030320b5a80661e88039f59060d1790298b4718944a65a7f2aeda3d9e9/nvidia_cuda_nvrtc_cu12-12.6.77-py3-none-manylinux2014_x86_64.whl (23.7 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 23.7/23.7 MB 1.1 MB/s eta 0:00:00 Collecting nvidia-cuda-runtime-cu12==12.6.77 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/e1/23/e717c5ac26d26cf39a27fbc076240fad2e3b817e5889d671b67f4f9f49c5/nvidia_cuda_runtime_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (897 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 897.7/897.7 kB 1.7 MB/s eta 0:00:00 Collecting nvidia-cuda-cupti-cu12==12.6.80 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/49/60/7b6497946d74bcf1de852a21824d63baad12cd417db4195fc1bfe59db953/nvidia_cuda_cupti_cu12-12.6.80-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (8.9 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.9/8.9 MB 1.1 MB/s eta 0:00:00 Collecting nvidia-cudnn-cu12==9.5.1.17 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/2a/78/4535c9c7f859a64781e43c969a3a7e84c54634e319a996d43ef32ce46f83/nvidia_cudnn_cu12-9.5.1.17-py3-none-manylinux_2_28_x86_64.whl (571.0 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 571.0/571.0 MB 1.1 MB/s eta 0:00:00 Collecting nvidia-cublas-cu12==12.6.4.1 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/af/eb/ff4b8c503fa1f1796679dce648854d58751982426e4e4b37d6fce49d259c/nvidia_cublas_cu12-12.6.4.1-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (393.1 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 393.1/393.1 MB 889.2 kB/s eta 0:00:00 Collecting nvidia-cufft-cu12==11.3.0.4 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/8f/16/73727675941ab8e6ffd86ca3a4b7b47065edcca7a997920b831f8147c99d/nvidia_cufft_cu12-11.3.0.4-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (200.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 200.2/200.2 MB 761.5 kB/s eta 0:00:00 Collecting nvidia-curand-cu12==10.3.7.77 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/73/1b/44a01c4e70933637c93e6e1a8063d1e998b50213a6b65ac5a9169c47e98e/nvidia_curand_cu12-10.3.7.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (56.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 56.3/56.3 MB 1.1 MB/s eta 0:00:00 Collecting nvidia-cusolver-cu12==11.7.1.2 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/f0/6e/c2cf12c9ff8b872e92b4a5740701e51ff17689c4d726fca91875b07f655d/nvidia_cusolver_cu12-11.7.1.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (158.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 158.2/158.2 MB 881.4 kB/s eta 0:00:00 Collecting nvidia-cusparse-cu12==12.5.4.2 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/06/1e/b8b7c2f4099a37b96af5c9bb158632ea9e5d9d27d7391d7eb8fc45236674/nvidia_cusparse_cu12-12.5.4.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (216.6 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 216.6/216.6 MB 1.4 MB/s eta 0:00:00 Collecting nvidia-cusparselt-cu12==0.6.3 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/3b/9a/72ef35b399b0e183bc2e8f6f558036922d453c4d8237dab26c666a04244b/nvidia_cusparselt_cu12-0.6.3-py3-none-manylinux2014_x86_64.whl (156.8 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 156.8/156.8 MB 887.6 kB/s eta 0:00:00 Collecting nvidia-nccl-cu12==2.26.2 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/67/ca/f42388aed0fddd64ade7493dbba36e1f534d4e6fdbdd355c6a90030ae028/nvidia_nccl_cu12-2.26.2-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (201.3 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 201.3/201.3 MB 1.1 MB/s eta 0:00:00 Collecting nvidia-nvtx-cu12==12.6.77 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/56/9a/fff8376f8e3d084cd1530e1ef7b879bb7d6d265620c95c1b322725c694f4/nvidia_nvtx_cu12-12.6.77-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (89 kB) Collecting nvidia-nvjitlink-cu12==12.6.85 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/9d/d7/c5383e47c7e9bf1c99d5bd2a8c935af2b6d705ad831a7ec5c97db4d82f4f/nvidia_nvjitlink_cu12-12.6.85-py3-none-manylinux2010_x86_64.manylinux_2_12_x86_64.whl (19.7 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 19.7/19.7 MB 1.9 MB/s eta 0:00:00 Collecting nvidia-cufile-cu12==1.11.1.6 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/b2/66/cc9876340ac68ae71b15c743ddb13f8b30d5244af344ec8322b449e35426/nvidia_cufile_cu12-1.11.1.6-py3-none-manylinux2014_x86_64.manylinux_2_17_x86_64.whl (1.1 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.1/1.1 MB 1.9 MB/s eta 0:00:00 Collecting triton==3.3.1 (from torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/8d/a9/549e51e9b1b2c9b854fd761a1d23df0ba2fbc60bd0c13b489ffa518cfcb7/triton-3.3.1-cp310-cp310-manylinux_2_27_x86_64.manylinux_2_28_x86_64.whl (155.6 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 155.6/155.6 MB 758.4 kB/s eta 0:00:00 Collecting setuptools>=40.8.0 (from triton==3.3.1->torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/a3/dc/17031897dae0efacfea57dfd3a82fdd2a2aeb58e0ff71b77b87e44edc772/setuptools-80.9.0-py3-none-any.whl (1.2 MB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.2/1.2 MB 872.9 kB/s eta 0:00:00 Collecting mpmath<1.4,>=1.1.0 (from sympy>=1.13.3->torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/43/e3/7d92a15f894aa0c9c4b49b8ee9ac9850d6e63b03c9c32c0367a13ae62209/mpmath-1.3.0-py3-none-any.whl (536 kB) ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 536.2/536.2 kB 919.5 kB/s eta 0:00:00 Collecting MarkupSafe>=2.0 (from jinja2->torch->selective_scan==0.0.2) Downloading https://mirrors.tuna.tsinghua.edu.cn/pypi/web/packages/22/35/137da042dfb4720b638d2937c38a9c2df83fe32d20e8c8f3185dbfef05f7/MarkupSafe-3.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (20 kB) Building wheels for collected packages: selective_scan DEPRECATION: Building 'selective_scan' using the legacy setup.py bdist_wheel mechanism, which will be removed in a future version. pip 25.3 will enforce this behaviour change. A possible replacement is to use the standardized build interface by setting the --use-pep517 option, (possibly combined with --no-build-isolation), or adding a pyproject.toml file to the source tree of 'selective_scan'. Discussion can be found at https://github.com/pypa/pip/issues/6334 Building wheel for selective_scan (setup.py): started Running command python setup.py bdist_wheel A module that was compiled using NumPy 1.x cannot be run in NumPy 2.2.6 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. If you are a user of the module, the easiest solution will be to downgrade to 'numpy<2' or try to upgrade the affected module. We expect that some modules will need time to support NumPy 2. Traceback (most recent call last): File "<string>", line 2, in <module> File "", line 35, in <module> File "/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/setup.py", line 17, in <module> import torch File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/__init__.py", line 1471, in <module> from .functional import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/functional.py", line 9, in <module> import torch.nn.functional as F File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/__init__.py", line 1, in <module> from .modules import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/__init__.py", line 35, in <module> from .transformer import TransformerEncoder, TransformerDecoder, \ File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py", line 20, in <module> device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: _ARRAY_API not found (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.) device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), torch.__version__ = 2.2.0+cu118 CUDA_HOME = /home/lenovo/anaconda3/envs/li CUDA version: 11.8 /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. !! ******************************************************************************** Please consider removing the following classifiers in favor of a SPDX license expression: License :: OSI Approved :: BSD License See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. ******************************************************************************** !! self._finalize_license_expression() running bdist_wheel running build running build_ext /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py:425: UserWarning: There are no g++ version bounds defined for CUDA version 11.8 warnings.warn(f'There are no {compiler_name} version bounds defined for CUDA version {cuda_str_version}') building 'selective_scan_cuda_core' extension creating /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus Emitting ninja build file /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/build.ninja... Compiling objects... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) [1/3] c++ -MMD -MF '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan.o'.d -pthread -B /home/lenovo/anaconda3/envs/li/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /home/lenovo/anaconda3/envs/li/include -fPIC -O2 -isystem /home/lenovo/anaconda3/envs/li/include -fPIC '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan.cpp' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan.o' -O3 -std=c++17 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 FAILED: /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan.o c++ -MMD -MF '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan.o'.d -pthread -B /home/lenovo/anaconda3/envs/li/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O2 -Wall -fPIC -O2 -isystem /home/lenovo/anaconda3/envs/li/include -fPIC -O2 -isystem /home/lenovo/anaconda3/envs/li/include -fPIC '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan.cpp' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan.o' -O3 -std=c++17 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 In file included from /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/ATen/cuda/CUDAContext.h:3, from /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan.cpp:5: /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/ATen/cuda/CUDAContextLight.h:6:10: fatal error: cuda_runtime_api.h: 没有那个文件或目录 6 | #include <cuda_runtime_api.h> | ^~~~~~~~~~~~~~~~~~~~ compilation terminated. [2/3] /home/lenovo/anaconda3/envs/li/bin/nvcc --generate-dependencies-with-compile --dependency-output '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_fwd.o'.d '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan_core_fwd.cu' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_fwd.o' -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -O3 -std=c++17 -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_BFLOAT16_OPERATORS__ -U__CUDA_NO_BFLOAT16_CONVERSIONS__ -U__CUDA_NO_BFLOAT162_OPERATORS__ -U__CUDA_NO_BFLOAT162_CONVERSIONS__ --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -lineinfo -arch=sm_86 --threads 4 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 FAILED: /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_fwd.o /home/lenovo/anaconda3/envs/li/bin/nvcc --generate-dependencies-with-compile --dependency-output '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_fwd.o'.d '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan_core_fwd.cu' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_fwd.o' -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -O3 -std=c++17 -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_BFLOAT16_OPERATORS__ -U__CUDA_NO_BFLOAT16_CONVERSIONS__ -U__CUDA_NO_BFLOAT162_OPERATORS__ -U__CUDA_NO_BFLOAT162_CONVERSIONS__ --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -lineinfo -arch=sm_86 --threads 4 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 <command-line>: fatal error: cuda_runtime.h: 没有那个文件或目录 compilation terminated. <command-line>: fatal error: cuda_runtime.h: 没有那个文件或目录 compilation terminated. fatal : Could not open input file /tmp/tmpxft_00005756_00000000-7_selective_scan_core_fwd.cpp1.ii [3/3] /home/lenovo/anaconda3/envs/li/bin/nvcc --generate-dependencies-with-compile --dependency-output '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_bwd.o'.d '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan_core_bwd.cu' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_bwd.o' -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -O3 -std=c++17 -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_BFLOAT16_OPERATORS__ -U__CUDA_NO_BFLOAT16_CONVERSIONS__ -U__CUDA_NO_BFLOAT162_OPERATORS__ -U__CUDA_NO_BFLOAT162_CONVERSIONS__ --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -lineinfo -arch=sm_86 --threads 4 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 FAILED: /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_bwd.o /home/lenovo/anaconda3/envs/li/bin/nvcc --generate-dependencies-with-compile --dependency-output '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_bwd.o'.d '-I/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan' -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/torch/csrc/api/include -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/TH -I/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/include/THC -I/home/lenovo/anaconda3/envs/li/include -I/home/lenovo/anaconda3/envs/li/include/python3.10 -c -c '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/csrc/selective_scan/cus/selective_scan_core_bwd.cu' -o '/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/build/temp.linux-x86_64-cpython-310/csrc/selective_scan/cus/selective_scan_core_bwd.o' -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -O3 -std=c++17 -U__CUDA_NO_HALF_OPERATORS__ -U__CUDA_NO_HALF_CONVERSIONS__ -U__CUDA_NO_BFLOAT16_OPERATORS__ -U__CUDA_NO_BFLOAT16_CONVERSIONS__ -U__CUDA_NO_BFLOAT162_OPERATORS__ -U__CUDA_NO_BFLOAT162_CONVERSIONS__ --expt-relaxed-constexpr --expt-extended-lambda --use_fast_math --ptxas-options=-v -lineinfo -arch=sm_86 --threads 4 -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1011"' -DTORCH_EXTENSION_NAME=selective_scan_cuda_core -D_GLIBCXX_USE_CXX11_ABI=0 <command-line>: fatal error: cuda_runtime.h: 没有那个文件或目录 compilation terminated. <command-line>: fatal error: cuda_runtime.h: 没有那个文件或目录 compilation terminated. fatal : Could not open input file /tmp/tmpxft_00005755_00000000-7_selective_scan_core_bwd.cpp1.ii ninja: build stopped: subcommand failed. Traceback (most recent call last): File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py", line 2096, in _run_ninja_build subprocess.run( File "/home/lenovo/anaconda3/envs/li/lib/python3.10/subprocess.py", line 526, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "<string>", line 2, in <module> File "", line 35, in <module> File "/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/setup.py", line 146, in <module> setup( File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/__init__.py", line 117, in setup return distutils.core.setup(**attrs) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 186, in setup return run_commands(dist) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/core.py", line 202, in run_commands dist.run_commands() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 1002, in run_commands self.run_command(cmd) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py", line 1104, in run_command super().run_command(command) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 1021, in run_command cmd_obj.run() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/command/bdist_wheel.py", line 370, in run self.run_command("build") File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 357, in run_command self.distribution.run_command(command) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py", line 1104, in run_command super().run_command(command) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 1021, in run_command cmd_obj.run() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/command/build.py", line 135, in run self.run_command(cmd_name) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/cmd.py", line 357, in run_command self.distribution.run_command(command) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py", line 1104, in run_command super().run_command(command) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/dist.py", line 1021, in run_command cmd_obj.run() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 99, in run _build_ext.run(self) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 368, in run self.build_extensions() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py", line 871, in build_extensions build_ext.build_extensions(self) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 484, in build_extensions self._build_extensions_serial() File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 510, in _build_extensions_serial self.build_extension(ext) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/command/build_ext.py", line 264, in build_extension _build_ext.build_extension(self, ext) File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/_distutils/command/build_ext.py", line 565, in build_extension objects = self.compiler.compile( File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py", line 684, in unix_wrap_ninja_compile _write_ninja_file_and_compile_objects( File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py", line 1774, in _write_ninja_file_and_compile_objects _run_ninja_build( File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/utils/cpp_extension.py", line 2112, in _run_ninja_build raise RuntimeError(message) from e RuntimeError: Error compiling objects for extension error: subprocess-exited-with-error × python setup.py bdist_wheel did not run successfully. │ exit code: 1 ╰─> See above for output. note: This error originates from a subprocess, and is likely not a problem with pip. full command: /home/lenovo/anaconda3/envs/li/bin/python3.10 -u -c ' exec(compile('"'"''"'"''"'"' # This is -- a caller that pip uses to run setup.py # # - It imports setuptools before invoking setup.py, to enable projects that directly # import from distutils.core to work with newer packaging standards. # - It provides a clear error message when setuptools is not installed. # - It sets sys.argv[0] to the underlying setup.py, when invoking setup.py so # setuptools doesn'"'"'t think the script is -c. This avoids the following warning: # manifest_maker: standard file '"'"'-c'"'"' not found". # - It generates a shim setup.py, for handling setup.cfg-only projects. import os, sys, tokenize, traceback try: import setuptools except ImportError: print( "ERROR: Can not execute setup.py since setuptools failed to import in " "the build environment with exception:", file=sys.stderr, ) traceback.print_exc() sys.exit(1) __file__ = %r sys.argv[0] = __file__ if os.path.exists(__file__): filename = __file__ with tokenize.open(__file__) as f: setup_py_code = f.read() else: filename = "<auto-generated setuptools caller>" setup_py_code = "from setuptools import setup; setup()" exec(compile(setup_py_code, filename, "exec")) '"'"''"'"''"'"' % ('"'"'/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/setup.py'"'"',), "", "exec"))' bdist_wheel -d /tmp/pip-wheel-fm2tmb40 cwd: /media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/ Building wheel for selective_scan (setup.py): finished with status 'error' ERROR: Failed building wheel for selective_scan Running setup.py clean for selective_scan Running command python setup.py clean A module that was compiled using NumPy 1.x cannot be run in NumPy 2.2.6 as it may crash. To support both 1.x and 2.x versions of NumPy, modules must be compiled with NumPy 2.0. Some module may need to rebuild instead e.g. with 'pybind11>=2.12'. If you are a user of the module, the easiest solution will be to downgrade to 'numpy<2' or try to upgrade the affected module. We expect that some modules will need time to support NumPy 2. Traceback (most recent call last): File "<string>", line 2, in <module> File "", line 35, in <module> File "/media/lenovo/PHILIPS/Mamba_20250409_ torch22_cuda118/VMamba-main/kernels/selective_scan/setup.py", line 17, in <module> import torch File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/__init__.py", line 1471, in <module> from .functional import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/functional.py", line 9, in <module> import torch.nn.functional as F File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/__init__.py", line 1, in <module> from .modules import * # noqa: F403 File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/__init__.py", line 35, in <module> from .transformer import TransformerEncoder, TransformerDecoder, \ File "/home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py", line 20, in <module> device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/torch/nn/modules/transformer.py:20: UserWarning: Failed to initialize NumPy: _ARRAY_API not found (Triggered internally at ../torch/csrc/utils/tensor_numpy.cpp:84.) device: torch.device = torch.device(torch._C._get_default_device()), # torch.device('cpu'), torch.__version__ = 2.2.0+cu118 CUDA_HOME = /home/lenovo/anaconda3/envs/li CUDA version: 11.8 /home/lenovo/anaconda3/envs/li/lib/python3.10/site-packages/setuptools/dist.py:759: SetuptoolsDeprecationWarning: License classifiers are deprecated. !! ******************************************************************************** Please consider removing the following classifiers in favor of a SPDX license expression: License :: OSI Approved :: BSD License See https://packaging.python.org/en/latest/guides/writing-pyproject-toml/#license for details. ******************************************************************************** !! self._finalize_license_expression() running clean removing 'build/temp.linux-x86_64-cpython-310' (and everything under it) 'build/lib.linux-x86_64-cpython-310' does not exist -- can't clean it 'build/bdist.linux-x86_64' does not exist -- can't clean it 'build/scripts-3.10' does not exist -- can't clean it removing 'build' Failed to build selective_scan ERROR: Failed to build installable wheels for some pyproject.toml based projects (selective_scan)

Traceback (most recent call last): File "tools/test.py", line 230, in <module> main() File "tools/test.py", line 184, in main checkpoint = load_checkpoint(model, args.checkpoint, map_location="cpu") File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/checkpoint.py", line 531, in load_checkpoint checkpoint = _load_checkpoint(filename, map_location, logger) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/checkpoint.py", line 470, in _load_checkpoint return CheckpointLoader.load_checkpoint(filename, map_location, logger) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/checkpoint.py", line 249, in load_checkpoint return checkpoint_loader(filename, map_location) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/checkpoint.py", line 266, in load_from_local checkpoint = torch.load(filename, map_location=map_location) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/serialization.py", line 608, in load return _legacy_load(opened_file, map_location, pickle_module, **pickle_load_args) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/serialization.py", line 777, in _legacy_load magic_number = pickle_module.load(f, **pickle_load_args) EOFError: Ran out of input -------------------------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code. Per user-direction, the job has been aborted. -------------------------------------------------------------------------- -------------------------------------------------------------------------- mpirun detected that one or more processes exited with non-zero status, thus causing the job to be terminated. The first process to do so was: Process name: [[15407,1],0]

PREFIX=/home/ubuntu22/miniconda3 Unpacking payload ... entry_point.py:256: DeprecationWarning: Python 3.14 will, by default, filter extracted tar archives and reject files or modify their metadata. Use the filter argument to control this behavior. entry_point.py:256: DeprecationWarning: Python 3.14 will, by default, filter extracted tar archives and reject files or modify their metadata. Use the filter argument to control this behavior. concurrent.futures.process._RemoteTraceback: ''' Traceback (most recent call last): File "concurrent/futures/process.py", line 432, in wait_result_broken_or_wakeup File "multiprocessing/connection.py", line 251, in recv TypeError: InvalidArchiveError.__init__() missing 1 required positional argument: 'msg' ''' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "entry_point.py", line 244, in _constructor_extract_conda_pkgs File "concurrent/futures/_base.py", line 449, in result File "concurrent/futures/_base.py", line 401, in __get_result concurrent.futures.process.BrokenProcessPool: A process in the process pool was terminated abruptly while the future was running or pending. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "entry_point.py", line 691, in <module> File "entry_point.py", line 681, in main File "entry_point.py", line 582, in _constructor_subcommand File "entry_point.py", line 246, in _constructor_extract_conda_pkgs RuntimeError: Failed to extract /home/ubuntu22/miniconda3/pkgs/python-3.13.2-hf623796_100_cp313.conda: A process in the process pool was terminated abruptly while the future was running or pending. [PYI-3543:ERROR] Failed to execute script 'entry_point' due to unhandled exception!

load checkpoint from local path: pretrained/bevfusion-det.pth [ ] 0/81, elapsed: 0s, ETA:Traceback (most recent call last): File "tools/test.py", line 230, in <module> main() File "tools/test.py", line 203, in main outputs = multi_gpu_test(model, data_loader, args.tmpdir, args.gpu_collect) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmdet/apis/test.py", line 98, in multi_gpu_test result = model(return_loss=False, rescale=True, **data) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in _call_impl return forward_call(*input, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/parallel/distributed.py", line 799, in forward output = self.module(*inputs[0], **kwargs[0]) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in _call_impl return forward_call(*input, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/fp16_utils.py", line 128, in new_func output = old_func(*new_args, **new_kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/fusion_models/bevfusion.py", line 253, in forward outputs = self.forward_single( File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/fp16_utils.py", line 128, in new_func output = old_func(*new_args, **new_kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/fusion_models/bevfusion.py", line 301, in forward_single feature = self.extract_camera_features( File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/fusion_models/bevfusion.py", line 133, in extract_camera_features x = self.encoders["camera"]["vtransform"]( File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in _call_impl return forward_call(*input, **kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/vtransforms/depth_lss.py", line 100, in forward x = super().forward(*args, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/fp16_utils.py", line 214, in new_func output = old_func(*new_args, **new_kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/vtransforms/base.py", line 349, in forward x = self.get_cam_feats(img, depth, mats_dict) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/mmcv/runner/fp16_utils.py", line 214, in new_func output = old_func(*new_args, **new_kwargs) File "/media/wangbaihui/1ecf654b-afad-4dab-af7b-e34b00dda87a/bevfusion/mmdet3d/models/vtransforms/depth_lss.py", line 88, in get_cam_feats d = self.dtransform(d) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in _call_impl return forward_call(*input, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/container.py", line 139, in forward input = module(input) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/module.py", line 1051, in _call_impl return forward_call(*input, **kwargs) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 443, in forward return self._conv_forward(input, self.weight, self.bias) File "/home/wangbaihui/anaconda3/envs/vad/lib/python3.8/site-packages/torch/nn/modules/conv.py", line 439, in _conv_forward return F.conv2d(input, weight, bias, self.stride, RuntimeError: Given groups=1, weight of size [8, 1, 1, 1], expected input[6, 6, 256, 704] to have 1 channels, but got 6 channels instead -------------------------------------------------------------------------- Primary job terminated normally, but 1 process returned a non-zero exit code. Per user-direction, the job has been aborted.

[/public/home/pengjy/anaconda3] >>> PREFIX=/public/home/pengjy/anaconda3 WARNING: md5sum mismatch of tar archive expected: 8a581514493c9e0a1cbd425bc1c7dd90 got: 614f6284c34f91affd38a1be2e4be076 - Unpacking payload ... Traceback (most recent call last): File "entry_point.py", line 76, in <module> File "tarfile.py", line 2024, in extractall File "tarfile.py", line 2065, in extract File "tarfile.py", line 2137, in _extract_member File "tarfile.py", line 2186, in makefile File "tarfile.py", line 249, in copyfileobj tarfile.ReadError: unexpected end of data [210095] Failed to execute script entry_point concurrent.futures.process._RemoteTraceback: ''' Traceback (most recent call last): File "concurrent/futures/process.py", line 368, in _queue_management_worker File "multiprocessing/connection.py", line 251, in recv TypeError: __init__() missing 1 required positional argument: 'msg' ''' The above exception was the direct cause of the following exception: Traceback (most recent call last): File "entry_point.py", line 69, in <module> File "concurrent/futures/process.py", line 484, in _chain_from_iterable_of_lists File "concurrent/futures/_base.py", line 611, in result_iterator File "concurrent/futures/_base.py", line 439, in result File "concurrent/futures/_base.py", line 388, in __get_result concurrent.futures.process.BrokenProcessPool: A process in the process pool was terminated abruptly while the future was running or pending. [210105] Failed to execute script entry_point 是什么问题,如何解决?

rank1]:[E830 06:13:19.780493643 ProcessGroupNCCL.cpp:632] [Rank 1] Watchdog caught collective operation timeout: WorkNCCL(SeqNum=146806, OpType=ALLREDUCE, NumelIn=3675597, NumelOut=3675597, Timeout(ms)=600000) ran for 600472 milliseconds before timing out.████████████████ | 519/804 [18:24<09:52, 2.08s/it, loss=0.0163] [rank1]:[E830 06:13:57.611628340 ProcessGroupNCCL.cpp:2271] [PG ID 0 PG GUID 0(default_pg) Rank 1] failure detected by watchdog at work sequence id: 146806 PG status: last enqueued work: 146806, last completed work: 146804 [rank1]:[E830 06:13:57.634486146 ProcessGroupNCCL.cpp:670] Stack trace of the failed collective not found, potentially because FlightRecorder is disabled. You can enable it by setting TORCH_NCCL_TRACE_BUFFER_SIZE to a non-zero value. [rank1]:[E830 06:14:02.704255369 ProcessGroupNCCL.cpp:2106] [PG ID 0 PG GUID 0(default_pg) Rank 1] First PG on this rank to signal dumping. [rank0]:[E830 06:14:02.725916508 ProcessGroupNCCL.cpp:1685] [PG ID 0 PG GUID 0(default_pg) Rank 0] Observed flight recorder dump signal from another rank via TCPStore. [rank1]:[E830 06:14:02.768576450 ProcessGroupNCCL.cpp:1746] [PG ID 0 PG GUID 0(default_pg) Rank 1] Received a dump signal due to a collective timeout from rank 1 and we will try our best to dump the debug info. Last enqueued NCCL work: 146806, last completed NCCL work: 146804.This is most likely caused by incorrect usages of collectives, e.g., wrong sizes used across ranks, the order of collectives is not same for all ranks or the scheduled collective, for some reason, didn't run. Additionally, this can be caused by GIL deadlock or other reasons such as network errors or bugs in the communications library (e.g. NCCL), etc. [rank0]:[E830 06:14:02.768580472 ProcessGroupNCCL.cpp:1746] [PG ID 0 PG GUID 0(default_pg) Rank 0] Received a dump signal due to a collective timeout from rank 1 and we will try our best to dump the debug info. Last enqueued NCCL work: 146804, last completed NCCL work: 146804.This is most likely caused by incorrect usages of collectives, e.g., wrong sizes used across ranks, the order of collectives is not same for all ranks or the scheduled collective, for some reason, didn't run. Additionally, this can be caused by GIL deadlock or other reasons such as network errors or bugs in the communications library (e.g. NCCL), etc. [rank0]:[E830 06:14:08.324151488 ProcessGroupNCCL.cpp:1536] [PG ID 0 PG GUID 0(default_pg) Rank 0] ProcessGroupNCCL preparing to dump debug info. Include stack trace: 1 [rank1]:[E830 06:14:08.324155500 ProcessGroupNCCL.cpp:1536] [PG ID 0 PG GUID 0(default_pg) Rank 1] ProcessGroupNCCL preparing to dump debug info. Include stack trace: 1 [rank1]:[E830 06:16:17.610834920 ProcessGroupNCCL.cpp:684] [Rank 1] Some NCCL operations have failed or timed out. Due to the asynchronous nature of CUDA kernels, subsequent GPU operations might run on corrupted/incomplete data. [rank1]:[E830 06:16:17.667404073 ProcessGroupNCCL.cpp:698] [Rank 1] To avoid data inconsistency, we are taking the entire process down. [rank1]:[E830 06:16:45.611443648 ProcessGroupNCCL.cpp:1899] [PG ID 0 PG GUID 0(default_pg) Rank 1] Process group watchdog thread terminated with exception: [Rank 1] Watchdog caught collective operation timeout: WorkNCCL(SeqNum=146806, OpType=ALLREDUCE, NumelIn=3675597, NumelOut=3675597, Timeout(ms)=600000) ran for 600472 milliseconds before timing out. Exception raised from checkTimeout at /pytorch/torch/csrc/distributed/c10d/ProcessGroupNCCL.cpp:635 (most recent call first): frame #0: c10::Error::Error(c10::SourceLocation, std::__cxx11::basic_string<char, std::char_traits<char>, std::allocator<char> >) + 0x98 (0x7a36a8d785e8 in /root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/lib/libc10.so) frame #1: c10d::ProcessGroupNCCL::WorkNCCL::checkTimeout(std::optional<std::chrono::duration<long, std::ratio<1l, 1000l> > >) + 0x23d (0x7a36533e2a6d in /root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so) frame #2: c10d::ProcessGroupNCCL::watchdogHandler() + 0xc80 (0x7a36533e47f0 in /root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so) frame #3: c10d::ProcessGroupNCCL::ncclCommWatchdog() + 0x14d (0x7a36533e5efd in /root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/lib/libtorch_cuda.so) frame #4: <unknown function> + 0xdbbf4 (0x7a3642edbbf4 in /root/anaconda3/envs/torch/bin/../lib/libstdc++.so.6) frame #5: <unknown function> + 0x94ac3 (0x7a36aa694ac3 in /lib/x86_64-linux-gnu/libc.so.6) frame #6: <unknown function> + 0x126850 (0x7a36aa726850 in /lib/x86_64-linux-gnu/libc.so.6) [rank0]:[F830 06:22:40.572585407 ProcessGroupNCCL.cpp:1557] [PG ID 0 PG GUID 0(default_pg) Rank 0] [PG ID 0 PG GUID 0(default_pg) Rank 0] Terminating the process after attempting to dump debug info, due to collective timeout or exception. [rank1]: Traceback (most recent call last): [rank1]: File "/root/anaconda3/envs/torch/lib/python3.12/multiprocessing/resource_sharer.py", line 139, in _serve [rank1]: msg = conn.recv() [rank1]: ^^^^^^^^^^^ [rank1]: File "/root/anaconda3/envs/torch/lib/python3.12/multiprocessing/connection.py", line 249, in recv [rank1]: buf = self._recv_bytes() [rank1]: ^^^^^^^^^^^^^^^^^^ [rank1]: File "/root/anaconda3/envs/torch/lib/python3.12/multiprocessing/connection.py", line 413, in _recv_bytes [rank1]: buf = self._recv(4) [rank1]: ^^^^^^^^^^^^^ [rank1]: File "/root/anaconda3/envs/torch/lib/python3.12/multiprocessing/connection.py", line 382, in _recv [rank1]: raise EOFError [rank1]: EOFError W0830 06:40:48.263000 3372 site-packages/torch/distributed/elastic/multiprocessing/api.py:900] Sending process 3385 closing signal SIGTERM E0830 06:40:48.433000 3372 site-packages/torch/distributed/elastic/multiprocessing/api.py:874] failed (exitcode: -6) local_rank: 1 (pid: 3386) of binary: /root/anaconda3/envs/torch/bin/python3.12 Traceback (most recent call last): File "/root/anaconda3/envs/torch/bin/torchrun", line 8, in <module> sys.exit(main()) ^^^^^^ File "/root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/distributed/elastic/multiprocessing/errors/__init__.py", line 355, in wrapper return f(*args, **kwargs) ^^^^^^^^^^^^^^^^^^ File "/root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/distributed/run.py", line 892, in main run(args) File "/root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/distributed/run.py", line 883, in run elastic_launch( File "/root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/distributed/launcher/api.py", line 139, in __call__ return launch_agent(self._config, self._entrypoint, list(args)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/root/anaconda3/envs/torch/lib/python3.12/site-packages/torch/distributed/launcher/api.py", line 270, in launch_agent raise ChildFailedError( torch.distributed.elastic.multiprocessing.errors.ChildFailedError: ===================================================== train.py FAILED ----------------------------------------------------- Failures: <NO_OTHER_FAILURES> ----------------------------------------------------- Root Cause (first observed failure): [0]: time : 2025-08-30_06:40:48 host : ub-MS-7A93 rank : 1 (local_rank: 1) exitcode : -6 (pid: 3386) error_file: <N/A> traceback : Signal 6 (SIGABRT) received by PID 3386

(openpcdet) autodriver@ubuntu:~/wzk2/OpenPCDet$ python setup.py develop running develop /home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/command/develop.py:41: EasyInstallDeprecationWarning: easy_install command is deprecated. !! ******************************************************************************** Please avoid running setup.py and easy_install. Instead, use pypa/build, pypa/installer or other standards-based tools. See https://github.com/pypa/setuptools/issues/917 for details. ******************************************************************************** !! easy_install.initialize_options(self) /home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/cmd.py:66: SetuptoolsDeprecationWarning: setup.py install is deprecated. !! ******************************************************************************** Please avoid running setup.py directly. Instead, use pypa/build, pypa/installer or other standards-based tools. See https://blog.ganssle.io/articles/2021/10/setup-py-deprecated.html for details. ******************************************************************************** !! self.initialize_options() running egg_info writing pcdet.egg-info/PKG-INFO writing dependency_links to pcdet.egg-info/dependency_links.txt writing requirements to pcdet.egg-info/requires.txt writing top-level names to pcdet.egg-info/top_level.txt reading manifest file 'pcdet.egg-info/SOURCES.txt' adding license file 'LICENSE' writing manifest file 'pcdet.egg-info/SOURCES.txt' running build_ext building 'pcdet.ops.iou3d_nms.iou3d_nms_cuda' extension Emitting ninja build file /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/build.ninja... Compiling objects... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) ninja: no work to do. g++ -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -shared /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/iou3d_nms/src/iou3d_cpu.o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/iou3d_nms/src/iou3d_nms.o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/iou3d_nms/src/iou3d_nms_api.o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/iou3d_nms/src/iou3d_nms_kernel.o -L/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/lib -L/usr/local/cuda-11.4/lib64 -lc10 -ltorch -ltorch_cpu -ltorch_python -lcudart -lc10_cuda -ltorch_cuda -o build/lib.linux-aarch64-cpython-38/pcdet/ops/iou3d_nms/iou3d_nms_cuda.cpython-38-aarch64-linux-gnu.so building 'pcdet.ops.roiaware_pool3d.roiaware_pool3d_cuda' extension Emitting ninja build file /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/build.ninja... Compiling objects... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) ninja: no work to do. g++ -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -shared /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/roiaware_pool3d/src/roiaware_pool3d.o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/roiaware_pool3d/src/roiaware_pool3d_kernel.o -L/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/lib -L/usr/local/cuda-11.4/lib64 -lc10 -ltorch -ltorch_cpu -ltorch_python -lcudart -lc10_cuda -ltorch_cuda -o build/lib.linux-aarch64-cpython-38/pcdet/ops/roiaware_pool3d/roiaware_pool3d_cuda.cpython-38-aarch64-linux-gnu.so building 'pcdet.ops.roipoint_pool3d.roipoint_pool3d_cuda' extension Emitting ninja build file /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/build.ninja... Compiling objects... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) ninja: no work to do. g++ -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -shared /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/roipoint_pool3d/src/roipoint_pool3d.o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/roipoint_pool3d/src/roipoint_pool3d_kernel.o -L/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/lib -L/usr/local/cuda-11.4/lib64 -lc10 -ltorch -ltorch_cpu -ltorch_python -lcudart -lc10_cuda -ltorch_cuda -o build/lib.linux-aarch64-cpython-38/pcdet/ops/roipoint_pool3d/roipoint_pool3d_cuda.cpython-38-aarch64-linux-gnu.so building 'pcdet.ops.pointnet2.pointnet2_stack.pointnet2_stack_cuda' extension Emitting ninja build file /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/build.ninja... Compiling objects... Allowing ninja to set a default number of workers... (overridable by setting the environment variable MAX_JOBS=N) [1/6] c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 FAILED: /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.o c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/ball_query.cpp:10:10: fatal error: THC/THC.h: 没有那个文件或目录 10 | #include <THC/THC.h> | ^~~~~~~~~~~ compilation terminated. [2/6] c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 FAILED: /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.o c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/interpolate.cpp:10:10: fatal error: THC/THC.h: 没有那个文件或目录 10 | #include <THC/THC.h> | ^~~~~~~~~~~ compilation terminated. [3/6] c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 FAILED: /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.o c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query.cpp:3:10: fatal error: THC/THC.h: 没有那个文件或目录 3 | #include <THC/THC.h> | ^~~~~~~~~~~ compilation terminated. [4/6] c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 FAILED: /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.o c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/group_points.cpp:12:10: fatal error: THC/THC.h: 没有那个文件或目录 12 | #include <THC/THC.h> | ^~~~~~~~~~~ compilation terminated. [5/6] c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 FAILED: /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.o c++ -MMD -MF /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.o.d -pthread -B /home/autodriver/miniconda3/envs/openpcdet/compiler_compat -Wno-unused-result -Wsign-compare -DNDEBUG -fwrapv -O3 -Wall -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -O3 -isystem /home/autodriver/miniconda3/envs/openpcdet/include -fPIC -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.cpp -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.o -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -std=c++14 /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/sampling.cpp:4:10: fatal error: THC/THC.h: 没有那个文件或目录 4 | #include <THC/THC.h> | ^~~~~~~~~~~ compilation terminated. [6/6] /usr/local/cuda-11.4/bin/nvcc -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/torch/csrc/api/include -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/TH -I/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/include/THC -I/usr/local/cuda-11.4/include -I/home/autodriver/miniconda3/envs/openpcdet/include/python3.8 -c -c /home/autodriver/wzk2/OpenPCDet/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query_gpu.cu -o /home/autodriver/wzk2/OpenPCDet/build/temp.linux-aarch64-cpython-38/pcdet/ops/pointnet2/pointnet2_stack/src/voxel_query_gpu.o -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ --expt-relaxed-constexpr --compiler-options ''"'"'-fPIC'"'"'' -DTORCH_API_INCLUDE_EXTENSION_H '-DPYBIND11_COMPILER_TYPE="_gcc"' '-DPYBIND11_STDLIB="_libstdcpp"' '-DPYBIND11_BUILD_ABI="_cxxabi1013"' -DTORCH_EXTENSION_NAME=pointnet2_stack_cuda -D_GLIBCXX_USE_CXX11_ABI=1 -gencode=arch=compute_87,code=compute_87 -gencode=arch=compute_87,code=sm_87 -std=c++14 ninja: build stopped: subcommand failed. Traceback (most recent call last): File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/utils/cpp_extension.py", line 1740, in _run_ninja_build subprocess.run( File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/subprocess.py", line 516, in run raise CalledProcessError(retcode, process.args, subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1. The above exception was the direct cause of the following exception: Traceback (most recent call last): File "setup.py", line 34, in <module> setup( File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/__init__.py", line 117, in setup return distutils.core.setup(**attrs) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/core.py", line 183, in setup return run_commands(dist) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/core.py", line 199, in run_commands dist.run_commands() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/dist.py", line 954, in run_commands self.run_command(cmd) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/dist.py", line 999, in run_command super().run_command(command) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/dist.py", line 973, in run_command cmd_obj.run() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/command/develop.py", line 35, in run self.install_for_development() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/command/develop.py", line 112, in install_for_development self.run_command('build_ext') File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/cmd.py", line 316, in run_command self.distribution.run_command(command) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/dist.py", line 999, in run_command super().run_command(command) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/dist.py", line 973, in run_command cmd_obj.run() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/command/build_ext.py", line 98, in run _build_ext.run(self) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/command/build_ext.py", line 359, in run self.build_extensions() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/utils/cpp_extension.py", line 741, in build_extensions build_ext.build_extensions(self) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/command/build_ext.py", line 476, in build_extensions self._build_extensions_serial() File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/command/build_ext.py", line 502, in _build_extensions_serial self.build_extension(ext) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/command/build_ext.py", line 263, in build_extension _build_ext.build_extension(self, ext) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/Cython/Distutils/build_ext.py", line 136, in build_extension super().build_extension(ext) File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/setuptools/_distutils/command/build_ext.py", line 557, in build_extension objects = self.compiler.compile( File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/utils/cpp_extension.py", line 562, in unix_wrap_ninja_compile _write_ninja_file_and_compile_objects( File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/utils/cpp_extension.py", line 1419, in _write_ninja_file_and_compile_objects _run_ninja_build( File "/home/autodriver/miniconda3/envs/openpcdet/lib/python3.8/site-packages/torch/utils/cpp_extension.py", line 1756, in _run_ninja_build raise RuntimeError(message) from e RuntimeError: Error compiling objects for extension

--------------------------------------------------------------------------- ProcessExitedException Traceback (most recent call last) Cell In[11], line 37 34 if __name__ == '__main__': 35 # 启动多进程训练 36 world_size = torch.cuda.device_count() # 使用所有可用GPU ---> 37 mp.spawn( 38 train, # 目标函数 39 args=(world_size,), # 传递给train的参数 40 nprocs=world_size, # 进程数=GPU数量 41 join=True # 等待所有进程结束 42 ) File ~/miniconda3/envs/dev0.4/lib/python3.8/site-packages/torch/multiprocessing/spawn.py:240, in spawn(fn, args, nprocs, join, daemon, start_method) 236 msg = ('This method only supports start_method=spawn (got: %s).\n' 237 'To use a different start_method use:\n\t\t' 238 ' torch.multiprocessing.start_processes(...)' % start_method) 239 warnings.warn(msg) --> 240 return start_processes(fn, args, nprocs, join, daemon, start_method='spawn') File ~/miniconda3/envs/dev0.4/lib/python3.8/site-packages/torch/multiprocessing/spawn.py:198, in start_processes(fn, args, nprocs, join, daemon, start_method) 195 return context 197 # Loop on join until it returns True or raises an exception. --> 198 while not context.join(): 199 pass File ~/miniconda3/envs/dev0.4/lib/python3.8/site-packages/torch/multiprocessing/spawn.py:149, in ProcessContext.join(self, timeout) 140 raise ProcessExitedException( 141 "process %d terminated with signal %s" % 142 (error_index, name), (...) 146 signal_name=name 147 ) 148 else: --> 149 raise ProcessExitedException( 150 "process %d terminated with exit code %d" % 151 (error_index, exitcode), 152 error_index=error_index, 153 error_pid=failed_process.pid, 154 exit_code=exitcode 155 ) 157 original_trace = self.error_queues[error_index].get() 158 msg = "\n\n-- Process %d terminated with the following error:\n" % error_index ProcessExitedException: process 2 terminated with exit code 1

最新推荐

recommend-type

collect2: fatal error: ld terminated with signal 11 [Segmentation fault], core d

在编程过程中,我们时常会遇到各种编译错误,其中一种比较棘手的错误是"collect2: fatal error: ld terminated with signal 11 [Segmentation fault], core dumped"。这个错误通常发生在尝试使用g++或gcc编译C++或C...
recommend-type

基于YOLO检测器的高效、与布局无关的自动车牌识别系统的自动ALPR推理流程_Automating ALPR infer

基于YOLO检测器的高效、与布局无关的自动车牌识别系统的自动ALPR推理流程_Automating ALPR inference flow of “An Efficient and Layout-Independent Automatic License Plate Recognition System Based on the YOLO Detector”.zip
recommend-type

清华大学袁春与华为李航合著的统计学习课件

统计学习方法是机器学习领域的重要分支,它将统计学的原理和方法应用在数据分析和模型构建上,尤其适用于解决那些不确定性和复杂性的数据问题。在本次分享的课件中,袁春和李航作为来自清华大学深圳研究生院和华为诺亚方舟实验室的专家,将为我们展示统计学习方法的理论与实践。 课件内容可能涵盖了以下几个主要知识点: 1. 统计学习的基本概念:首先,课件可能会介绍统计学习的定义,包括它是如何从统计学中独立出来,并与机器学习相结合,形成一套独立的理论体系的。同时,解释统计学习的核心思想,即利用数据来发现知识,并构建预测模型。 2. 常见统计学习模型:课程内容可能会包括线性回归、逻辑回归、决策树、随机森林、支持向量机(SVM)、神经网络等模型,这些模型是统计学习方法中经常使用到的。对于每一种模型,课件可能会详细讲解其数学原理、优缺点、适用场景以及如何在实际中应用。 3. 模型评估与选择:统计学习中的模型评估是确保模型性能和泛化能力的关键。课件可能会介绍交叉验证、AIC、BIC、ROC曲线、混淆矩阵等评估指标和方法。此外,还会讲解如何根据业务需求和评估结果选择合适的模型。 4. 正则化与模型优化:为防止过拟合,提升模型的泛化能力,统计学习方法中常常使用正则化技术。课件可能会涉及L1和L2正则化、岭回归(Ridge Regression)、套索回归(Lasso Regression)等技术,并展示如何通过正则化调整模型复杂度。 5. 统计学习在机器学习中的应用:统计学习方法不仅仅是一个理论体系,它在实际中也有广泛应用。例如,金融风险评估、生物信息学、推荐系统、自然语言处理等领域,课件可能会选取一些实际案例来说明统计学习方法的应用。 6. 统计学习前沿发展:课件可能会介绍统计学习领域的最新研究动态,包括最新算法的提出、统计学习与深度学习的结合等。 7. 课件中的实践操作:通常在理论学习之后,都会有实际操作环节,这可能包括使用R、Python等数据分析工具进行数据处理和模型构建。通过实际数据操作,学生可以更加深刻地理解统计学习方法的应用。 8. 与传统统计方法的区别:尽管统计学习方法和传统统计方法在很多方面是相通的,但它们在处理大数据、复杂数据结构以及预测模型构建方面存在明显差异。课件可能会明确指出这些差异,并突出统计学习方法的优势。 以上内容是根据标题、描述、标签以及文件名称列表推测的课件可能包含的知识点。学习这些内容,不仅能帮助学生掌握统计学习方法,而且能为学生在未来从事相关领域的工作和研究打下坚实的基础。
recommend-type

概率空间中的监督分类与文档分类方法探索

### 概率空间中的监督分类与文档分类方法探索 #### 1. 监督分类概述 在文档分类中,监督分类方法具有重要地位。这里主要聚焦于统计方法中的似然比方法。该方法通过不同类别依赖的统计模型来估计给定文档的概率,并利用这些概率的比率确定文档最可能所属的类别。 #### 2. 似然比框架 对于一个给定数据集划分为两个互斥类别A和B的情况,根据贝叶斯规则,给定观察文档D时每个类别的发生概率可表示为: - \(p(A|D) = p(D|A)× p(A)/p(D)\) (11.32a) - \(p(B|D) = p(D|B)× p(B)/p(D)\) (11.32b) 两式相除可得: \(p(
recommend-type

永磁同步电机矢量控制PID闭环函数

在永磁同步电机矢量控制中,PID(比例 - 积分 - 微分)控制器是常用的闭环控制方法,用于调节电机的各种参数,如电流、速度和位置等。以下是不同控制环中PID闭环函数的详细介绍: ### 电流环PID闭环函数 电流环主要用于控制电机的定子电流,通常分为直轴($d$轴)和交轴($q$轴)电流控制。PID控制器的输出是电压指令,用于控制逆变器的输出电压。 PID控制器的一般公式为: $$u(t) = K_p e(t) + K_i \int_{0}^{t} e(\tau) d\tau + K_d \frac{de(t)}{dt}$$ 在离散系统中,PID控制器的公式可以表示为: $$u(k)
recommend-type

GitHub新手入门:创建并发布您的第一个网站

标题所指的"emvillanueva.github.io"是一个网站地址,该地址使用GitHub Pages服务创建,它是GitHub提供的一项功能,允许开发者直接通过GitHub托管和发布静态网站。这一服务特别受那些希望简化网站部署流程的开发者欢迎,因为它无需传统服务器配置即可将网站上线。 从描述中,我们可以了解到几个关键知识点: 1. GitHub是一个为数百万开发人员共用的代码托管平台,支持版本控制和协作。它广泛用于开源项目,同时也有私有项目的服务,为开发人员提供了代码仓库、问题追踪、代码审查以及一系列其他功能。 2. GitHub Pages是GitHub提供的静态网站托管服务,允许用户通过特定的仓库来发布网站。用户可以通过简单地推送HTML、CSS、JavaScript文件到指定的分支(通常是master或main分支),而GitHub Pages将会自动发布这些文件到一个由GitHub分配的URL上。 3. 在GitHub中,"仓库"(repository)相当于项目的虚拟文件夹或容器,它包含项目的全部文件,包括源代码、文档、图像、数据等。此外,仓库还负责记录所有文件的更改历史,以及对这些文件进行版本控制。这种变更追踪机制意味着开发人员可以回滚到任何之前的状态,便于错误修复和版本迭代。 4. 在上述描述中,提到的一个项目仓库里包含三个关键文件:HTML文件、CSS文件和README文件。这些文件共同构成一个基本的网站结构: - HTML文件是网页的骨架,负责定义网页的结构和内容; - CSS文件负责网页的样式,包括颜色、字体以及其他视觉表现形式; - README文件通常用来描述项目的信息,例如项目的目的、使用说明和作者信息等。它通常以纯文本格式编写,但也可以用Markdown格式,以便于排版和展现更丰富的文档信息。 5. 描述中还提到了“JavaScript”,这是一种广泛应用于网页开发的编程语言,负责实现网页上的交互性和动态效果。虽然在这个上下文中并未明确指出JavaScript文件,但考虑到一个功能齐全的网站通常需要JavaScript来增加其功能性,因此可以推测仓库中可能还包含一个或多个JavaScript文件。 标签“JavaScript”强调了这一语言在现代网页开发中的重要性。它常用于网页的客户端脚本,可以操作文档对象模型(DOM),处理用户输入,制作动画效果,发送和接受服务器数据等。 压缩包子文件的文件名称列表中提到的"emvillanueva.github.io-master",意味着存在一个GitHub项目仓库的压缩包文件。通常,这种文件可以用来备份或传输仓库内容。在GitHub上,"master"或"main"分支是默认的主分支,存放着项目的最新稳定代码。 总结上述知识点,我们可以了解GitHub Pages服务的使用方法和优势、仓库的概念、文件类型对网站构成的重要性以及JavaScript在网页开发中的核心作用。这对于任何希望开始使用GitHub进行项目协作和网站发布的IT专业人士或爱好者都是非常重要的基础概念。
recommend-type

电影评论内容的极性分析与信息提取

### 电影评论内容的极性分析与信息提取 #### 1. 极性分析结论与方法概述 从相关结果中我们可以得出一些重要结论。首先,利用基于词频和分布的数学模型,特别是这里使用的向量空间模型方法,确实能够区分有观点内容的极性。而且,模型生成的面向领域和面向极性的表示维度的判别能力似乎相互独立。此外,结果还表明,至少对于本节所考虑的特定类型文本数据,极性判别比领域判别更具挑战性。 在极性估计方面,分析有观点内容的重要任务包括极性检测、强度估计、主观性检测和情感测量。这里我们主要关注前两项任务。极性检测是识别正负两个极性类别,而强度估计则是衡量给定内容的正负程度。从统计角度来看,我们采用基于似然比方
recommend-type

STM32F103C8T6双串口透传

虽然给定引用中未直接提及STM32F103C8T6实现双串口透传的方法,但可结合其串口通信相关内容推测大致实现思路。 STM32F103C8T6有3个USART,可利用其中两个串口来实现双串口透传。其基本方法如下: 1. **硬件连接**:连接两个外部设备到STM32F103C8T6的两个不同串口,比如一个设备连接到USART1,另一个连接到USART2。同时要保证电源、地等连接正确。 2. **工程配置**:可使用STM32CubeMx工具对工程进行配置。开启需要使用的两个串口,并设置好波特率、数据位、停止位等参数。 3. **代码实现**:在代码中,利用HAL库函数实现串口数据的接收和发
recommend-type

DevSENAI项目深度解析与技术指南

由于提供的文件信息中【标题】和【描述】都是"DevSENAI",且【标签】为空,【压缩包子文件的文件名称列表】中只有一个元素"DevSENAI-main",这些信息对于生成知识点来说非常有限。由于没有具体的上下文,我将假设"DevSENAI"是一个特定的软件开发项目名称,并且"DevSENAI-main"是该项目的主分支或主文件夹。以下是对“DevSENAI”项目可能涉及的知识点的详细说明: 1. 软件开发流程:DevSENAI项目可能遵循传统的软件开发生命周期,包括需求收集、分析、设计、实现、测试、部署和维护等阶段。每个阶段都有其特定的工作和产出,如需求规格说明书、设计文档、代码、用户手册和维护计划。 2. 版本控制系统:从文件名称"DevSENAI-main"可以推断,该项目可能使用了版本控制系统,如Git。"main"通常是Git中默认的主分支名称,它用于存放当前稳定版本的代码。版本控制对于团队协作开发项目是必不可少的,它可以帮助开发者追踪代码变更历史,合并代码,以及协作解决问题。 3. 项目管理方法论:DevSENAI可能采用了一系列项目管理实践来确保项目的顺利进行。这可能包括敏捷开发方法,比如Scrum或Kanban,以快速迭代的方式开发和交付软件。项目管理也涉及资源分配、进度跟踪、风险管理以及与利益相关者的沟通。 4. 软件工程最佳实践:为确保软件质量和项目效率,DevSENAI项目可能会采用各种软件工程的最佳实践。这可能包括代码审查、持续集成(CI)、测试驱动开发(TDD)、以及重构等技术。这些实践有助于保证代码的可维护性和项目的可持续发展。 5. 技术栈与开发工具:DevSENAI项目所涉及的技术栈将决定它使用哪些编程语言、框架、数据库、开发环境和部署工具。例如,项目可能使用Java作为后端开发语言、React作为前端框架、MySQL作为数据库,以及Docker容器化技术和Kubernetes作为容器编排工具。 6. 质量保证和测试:在"DevSENAI"项目中,质量保证是一个核心环节。这可能包括单元测试、集成测试、性能测试、安全测试等,确保软件满足既定的功能和非功能需求。自动化测试工具如Selenium、JUnit、Mocha等可能是项目中的重要组成部分。 7. 文档和知识管理:任何软件项目都需要相应的文档来记录其开发过程和最终产品。文档可能包括API文档、系统设计文档、用户手册、操作指南等。在DevSENAI项目中,有效的知识管理还可能涉及到团队内部知识共享的实践和平台。 8. 部署和运维:项目开发完成之后,还需要考虑软件的部署和运维工作。这可能包括持续部署(CD)、监控、日志记录、错误报告和恢复计划等。DevOps文化可能在DevSENAI项目中得到实践,以提高交付效率和运维效能。 9. 用户界面和体验:对于面向终端用户的软件,用户界面(UI)和用户体验(UX)设计变得尤为重要。在DevSENAI项目中,可能需要专业的UI/UX设计师参与,以确保软件产品既美观又易于使用。 10. 数据安全与合规性:在开发过程中,需要考虑到数据安全和符合相关法律法规的要求。DevSENAI项目可能需要实现安全措施,如数据加密、安全认证、访问控制以及遵循如GDPR或HIPAA等数据保护法规。 由于文件提供的信息非常有限,以上知识点是基于一般性的假设。实际的DevSENAI项目可能涉及更多具体的技术细节和专业知识,这些将需要更详尽的项目文档和信息来进行深入探讨。
recommend-type

文本关键词提取与摘要生成技术解析

### 文本关键词提取与摘要生成技术解析 #### 1. 角色重要性分析 在相关文本分析中,对角色的重要性进行了评估,以下是部分角色的相关数据: | 角色 | 出现次数 | 占比 | 相关指标值 | 另一指标值 | | ---- | ---- | ---- | ---- | ---- | | {'Sancho'} | 13 | 0.052632 | 17.369 | 0.11424 | | {'Christian'} | 9 | 0.043478 | 14.905 | 0.083947 | | {'Pedro'} | 3 | 0.032258 | 0 | 0.031398 | | {'Gin