Need help building Pytorch from source

Hello everyone,

I set USE_CUDA=1 in setup.py and successfully build Pytorch from source, but I got this error:

AssertionError: Torch not compiled with CUDA enabled

How can I fix it?
Here is my CUDA info:

$ nvidia-smi
Thu Aug 31 19:44:49 2023       
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 470.199.02   Driver Version: 470.199.02   CUDA Version: 11.4     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  Quadro K600         Off  | 00000000:11:00.0  On |                  N/A |
| 29%   58C    P0    N/A /  N/A |    144MiB /   981MiB |      2%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
                                                                               
+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|    0   N/A  N/A      3268      G   /usr/libexec/Xorg                 108MiB |
|    0   N/A  N/A      3395      G   /usr/bin/gnome-shell               28MiB |
|    0   N/A  N/A     11847      G   /usr/lib64/firefox/firefox          2MiB |
+-----------------------------------------------------------------------------+

Check the build logs and make sure your CUDA compiler etc. was properly detected.

How to get the build logs?

During the build process while you are executing python setup.py install the process will print logs to your terminal pointing to your setup, which CUDA toolkit and host compiler were detected etc.

There is some info in the log:

-- Not using cuDNN
-- Not using CUDA

In that case no CUDA toolkit was detected, which should have been shown in the logs as well.

This is the full log:

(base) notooth@localhost:pytorch$ python setup.py develop
Building wheel torch-2.1.0a0+gitfe1f26a
-- Building version 2.1.0a0+gitfe1f26a
cmake --build . --target install --config Release
[0/1] Install the project...
-- Install configuration: "Release"
running develop
/home/notooth/miniconda3/lib/python3.11/site-packages/setuptools/command/develop.py:40: EasyInstallDeprecationWarning: easy_install command is deprecated.
!!

        ********************************************************************************
        Please avoid running ``setup.py`` and ``easy_install``.
        Instead, use pypa/build, pypa/installer, pypa/build or
        other standards-based tools.

        See https://github.com/pypa/setuptools/issues/917 for details.
        ********************************************************************************

!!
  easy_install.initialize_options(self)
/home/notooth/miniconda3/lib/python3.11/site-packages/setuptools/_distutils/cmd.py:66: SetuptoolsDeprecationWarning: setup.py install is deprecated.
!!

        ********************************************************************************
        Please avoid running ``setup.py`` directly.
        Instead, use pypa/build, pypa/installer, pypa/build or
        other standards-based tools.

        See https://blog.ganssle.io/articles/2021/10/setup-py-deprecated.html for details.
        ********************************************************************************

!!
  self.initialize_options()
running egg_info
writing torch.egg-info/PKG-INFO
writing dependency_links to torch.egg-info/dependency_links.txt
writing entry points to torch.egg-info/entry_points.txt
writing requirements to torch.egg-info/requires.txt
writing top-level names to torch.egg-info/top_level.txt
reading manifest file 'torch.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
warning: no previously-included files matching '*.o' found anywhere in distribution
warning: no previously-included files matching '*.dylib' found anywhere in distribution
warning: no previously-included files matching '*.swp' found anywhere in distribution
adding license file 'LICENSE'
adding license file 'NOTICE'
writing manifest file 'torch.egg-info/SOURCES.txt'
running build_ext
-- Building with NumPy bindings
-- Not using cuDNN
-- Not using CUDA
-- Using MKLDNN
-- Not using Compute Library for the Arm architecture with MKLDNN
-- Not using CBLAS in MKLDNN
-- Not using NCCL
-- Building with distributed package: 
  -- USE_TENSORPIPE=True
  -- USE_GLOO=True
  -- USE_MPI=False
-- Building Executorch
-- Using ITT
-- Not Building nvfuser
Copying functorch._C from functorch/functorch.so to /mnt/Archive/Downloads/1/pytorch/build/lib.linux-x86_64-cpython-311/functorch/_C.cpython-311-x86_64-linux-gnu.so
Creating /home/notooth/miniconda3/lib/python3.11/site-packages/torch.egg-link (link to .)
torch 2.1.0a0+gitfe1f26a is already the active version in easy-install.pth
Installing convert-caffe2-to-onnx script to /home/notooth/miniconda3/bin
Installing convert-onnx-to-caffe2 script to /home/notooth/miniconda3/bin
Installing torchrun script to /home/notooth/miniconda3/bin

Installed /mnt/Archive/Downloads/1/pytorch
Processing dependencies for torch==2.1.0a0+gitfe1f26a
Searching for fsspec==2023.6.0
Best match: fsspec 2023.6.0
Adding fsspec 2023.6.0 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for Jinja2==3.1.2
Best match: Jinja2 3.1.2
Adding Jinja2 3.1.2 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for networkx==3.1
Best match: networkx 3.1
Adding networkx 3.1 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for sympy==1.12
Best match: sympy 1.12
Adding sympy 1.12 to easy-install.pth file
Installing isympy script to /home/notooth/miniconda3/bin

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for typing-extensions==4.7.1
Best match: typing-extensions 4.7.1
Adding typing-extensions 4.7.1 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for filelock==3.12.3
Best match: filelock 3.12.3
Adding filelock 3.12.3 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for MarkupSafe==2.1.3
Best match: MarkupSafe 2.1.3
Adding MarkupSafe 2.1.3 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Searching for mpmath==1.3.0
Best match: mpmath 1.3.0
Adding mpmath 1.3.0 to easy-install.pth file

Using /home/notooth/miniconda3/lib/python3.11/site-packages
Finished processing dependencies for torch==2.1.0a0+gitfe1f26a```

That’s not the full build log as nothing is being compiled so I guess you are just rerunning the build command in the source folder which already contains the previous build?

What is the compile command?

As previously mentioned:

This is the log while executing python setup.py install

(base) notooth@localhost:pytorch$ python setup.py install
Building wheel torch-2.1.0a0+gitfe1f26a
-- Building version 2.1.0a0+gitfe1f26a
cmake --build . --target install --config Release
[0/1] Install the project...
-- Install configuration: "Release"
running install
/home/notooth/miniconda3/lib/python3.11/site-packages/setuptools/_distutils/cmd.py:66: SetuptoolsDeprecationWarning: setup.py install is deprecated.
!!

        ********************************************************************************
        Please avoid running ``setup.py`` directly.
        Instead, use pypa/build, pypa/installer, pypa/build or
        other standards-based tools.

        See https://blog.ganssle.io/articles/2021/10/setup-py-deprecated.html for details.
        ********************************************************************************

!!
  self.initialize_options()
running build
running build_py
copying torchgen/packaged/ATen/native/native_functions.yaml -> build/lib.linux-x86_64-cpython-311/torchgen/packaged/ATen/native
copying torchgen/packaged/ATen/native/tags.yaml -> build/lib.linux-x86_64-cpython-311/torchgen/packaged/ATen/native
running build_ext
-- Building with NumPy bindings
-- Not using cuDNN
-- Not using CUDA
-- Using MKLDNN
-- Not using Compute Library for the Arm architecture with MKLDNN
-- Not using CBLAS in MKLDNN
-- Not using NCCL
-- Building with distributed package: 
  -- USE_TENSORPIPE=True
  -- USE_GLOO=True
  -- USE_MPI=False
-- Building Executorch
-- Using ITT
-- Not Building nvfuser
Copying functorch._C from functorch/functorch.so to /mnt/Archive/Downloads/1/pytorch/build/lib.linux-x86_64-cpython-311/functorch/_C.cpython-311-x86_64-linux-gnu.so
running install_lib
copying build/lib.linux-x86_64-cpython-311/torchgen/packaged/ATen/native/native_functions.yaml -> /home/notooth/miniconda3/lib/python3.11/site-packages/torchgen/packaged/ATen/native
copying build/lib.linux-x86_64-cpython-311/torchgen/packaged/ATen/native/tags.yaml -> /home/notooth/miniconda3/lib/python3.11/site-packages/torchgen/packaged/ATen/native
running install_egg_info
running egg_info
writing torch.egg-info/PKG-INFO
writing dependency_links to torch.egg-info/dependency_links.txt
writing entry points to torch.egg-info/entry_points.txt
writing requirements to torch.egg-info/requires.txt
writing top-level names to torch.egg-info/top_level.txt
reading manifest file 'torch.egg-info/SOURCES.txt'
reading manifest template 'MANIFEST.in'
warning: no previously-included files matching '*.o' found anywhere in distribution
warning: no previously-included files matching '*.dylib' found anywhere in distribution
warning: no previously-included files matching '*.swp' found anywhere in distribution
adding license file 'LICENSE'
adding license file 'NOTICE'
writing manifest file 'torch.egg-info/SOURCES.txt'
removing '/home/notooth/miniconda3/lib/python3.11/site-packages/torch-2.1.0a0+gitfe1f26a-py3.11.egg-info' (and everything under it)
Copying torch.egg-info to /home/notooth/miniconda3/lib/python3.11/site-packages/torch-2.1.0a0+gitfe1f26a-py3.11.egg-info
running install_scripts
Installing convert-caffe2-to-onnx script to /home/notooth/miniconda3/bin
Installing convert-onnx-to-caffe2 script to /home/notooth/miniconda3/bin
Installing torchrun script to /home/notooth/miniconda3/bin

You are most likely still using the build cache or your terminal swallows the full output. Note that a full rebuild will take more than a few seconds.
In any case, the full log isn’t needed since it already shows that no CUDA toolkit was found, so make sure your CUDA toolkit is properly installed and you are able to compile CUDA code.

I already have CUDA 11.4:

$ nvidia-smi
Thu Aug 31 19:44:49 2023       
+-----------------------------------------------------------------------------+
| NVIDIA-SMI 470.199.02   Driver Version: 470.199.02   CUDA Version: 11.4     |
|-------------------------------+----------------------+----------------------+
| GPU  Name        Persistence-M| Bus-Id        Disp.A | Volatile Uncorr. ECC |
| Fan  Temp  Perf  Pwr:Usage/Cap|         Memory-Usage | GPU-Util  Compute M. |
|                               |                      |               MIG M. |
|===============================+======================+======================|
|   0  Quadro K600         Off  | 00000000:11:00.0  On |                  N/A |
| 29%   58C    P0    N/A /  N/A |    144MiB /   981MiB |      2%      Default |
|                               |                      |                  N/A |
+-------------------------------+----------------------+----------------------+
                                                                               
+-----------------------------------------------------------------------------+
| Processes:                                                                  |
|  GPU   GI   CI        PID   Type   Process name                  GPU Memory |
|        ID   ID                                                   Usage      |
|=============================================================================|
|    0   N/A  N/A      3268      G   /usr/libexec/Xorg                 108MiB |
|    0   N/A  N/A      3395      G   /usr/bin/gnome-shell               28MiB |
|    0   N/A  N/A     11847      G   /usr/lib64/firefox/firefox          2MiB |
+-----------------------------------------------------------------------------+

nvidia-smi does not show if the full CUDA toolkit including the compiler is installed, but shows the driver corresponding to a CUDA release version. Check the cuda-samples and make sure you are able to build and execute them.

I am not able to build the cuda-samples:

(base) notooth@localhost:cuda-samples$ make
...
simplePitchLinearTexture.cu:179:103: warning: ‘cudaError_t cudaMemcpyToArray(cudaArray_t, size_t, size_t, const void*, size_t, cudaMemcpyKind)’ is deprecated [-Wdeprecated-declarations]
   checkCudaErrors(cudaMemcpyToArray(d_idataArray, 0, 0, h_idata,
                                                                                                       ^
reduction_kernel.cu(557): error: namespace "cooperative_groups" has no member "block_tile_memory"
          detected during instantiation of "void reduce(int, int, int, int, T *, T *) [with T=int]" 
(1028): here

reduction_kernel.cu(560): error: too many arguments in function call
          detected during instantiation of "void reduce(int, int, int, int, T *, T *) [with T=int]" 
(1028): here

/usr/local/cuda/bin/../targets/x86_64-linux/include/cooperative_groups.h(1701): error: static assertion failed with "Tiled partition with Size > 32 is supported only by cooperative_groups::experimental::tiled_partition available with experimental features enabled"
          detected during:
            instantiation of "cooperative_groups::__v1::thread_block_tile<Size, ParentT> cooperative_groups::__v1::tiled_partition<Size,ParentT>(const ParentT &) [with Size=512U, ParentT=cooperative_groups::__v1::thread_block]" 
reduction_kernel.cu(562): here
            instantiation of "void multi_warp_cg_reduce<T,BlockSize,MultiWarpGroupSize>(T *, T *, unsigned int) [with T=int, BlockSize=1024UL, MultiWarpGroupSize=512UL]" 
reduction_kernel.cu(996): here
            instantiation of "void reduce(int, int, int, int, T *, T *) [with T=int]" 
reduction_kernel.cu(1028): here

Hi,
I think you’ll also need to download the CUDA toolkit (for runtime APIs) from here; please select the version you want to download. After downloading, you’ll need to set the path. You can use the following commands (for cuda-12.2):

export PATH=/usr/local/cuda-12.2/bin/:$PATH

and

export LD_LIBRARY_PATH=/usr/local/cuda-12.2/lib64:$LD_LIBRARY_PATH

You can check CUDA compatibility with the CUDA driver from here.
I hope this helps to solve your problem. Thanks!

After installing cuda-12.2 and setting the PATH, I successfully built cuda-samples. But I still got -- Not using CUDA when building Pytorch.

As @ ptrblck mentioned earlier, I think it’s still using the previous build cache. Please run:

python setup.py clean

Set up the path again to confirm if the CUDA path is correct, and check what is the result of:

nvcc --version

Then rebuild pytorch again (follow all steps).

This is the log when I rebuild Pytorch:

(base) notooth@localhost:pytorch$ nvcc --version
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Tue_Aug_15_22:02:13_PDT_2023
Cuda compilation tools, release 12.2, V12.2.140
Build cuda_12.2.r12.2/compiler.33191640_0

(base) notooth@localhost:pytorch$ python setup.py clean
Building wheel torch-2.1.0a0+gitfe1f26a
running clean

(base) notooth@localhost:pytorch$ python setup.py install
Building wheel torch-2.1.0a0+gitfe1f26a
-- Building version 2.1.0a0+gitfe1f26a
cmake -GNinja -DBUILD_PYTHON=True -DBUILD_TEST=True -DCMAKE_BUILD_TYPE=Release -DCMAKE_INSTALL_PREFIX=/mnt/Archive/Downloads/1/pytorch/torch -DCMAKE_PREFIX_PATH=/home/notooth/miniconda3/lib/python3.11/site-packages -DNUMPY_INCLUDE_DIR=/home/notooth/miniconda3/lib/python3.11/site-packages/numpy/core/include -DPYTHON_EXECUTABLE=/home/notooth/miniconda3/bin/python -DPYTHON_INCLUDE_DIR=/home/notooth/miniconda3/include/python3.11 -DPYTHON_LIBRARY=/home/notooth/miniconda3/lib/libpython3.11.a -DTORCH_BUILD_VERSION=2.1.0a0+gitfe1f26a -DUSE_NUMPY=True /mnt/Archive/Downloads/1/pytorch
-- The CXX compiler identification is GNU 8.5.0
-- The C compiler identification is GNU 8.5.0
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- /usr/bin/c++ /mnt/Archive/Downloads/1/pytorch/torch/abi-check.cpp -o /mnt/Archive/Downloads/1/pytorch/build/abi-check
-- Determined _GLIBCXX_USE_CXX11_ABI=1
-- Not forcing any particular BLAS to be found
-- Could not find ccache. Consider installing ccache to speed up compilation.
-- Performing Test C_HAS_AVX_1
-- Performing Test C_HAS_AVX_1 - Failed
-- Performing Test C_HAS_AVX_2
-- Performing Test C_HAS_AVX_2 - Failed
-- Performing Test C_HAS_AVX_3
-- Performing Test C_HAS_AVX_3 - Failed
-- Performing Test C_HAS_AVX2_1
-- Performing Test C_HAS_AVX2_1 - Failed
-- Performing Test C_HAS_AVX2_2
-- Performing Test C_HAS_AVX2_2 - Failed
-- Performing Test C_HAS_AVX2_3
-- Performing Test C_HAS_AVX2_3 - Failed
-- Performing Test C_HAS_AVX512_1
-- Performing Test C_HAS_AVX512_1 - Failed
-- Performing Test C_HAS_AVX512_2
-- Performing Test C_HAS_AVX512_2 - Failed
-- Performing Test C_HAS_AVX512_3
-- Performing Test C_HAS_AVX512_3 - Failed
-- Performing Test CXX_HAS_AVX_1
-- Performing Test CXX_HAS_AVX_1 - Failed
-- Performing Test CXX_HAS_AVX_2
-- Performing Test CXX_HAS_AVX_2 - Failed
-- Performing Test CXX_HAS_AVX_3
-- Performing Test CXX_HAS_AVX_3 - Failed
-- Performing Test CXX_HAS_AVX2_1
-- Performing Test CXX_HAS_AVX2_1 - Failed
-- Performing Test CXX_HAS_AVX2_2
-- Performing Test CXX_HAS_AVX2_2 - Failed
-- Performing Test CXX_HAS_AVX2_3
-- Performing Test CXX_HAS_AVX2_3 - Failed
-- Performing Test CXX_HAS_AVX512_1
-- Performing Test CXX_HAS_AVX512_1 - Failed
-- Performing Test CXX_HAS_AVX512_2
-- Performing Test CXX_HAS_AVX512_2 - Failed
-- Performing Test CXX_HAS_AVX512_3
-- Performing Test CXX_HAS_AVX512_3 - Failed
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS
-- Performing Test CAFFE2_COMPILER_SUPPORTS_AVX512_EXTENSIONS - Success
-- Current compiler supports avx512f extension. Will build fbgemm.
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_VISIBILITY - Success
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY
-- Performing Test COMPILER_SUPPORTS_HIDDEN_INLINE_VISIBILITY - Success
-- Performing Test COMPILER_SUPPORTS_RDYNAMIC
-- Performing Test COMPILER_SUPPORTS_RDYNAMIC - Success
-- Found CUDA: /usr/local/cuda-12.2 (found version "12.2") 
-- The CUDA compiler identification is NVIDIA 12.2.140
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: /usr/local/cuda-12.2/bin/nvcc - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- Found CUDAToolkit: /usr/local/cuda-12.2/include (found version "12.2.140") 
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Failed
-- Check if compiler accepts -pthread
-- Check if compiler accepts -pthread - yes
-- Found Threads: TRUE  
-- Caffe2: CUDA detected: 12.2
-- Caffe2: CUDA nvcc is: /usr/local/cuda-12.2/bin/nvcc
-- Caffe2: CUDA toolkit directory: /usr/local/cuda-12.2
-- Caffe2: Header version is: 12.2
-- /usr/local/cuda-12.2/lib64/libnvrtc.so shorthash is 000ca627
-- Could NOT find CUDNN (missing: CUDNN_LIBRARY_PATH CUDNN_INCLUDE_PATH) 
CMake Warning at cmake/public/cuda.cmake:251 (message):
  Cannot find cuDNN library.  Turning the option off
Call Stack (most recent call first):
  cmake/Dependencies.cmake:44 (include)
  CMakeLists.txt:722 (include)


-- Could NOT find CUSPARSELT (missing: CUSPARSELT_LIBRARY_PATH CUSPARSELT_INCLUDE_PATH) 
CMake Warning at cmake/public/cuda.cmake:276 (message):
  Cannot find cuSPARSELt library.  Turning the option off
Call Stack (most recent call first):
  cmake/Dependencies.cmake:44 (include)
  CMakeLists.txt:722 (include)


-- Automatic GPU detection failed. Building for common architectures.
-- Autodetected CUDA architecture(s): 3.5;5.0;8.0;8.6;8.9;9.0
-- Added CUDA NVCC flags for: -gencode;arch=compute_35,code=sm_35;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90
-- Building using own protobuf under third_party per request.
-- Use custom protobuf build.
-- 
-- 3.13.0.0
-- Performing Test protobuf_HAVE_BUILTIN_ATOMICS
-- Performing Test protobuf_HAVE_BUILTIN_ATOMICS - Success
-- Caffe2 protobuf include directory: $<BUILD_INTERFACE:/mnt/Archive/Downloads/1/pytorch/third_party/protobuf/src>$<INSTALL_INTERFACE:include>
-- Trying to find preferred BLAS backend of choice: MKL
-- MKL_THREADING = OMP
-- Looking for sys/types.h
-- Looking for sys/types.h - found
-- Looking for stdint.h
-- Looking for stdint.h - found
-- Looking for stddef.h
-- Looking for stddef.h - found
-- Check size of void*
-- Check size of void* - done
CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_C)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  cmake/Modules/FindMKL.cmake:239 (FIND_PACKAGE)
  cmake/Modules/FindMKL.cmake:334 (CHECK_ALL_LIBRARIES)
  cmake/Dependencies.cmake:212 (find_package)
  CMakeLists.txt:722 (include)
This warning is for project developers.  Use -Wno-dev to suppress it.

CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_CXX)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  cmake/Modules/FindMKL.cmake:239 (FIND_PACKAGE)
  cmake/Modules/FindMKL.cmake:334 (CHECK_ALL_LIBRARIES)
  cmake/Dependencies.cmake:212 (find_package)
  CMakeLists.txt:722 (include)
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Looking for cblas_sgemm
-- Looking for cblas_sgemm - found
-- Looking for cblas_gemm_bf16bf16f32
-- Looking for cblas_gemm_bf16bf16f32 - found
-- MKL libraries: /home/notooth/miniconda3/lib/libmkl_intel_lp64.so;/home/notooth/miniconda3/lib/libmkl_gnu_thread.so;/home/notooth/miniconda3/lib/libmkl_core.so;-fopenmp;/usr/lib64/libpthread.so;/usr/lib64/libm.so;/usr/lib64/libdl.so
-- MKL include directory: /home/notooth/miniconda3/include
-- MKL OpenMP type: GNU
-- MKL OpenMP library: -fopenmp
-- The ASM compiler identification is GNU
-- Found assembler: /usr/bin/cc
-- Brace yourself, we are building NNPACK
-- Performing Test NNPACK_ARCH_IS_X86_32
-- Performing Test NNPACK_ARCH_IS_X86_32 - Failed
-- Found PythonInterp: /home/notooth/miniconda3/bin/python (found version "3.11.4") 
-- NNPACK backend is x86-64
-- Found Python: /home/notooth/miniconda3/bin/python3.11 (found version "3.11.4") found components: Interpreter 
-- Failed to find LLVM FileCheck
-- Found Git: /usr/bin/git (found version "2.39.3") 
-- git version: v1.6.1 normalized to 1.6.1
-- Version: 1.6.1
-- Looking for shm_open in rt
-- Looking for shm_open in rt - found
-- Performing Test HAVE_CXX_FLAG_STD_CXX11
-- Performing Test HAVE_CXX_FLAG_STD_CXX11 - Success
-- Performing Test HAVE_CXX_FLAG_WALL
-- Performing Test HAVE_CXX_FLAG_WALL - Success
-- Performing Test HAVE_CXX_FLAG_WEXTRA
-- Performing Test HAVE_CXX_FLAG_WEXTRA - Success
-- Performing Test HAVE_CXX_FLAG_WSHADOW
-- Performing Test HAVE_CXX_FLAG_WSHADOW - Success
-- Performing Test HAVE_CXX_FLAG_WERROR
-- Performing Test HAVE_CXX_FLAG_WERROR - Success
-- Performing Test HAVE_CXX_FLAG_WSUGGEST_OVERRIDE
-- Performing Test HAVE_CXX_FLAG_WSUGGEST_OVERRIDE - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC
-- Performing Test HAVE_CXX_FLAG_PEDANTIC - Success
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS
-- Performing Test HAVE_CXX_FLAG_PEDANTIC_ERRORS - Success
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32
-- Performing Test HAVE_CXX_FLAG_WSHORTEN_64_TO_32 - Failed
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_FSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED_DECLARATIONS
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED_DECLARATIONS - Success
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED
-- Performing Test HAVE_CXX_FLAG_WNO_DEPRECATED - Success
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING
-- Performing Test HAVE_CXX_FLAG_WSTRICT_ALIASING - Success
-- Performing Test HAVE_CXX_FLAG_WD654
-- Performing Test HAVE_CXX_FLAG_WD654 - Failed
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY
-- Performing Test HAVE_CXX_FLAG_WTHREAD_SAFETY - Failed
-- Performing Test HAVE_CXX_FLAG_COVERAGE
-- Performing Test HAVE_CXX_FLAG_COVERAGE - Success
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX
-- Performing Test HAVE_STD_REGEX -- success
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX
-- Performing Test HAVE_GNU_POSIX_REGEX -- failed to compile
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX
-- Performing Test HAVE_POSIX_REGEX -- success
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK
-- Performing Test HAVE_STEADY_CLOCK -- success
-- Performing Test COMPILER_SUPPORTS_AVX512
-- Performing Test COMPILER_SUPPORTS_AVX512 - Success
CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_C)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  third_party/fbgemm/CMakeLists.txt:129 (find_package)
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Found OpenMP_C: -fopenmp (found version "4.5") 
CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_CXX)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  third_party/fbgemm/CMakeLists.txt:129 (find_package)
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Found OpenMP_CXX: -fopenmp (found version "4.5") 
-- Found OpenMP: TRUE (found version "4.5")  
CMake Warning at third_party/fbgemm/CMakeLists.txt:131 (message):
  OpenMP found! OpenMP_C_INCLUDE_DIRS =


CMake Warning at third_party/fbgemm/CMakeLists.txt:224 (message):
  ==========


CMake Warning at third_party/fbgemm/CMakeLists.txt:225 (message):
  CMAKE_BUILD_TYPE = Release


CMake Warning at third_party/fbgemm/CMakeLists.txt:226 (message):
  CMAKE_CXX_FLAGS_DEBUG is -g


CMake Warning at third_party/fbgemm/CMakeLists.txt:227 (message):
  CMAKE_CXX_FLAGS_RELEASE is -O3 -DNDEBUG


CMake Warning at third_party/fbgemm/CMakeLists.txt:228 (message):
  ==========


-- Performing Test __CxxFlag__fno_threadsafe_statics
-- Performing Test __CxxFlag__fno_threadsafe_statics - Success
-- Performing Test __CxxFlag__fno_semantic_interposition
-- Performing Test __CxxFlag__fno_semantic_interposition - Success
-- Performing Test __CxxFlag__fmerge_all_constants
-- Performing Test __CxxFlag__fmerge_all_constants - Success
-- Performing Test __CxxFlag__fno_enforce_eh_specs
-- Performing Test __CxxFlag__fno_enforce_eh_specs - Success
** AsmJit Summary **
   ASMJIT_DIR=/mnt/Archive/Downloads/1/pytorch/third_party/fbgemm/third_party/asmjit
   ASMJIT_TEST=FALSE
   ASMJIT_TARGET_TYPE=STATIC
   ASMJIT_DEPS=pthread;rt
   ASMJIT_LIBS=asmjit;pthread;rt
   ASMJIT_CFLAGS=-DASMJIT_STATIC
   ASMJIT_PRIVATE_CFLAGS=-Wall;-Wextra;-Wconversion;-fno-math-errno;-fno-threadsafe-statics;-fno-semantic-interposition;-DASMJIT_STATIC
   ASMJIT_PRIVATE_CFLAGS_DBG=
   ASMJIT_PRIVATE_CFLAGS_REL=-O2;-fmerge-all-constants;-fno-enforce-eh-specs
-- Could NOT find Numa (missing: Numa_INCLUDE_DIR Numa_LIBRARIES) 
CMake Warning at cmake/Dependencies.cmake:903 (message):
  Not compiling with NUMA.  Suppress this warning with -DUSE_NUMA=OFF
Call Stack (most recent call first):
  CMakeLists.txt:722 (include)


-- Using third party subdirectory Eigen.
-- Found PythonInterp: /home/notooth/miniconda3/bin/python (found suitable version "3.11.4", minimum required is "3.0") 
-- Found PythonLibs: /home/notooth/miniconda3/lib/libpython3.11.a (found suitable version "3.11.4", minimum required is "3.0") 
-- Using third_party/pybind11.
-- pybind11 include dirs: /mnt/Archive/Downloads/1/pytorch/cmake/../third_party/pybind11/include
-- Checking for module 'mpi-c'
--   Package 'mpi-c', required by 'virtual:world', not found
-- Could NOT find MPI_C (missing: MPI_C_LIB_NAMES MPI_C_HEADER_DIR MPI_C_WORKS) 
-- Checking for module 'mpi-cxx'
--   Package 'mpi-cxx', required by 'virtual:world', not found
-- Could NOT find MPI_CXX (missing: MPI_CXX_LIB_NAMES MPI_CXX_HEADER_DIR MPI_CXX_WORKS) 
-- Could NOT find MPI (missing: MPI_C_FOUND MPI_CXX_FOUND) 
CMake Warning at cmake/Dependencies.cmake:1185 (message):
  Not compiling with MPI.  Suppress this warning with -DUSE_MPI=OFF
Call Stack (most recent call first):
  CMakeLists.txt:722 (include)


-- Adding OpenMP CXX_FLAGS: -fopenmp
-- Will link against OpenMP libraries: /usr/lib/gcc/x86_64-redhat-linux/8/libgomp.so;/usr/lib64/libpthread.so
-- Disabling kernel asserts for ROCm
-- Automatic GPU detection failed. Building for common architectures.
-- Autodetected CUDA architecture(s): 3.5;5.0;8.0;8.6;8.9;9.0
CMake Warning at cmake/External/nccl.cmake:70 (message):
  Enabling NCCL library slimming
Call Stack (most recent call first):
  cmake/Dependencies.cmake:1348 (include)
  CMakeLists.txt:722 (include)


-- Found CUB: /usr/local/cuda-12.2/include  
-- Converting CMAKE_CUDA_FLAGS to CUDA_NVCC_FLAGS:
    CUDA_NVCC_FLAGS                = -D_GLIBCXX_USE_CXX11_ABI=1;-Xfatbin;-compress-all;-DONNX_NAMESPACE=onnx_torch;-gencode;arch=compute_35,code=sm_35;-gencode;arch=compute_50,code=sm_50;-gencode;arch=compute_80,code=sm_80;-gencode;arch=compute_86,code=sm_86;-gencode;arch=compute_89,code=sm_89;-gencode;arch=compute_90,code=sm_90;-Xcudafe;--diag_suppress=cc_clobber_ignored,--diag_suppress=set_but_not_used,--diag_suppress=field_without_dll_interface,--diag_suppress=base_class_has_different_dll_interface,--diag_suppress=dll_interface_conflict_none_assumed,--diag_suppress=dll_interface_conflict_dllexport_assumed,--diag_suppress=bad_friend_decl;--expt-relaxed-constexpr;--expt-extended-lambda
    CUDA_NVCC_FLAGS_DEBUG          = -g
    CUDA_NVCC_FLAGS_RELEASE        = -O3;-DNDEBUG
    CUDA_NVCC_FLAGS_RELWITHDEBINFO = -O2;-g;-DNDEBUG
    CUDA_NVCC_FLAGS_MINSIZEREL     = -O1;-DNDEBUG
-- Performing Test UV_LINT_W4
-- Performing Test UV_LINT_W4 - Failed
-- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC
-- Performing Test UV_LINT_NO_UNUSED_PARAMETER_MSVC - Failed
-- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC
-- Performing Test UV_LINT_NO_CONDITIONAL_CONSTANT_MSVC - Failed
-- Performing Test UV_LINT_NO_NONSTANDARD_MSVC
-- Performing Test UV_LINT_NO_NONSTANDARD_MSVC - Failed
-- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC
-- Performing Test UV_LINT_NO_NONSTANDARD_EMPTY_TU_MSVC - Failed
-- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC
-- Performing Test UV_LINT_NO_NONSTANDARD_FILE_SCOPE_MSVC - Failed
-- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC
-- Performing Test UV_LINT_NO_NONSTANDARD_NONSTATIC_DLIMPORT_MSVC - Failed
-- Performing Test UV_LINT_NO_HIDES_LOCAL
-- Performing Test UV_LINT_NO_HIDES_LOCAL - Failed
-- Performing Test UV_LINT_NO_HIDES_PARAM
-- Performing Test UV_LINT_NO_HIDES_PARAM - Failed
-- Performing Test UV_LINT_NO_HIDES_GLOBAL
-- Performing Test UV_LINT_NO_HIDES_GLOBAL - Failed
-- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC
-- Performing Test UV_LINT_NO_CONDITIONAL_ASSIGNMENT_MSVC - Failed
-- Performing Test UV_LINT_NO_UNSAFE_MSVC
-- Performing Test UV_LINT_NO_UNSAFE_MSVC - Failed
-- Performing Test UV_LINT_WALL
-- Performing Test UV_LINT_WALL - Success
-- Performing Test UV_LINT_NO_UNUSED_PARAMETER
-- Performing Test UV_LINT_NO_UNUSED_PARAMETER - Success
-- Performing Test UV_LINT_STRICT_PROTOTYPES
-- Performing Test UV_LINT_STRICT_PROTOTYPES - Success
-- Performing Test UV_LINT_EXTRA
-- Performing Test UV_LINT_EXTRA - Success
-- Performing Test UV_LINT_UTF8_MSVC
-- Performing Test UV_LINT_UTF8_MSVC - Failed
-- Performing Test UV_F_STRICT_ALIASING
-- Performing Test UV_F_STRICT_ALIASING - Success
-- summary of build options:
    Install prefix:  /mnt/Archive/Downloads/1/pytorch/torch
    Target system:   Linux
    Compiler:
      C compiler:    /usr/bin/cc
      CFLAGS:         

-- Found uv: 1.38.1 (found version "1.38.1") 
CMake Warning (dev) at third_party/gloo/CMakeLists.txt:21 (option):
  Policy CMP0077 is not set: option() honors normal variables.  Run "cmake
  --help-policy CMP0077" for policy details.  Use the cmake_policy command to
  set the policy and suppress this warning.

  For compatibility with older versions of CMake, option is clearing the
  normal variable 'BUILD_BENCHMARK'.
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Gloo build as SHARED library
CMake Warning (dev) at third_party/gloo/cmake/Cuda.cmake:109 (find_package):
  Policy CMP0074 is not set: find_package uses <PackageName>_ROOT variables.
  Run "cmake --help-policy CMP0074" for policy details.  Use the cmake_policy
  command to set the policy and suppress this warning.

  CMake variable CUDAToolkit_ROOT is set to:

    /usr/local/cuda-12.2

  For compatibility, CMake is ignoring the variable.
Call Stack (most recent call first):
  third_party/gloo/cmake/Dependencies.cmake:115 (include)
  third_party/gloo/CMakeLists.txt:111 (include)
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Found CUDAToolkit: /usr/local/cuda-12.2/include (found suitable version "12.2.140", minimum required is "7.0") 
-- CUDA detected: 12.2.140
CMake Warning at cmake/Dependencies.cmake:1492 (message):
  Metal is only used in ios builds.
Call Stack (most recent call first):
  CMakeLists.txt:722 (include)


-- Found PythonInterp: /home/notooth/miniconda3/bin/python (found version "3.11.4") 
Generated: /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_onnx_torch-ml.proto
Generated: /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-operators_onnx_torch-ml.proto
Generated: /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-data_onnx_torch.proto
-- 
-- ******** Summary ********
--   CMake version             : 3.26.4
--   CMake command             : /home/notooth/miniconda3/bin/cmake
--   System                    : Linux
--   C++ compiler              : /usr/bin/c++
--   C++ compiler version      : 8.5.0
--   CXX flags                 :  -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -Wnon-virtual-dtor
--   Build type                : Release
--   Compile definitions       : ONNX_ML=1;ONNXIFI_ENABLE_EXT=1;__STDC_FORMAT_MACROS
--   CMAKE_PREFIX_PATH         : /home/notooth/miniconda3/lib/python3.11/site-packages;/usr/local/cuda-12.2;/usr/local/cuda-12.2
--   CMAKE_INSTALL_PREFIX      : /mnt/Archive/Downloads/1/pytorch/torch
--   CMAKE_MODULE_PATH         : /mnt/Archive/Downloads/1/pytorch/cmake/Modules;/mnt/Archive/Downloads/1/pytorch/cmake/public/../Modules_CUDA_fix
-- 
--   ONNX version              : 1.14.1rc2
--   ONNX NAMESPACE            : onnx_torch
--   ONNX_USE_LITE_PROTO       : OFF
--   USE_PROTOBUF_SHARED_LIBS  : OFF
--   Protobuf_USE_STATIC_LIBS  : ON
--   ONNX_DISABLE_EXCEPTIONS   : OFF
--   ONNX_WERROR               : OFF
--   ONNX_BUILD_TESTS          : OFF
--   ONNX_BUILD_BENCHMARKS     : OFF
-- 
--   Protobuf compiler         : 
--   Protobuf includes         : 
--   Protobuf libraries        : 
--   BUILD_ONNX_PYTHON         : OFF
-- 
-- ******** Summary ********
--   CMake version         : 3.26.4
--   CMake command         : /home/notooth/miniconda3/bin/cmake
--   System                : Linux
--   C++ compiler          : /usr/bin/c++
--   C++ compiler version  : 8.5.0
--   CXX flags             :  -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -Wnon-virtual-dtor
--   Build type            : Release
--   Compile definitions   : ONNX_ML=1;ONNXIFI_ENABLE_EXT=1
--   CMAKE_PREFIX_PATH     : /home/notooth/miniconda3/lib/python3.11/site-packages;/usr/local/cuda-12.2;/usr/local/cuda-12.2
--   CMAKE_INSTALL_PREFIX  : /mnt/Archive/Downloads/1/pytorch/torch
--   CMAKE_MODULE_PATH     : /mnt/Archive/Downloads/1/pytorch/cmake/Modules;/mnt/Archive/Downloads/1/pytorch/cmake/public/../Modules_CUDA_fix
-- 
--   ONNX version          : 1.4.1
--   ONNX NAMESPACE        : onnx_torch
--   ONNX_BUILD_TESTS      : OFF
--   ONNX_BUILD_BENCHMARKS : OFF
--   ONNX_USE_LITE_PROTO   : OFF
--   ONNXIFI_DUMMY_BACKEND : 
-- 
--   Protobuf compiler     : 
--   Protobuf includes     : 
--   Protobuf libraries    : 
--   BUILD_ONNX_PYTHON     : OFF
-- Found CUDA with FP16 support, compiling with torch.cuda.HalfTensor
-- Adding -DNDEBUG to compile flags
-- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2
-- Checking prototype magma_get_sgeqrf_nb for MAGMA_V2 - False
-- Compiling with MAGMA support
-- MAGMA INCLUDE DIRECTORIES: /home/notooth/miniconda3/include
-- MAGMA LIBRARIES: /home/notooth/miniconda3/lib/libmagma.a
-- MAGMA V2 check: 0
-- Could not find hardware support for NEON on this machine.
-- No OMAP3 processor on this machine.
-- No OMAP4 processor on this machine.
-- Found a library with LAPACK API (mkl).
disabling ROCM because NOT USE_ROCM is set
-- MIOpen not found. Compiling without MIOpen support
-- Will build oneDNN Graph
-- MKLDNN_CPU_RUNTIME = OMP
-- DNNL_TARGET_ARCH: X64
-- DNNL_LIBRARY_NAME: dnnl
CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_C)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  third_party/ideep/mkl-dnn/cmake/OpenMP.cmake:69 (find_package)
  third_party/ideep/mkl-dnn/CMakeLists.txt:117 (include)
This warning is for project developers.  Use -Wno-dev to suppress it.

the log continued:

-- Found OpenMP_C: -fopenmp (found version "4.5") 
CMake Warning (dev) at /home/notooth/miniconda3/share/cmake-3.26/Modules/FindPackageHandleStandardArgs.cmake:438 (message):
  The package name passed to `find_package_handle_standard_args` (OpenMP_CXX)
  does not match the name of the calling package (OpenMP).  This can lead to
  problems in calling code that expects `find_package` result variables
  (e.g., `_FOUND`) to follow a certain pattern.
Call Stack (most recent call first):
  cmake/Modules/FindOpenMP.cmake:584 (find_package_handle_standard_args)
  third_party/ideep/mkl-dnn/cmake/OpenMP.cmake:69 (find_package)
  third_party/ideep/mkl-dnn/CMakeLists.txt:117 (include)
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Found OpenMP_CXX: -fopenmp (found version "4.5") 
-- Could NOT find Doxyrest (missing: DOXYREST_EXECUTABLE) 
-- Found PythonInterp: /home/notooth/miniconda3/bin/python (found suitable version "3.11.4", minimum required is "2.7") 
-- Could NOT find Sphinx (missing: SPHINX_EXECUTABLE) 
-- Enabled workload: TRAINING
-- Enabled primitives: ALL
-- Enabled primitive CPU ISA: ALL
-- Enabled primitive GPU ISA: ALL
-- Primitive cache is enabled
-- Graph component is enabled
-- Found MKL-DNN: TRUE
-- Looking for clock_gettime in rt
-- Looking for clock_gettime in rt - found
-- Looking for mmap
-- Looking for mmap - found
-- Looking for shm_open
-- Looking for shm_open - found
-- Looking for shm_unlink
-- Looking for shm_unlink - found
-- Looking for malloc_usable_size
-- Looking for malloc_usable_size - found
-- Performing Test C_HAS_THREAD
-- Performing Test C_HAS_THREAD - Success
-- <FindZVECTOR>
-- check z16
-- Performing Test COMPILE_OUT_z16
-- Performing Test COMPILE_OUT_z16 - Failed
-- check z15
-- Performing Test COMPILE_OUT_z15
-- Performing Test COMPILE_OUT_z15 - Failed
-- check z14
-- Performing Test COMPILE_OUT_z14
-- Performing Test COMPILE_OUT_z14 - Failed
-- </FindZVECTOR>
-- Version: 10.1.0
-- Build type: Release
-- Using Kineto with CUPTI support
-- Configuring Kineto dependency:
--   KINETO_SOURCE_DIR = /mnt/Archive/Downloads/1/pytorch/third_party/kineto/libkineto
--   KINETO_BUILD_TESTS = OFF
--   KINETO_LIBRARY_TYPE = static
--   CUDA_SOURCE_DIR = /usr/local/cuda-12.2
--   CUDA_INCLUDE_DIRS = /usr/local/cuda-12.2/include
--   CUPTI_INCLUDE_DIR = /usr/local/cuda-12.2/include
--   CUDA_cupti_LIBRARY = /usr/local/cuda-12.2/lib64/libcupti.so
-- Found CUPTI
-- Found PythonInterp: /home/notooth/miniconda3/bin/python (found version "3.11.4") 
INFO ROCM_SOURCE_DIR = 
-- Kineto: FMT_SOURCE_DIR = /mnt/Archive/Downloads/1/pytorch/third_party/fmt
-- Kineto: FMT_INCLUDE_DIR = /mnt/Archive/Downloads/1/pytorch/third_party/fmt/include
INFO CUPTI_INCLUDE_DIR = /usr/local/cuda-12.2/include
INFO ROCTRACER_INCLUDE_DIR = /include/roctracer
INFO DYNOLOG_INCLUDE_DIR = /mnt/Archive/Downloads/1/pytorch/third_party/kineto/libkineto/third_party/dynolog/
INFO IPCFABRIC_INCLUDE_DIR = /mnt/Archive/Downloads/1/pytorch/third_party/kineto/libkineto/third_party/dynolog//dynolog/src/ipcfabric/
-- Configured Kineto
-- GCC 8.5.0: Adding gcc and gcc_s libs to link line
-- Performing Test HAS_WERROR_RETURN_TYPE
-- Performing Test HAS_WERROR_RETURN_TYPE - Success
-- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR
-- Performing Test HAS_WERROR_NON_VIRTUAL_DTOR - Success
-- Performing Test HAS_WERROR_BRACED_SCALAR_INIT
-- Performing Test HAS_WERROR_BRACED_SCALAR_INIT - Failed
-- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT
-- Performing Test HAS_WERROR_RANGE_LOOP_CONSTRUCT - Failed
-- Performing Test HAS_WERROR_BOOL_OPERATION
-- Performing Test HAS_WERROR_BOOL_OPERATION - Success
-- Performing Test HAS_WNARROWING
-- Performing Test HAS_WNARROWING - Success
-- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS
-- Performing Test HAS_WNO_MISSING_FIELD_INITIALIZERS - Success
-- Performing Test HAS_WNO_TYPE_LIMITS
-- Performing Test HAS_WNO_TYPE_LIMITS - Success
-- Performing Test HAS_WNO_ARRAY_BOUNDS
-- Performing Test HAS_WNO_ARRAY_BOUNDS - Success
-- Performing Test HAS_WNO_UNKNOWN_PRAGMAS
-- Performing Test HAS_WNO_UNKNOWN_PRAGMAS - Success
-- Performing Test HAS_WNO_UNUSED_PARAMETER
-- Performing Test HAS_WNO_UNUSED_PARAMETER - Success
-- Performing Test HAS_WNO_UNUSED_FUNCTION
-- Performing Test HAS_WNO_UNUSED_FUNCTION - Success
-- Performing Test HAS_WNO_UNUSED_RESULT
-- Performing Test HAS_WNO_UNUSED_RESULT - Success
-- Performing Test HAS_WNO_STRICT_OVERFLOW
-- Performing Test HAS_WNO_STRICT_OVERFLOW - Success
-- Performing Test HAS_WNO_STRICT_ALIASING
-- Performing Test HAS_WNO_STRICT_ALIASING - Success
-- Performing Test HAS_WVLA_EXTENSION
-- Performing Test HAS_WVLA_EXTENSION - Failed
-- Performing Test HAS_WNEWLINE_EOF
-- Performing Test HAS_WNEWLINE_EOF - Failed
-- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE
-- Performing Test HAS_WINCONSISTENT_MISSING_OVERRIDE - Failed
-- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE
-- Performing Test HAS_WINCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed
-- Performing Test HAS_WNO_ERROR_PEDANTIC
-- Performing Test HAS_WNO_ERROR_PEDANTIC - Success
-- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST
-- Performing Test HAS_WNO_ERROR_OLD_STYLE_CAST - Success
-- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE
-- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_OVERRIDE - Failed
-- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE
-- Performing Test HAS_WNO_ERROR_INCONSISTENT_MISSING_DESTRUCTOR_OVERRIDE - Failed
-- Performing Test HAS_WCONSTANT_CONVERSION
-- Performing Test HAS_WCONSTANT_CONVERSION - Failed
-- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION
-- Performing Test HAS_WNO_INVALID_PARTIAL_SPECIALIZATION - Success
-- Performing Test HAS_WNO_UNUSED_PRIVATE_FIELD
-- Performing Test HAS_WNO_UNUSED_PRIVATE_FIELD - Success
-- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE
-- Performing Test HAS_WNO_ALIGNED_ALLOCATION_UNAVAILABLE - Success
-- Performing Test HAS_WNO_MISSING_BRACES
-- Performing Test HAS_WNO_MISSING_BRACES - Success
-- Performing Test HAS_WUNUSED_LAMBDA_CAPTURE
-- Performing Test HAS_WUNUSED_LAMBDA_CAPTURE - Failed
-- Performing Test HAS_QUNUSED_ARGUMENTS
-- Performing Test HAS_QUNUSED_ARGUMENTS - Failed
-- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS
-- Performing Test HAS_FDIAGNOSTICS_COLOR_ALWAYS - Success
-- Performing Test HAS_FALIGNED_NEW
-- Performing Test HAS_FALIGNED_NEW - Success
-- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE
-- Performing Test HAS_WNO_UNUSED_BUT_SET_VARIABLE - Success
-- Performing Test HAS_WNO_MAYBE_UNINITIALIZED
-- Performing Test HAS_WNO_MAYBE_UNINITIALIZED - Success
-- Performing Test HAS_FSTANDALONE_DEBUG
-- Performing Test HAS_FSTANDALONE_DEBUG - Failed
-- Performing Test HAS_FNO_MATH_ERRNO
-- Performing Test HAS_FNO_MATH_ERRNO - Success
-- Performing Test HAS_FNO_TRAPPING_MATH
-- Performing Test HAS_FNO_TRAPPING_MATH - Success
-- Performing Test HAS_WERROR_FORMAT
-- Performing Test HAS_WERROR_FORMAT - Success
-- Performing Test HAS_WERROR_CAST_FUNCTION_TYPE
-- Performing Test HAS_WERROR_CAST_FUNCTION_TYPE - Success
-- Performing Test HAS_WNO_STRINGOP_OVERFLOW
-- Performing Test HAS_WNO_STRINGOP_OVERFLOW - Success
-- Looking for backtrace
-- Looking for backtrace - found
-- backtrace facility detected in default set of libraries
-- Found Backtrace: /usr/include  
-- don't use NUMA
-- headers outputs: 
-- sources outputs: 
-- declarations_yaml outputs: 
-- Using ATen parallel backend: OMP
CMake Deprecation Warning at third_party/sleef/CMakeLists.txt:91 (cmake_policy):
  The OLD behavior for policy CMP0066 will be removed from a future version
  of CMake.

  The cmake-policies(7) manual explains that the OLD behaviors of all
  policies are deprecated and that a policy should be set to OLD only under
  specific short-term circumstances.  Projects should be ported to the NEW
  behavior and not rely on setting a policy to OLD.


-- Found OpenSSL: /usr/lib64/libcrypto.so (found version "1.1.1k")  
-- Check size of long double
-- Check size of long double - done
-- Performing Test COMPILER_SUPPORTS_LONG_DOUBLE
-- Performing Test COMPILER_SUPPORTS_LONG_DOUBLE - Success
-- Performing Test COMPILER_SUPPORTS_FLOAT128
-- Performing Test COMPILER_SUPPORTS_FLOAT128 - Success
-- Performing Test COMPILER_SUPPORTS_SSE2
-- Performing Test COMPILER_SUPPORTS_SSE2 - Success
-- Performing Test COMPILER_SUPPORTS_SSE4
-- Performing Test COMPILER_SUPPORTS_SSE4 - Success
-- Performing Test COMPILER_SUPPORTS_AVX
-- Performing Test COMPILER_SUPPORTS_AVX - Success
-- Performing Test COMPILER_SUPPORTS_FMA4
-- Performing Test COMPILER_SUPPORTS_FMA4 - Success
-- Performing Test COMPILER_SUPPORTS_AVX2
-- Performing Test COMPILER_SUPPORTS_AVX2 - Success
-- Performing Test COMPILER_SUPPORTS_AVX512F
-- Performing Test COMPILER_SUPPORTS_AVX512F - Success
-- Performing Test COMPILER_SUPPORTS_OPENMP
-- Performing Test COMPILER_SUPPORTS_OPENMP - Success
-- Performing Test COMPILER_SUPPORTS_WEAK_ALIASES
-- Performing Test COMPILER_SUPPORTS_WEAK_ALIASES - Success
-- Performing Test COMPILER_SUPPORTS_BUILTIN_MATH
-- Performing Test COMPILER_SUPPORTS_BUILTIN_MATH - Success
-- Performing Test COMPILER_SUPPORTS_SYS_GETRANDOM
-- Performing Test COMPILER_SUPPORTS_SYS_GETRANDOM - Success
-- Configuring build for SLEEF-v3.6.0
   Target system: Linux-4.18.0-477.21.1.el8_8.x86_64
   Target processor: x86_64
   Host system: Linux-4.18.0-477.21.1.el8_8.x86_64
   Host processor: x86_64
   Detected C compiler: GNU @ /usr/bin/cc
   CMake: 3.26.4
   Make program: /home/notooth/miniconda3/bin/ninja
-- Using option `-Wall -Wno-unused -Wno-attributes -Wno-unused-result -Wno-psabi -ffp-contract=off -fno-math-errno -fno-trapping-math` to compile libsleef
-- Building shared libs : OFF
-- Building static test bins: OFF
-- MPFR : LIB_MPFR-NOTFOUND
-- GMP : /usr/lib64/libgmp.so
-- RT : /usr/lib64/librt.so
-- FFTW3 : LIBFFTW3-NOTFOUND
-- OPENSSL : 1.1.1k
-- SDE : SDE_COMMAND-NOTFOUND
-- RUNNING_ON_TRAVIS : 
-- COMPILER_SUPPORTS_OPENMP : 1
AT_INSTALL_INCLUDE_DIR include/ATen/core
core header install: /mnt/Archive/Downloads/1/pytorch/build/aten/src/ATen/core/TensorBody.h
core header install: /mnt/Archive/Downloads/1/pytorch/build/aten/src/ATen/core/aten_interned_strings.h
core header install: /mnt/Archive/Downloads/1/pytorch/build/aten/src/ATen/core/enum_tag.h
-- Generating sources for unboxing kernels /home/notooth/miniconda3/bin/python;-m;torchgen.gen_executorch;--source-path=/mnt/Archive/Downloads/1/pytorch/test/edge/../../test/edge;--install-dir=/mnt/Archive/Downloads/1/pytorch/build/out;--tags-path=/mnt/Archive/Downloads/1/pytorch/test/edge/../../aten/src/ATen/native/tags.yaml;--aten-yaml-path=/mnt/Archive/Downloads/1/pytorch/test/edge/../../aten/src/ATen/native/native_functions.yaml;--use-aten-lib;--op-selection-yaml-path=/mnt/Archive/Downloads/1/pytorch/test/edge/../../test/edge/selected_operators.yaml;--custom-ops-yaml-path=/mnt/Archive/Downloads/1/pytorch/test/edge/../../test/edge/custom_ops.yaml
-- Performing Test HAS_WNO_UNUSED_VARIABLE
-- Performing Test HAS_WNO_UNUSED_VARIABLE - Success
-- Performing Test HAS_WNO_UNUSED_BUT_SET_PARAMETER
-- Performing Test HAS_WNO_UNUSED_BUT_SET_PARAMETER - Success
-- _GLIBCXX_USE_CXX11_ABI=1 is already defined as a cmake variable
CMake Warning (dev) at torch/CMakeLists.txt:389:
  Syntax Warning in cmake code at column 107

  Argument not separated from preceding token by whitespace.
This warning is for project developers.  Use -Wno-dev to suppress it.

CMake Warning (dev) at torch/CMakeLists.txt:389:
  Syntax Warning in cmake code at column 115

  Argument not separated from preceding token by whitespace.
This warning is for project developers.  Use -Wno-dev to suppress it.

-- Automatic GPU detection failed. Building for common architectures.
-- Autodetected CUDA architecture(s): 3.5;5.0;8.0;8.6;8.9;9.0
-- Using lib/python3.11/site-packages as python relative installation path
CMake Warning at CMakeLists.txt:1119 (message):
  Generated cmake files are only fully tested if one builds with system glog,
  gflags, and protobuf.  Other settings may generate files that are not well
  tested.


-- 
-- ******** Summary ********
-- General:
--   CMake version         : 3.26.4
--   CMake command         : /home/notooth/miniconda3/bin/cmake
--   System                : Linux
--   C++ compiler          : /usr/bin/c++
--   C++ compiler id       : GNU
--   C++ compiler version  : 8.5.0
--   Using ccache if found : ON
--   Found ccache          : CCACHE_PROGRAM-NOTFOUND
--   CXX flags             :  -D_GLIBCXX_USE_CXX11_ABI=1 -fvisibility-inlines-hidden -DUSE_PTHREADPOOL -DNDEBUG -DUSE_KINETO -DLIBKINETO_NOROCTRACER -DUSE_FBGEMM -DUSE_QNNPACK -DUSE_PYTORCH_QNNPACK -DUSE_XNNPACK -DSYMBOLICATE_MOBILE_DEBUG_HANDLE -O2 -fPIC -Wall -Wextra -Werror=return-type -Werror=non-virtual-dtor -Werror=bool-operation -Wnarrowing -Wno-missing-field-initializers -Wno-type-limits -Wno-array-bounds -Wno-unknown-pragmas -Wno-unused-parameter -Wno-unused-function -Wno-unused-result -Wno-strict-overflow -Wno-strict-aliasing -Wno-stringop-overflow -Wno-psabi -Wno-error=pedantic -Wno-error=old-style-cast -Wno-invalid-partial-specialization -Wno-unused-private-field -Wno-aligned-allocation-unavailable -Wno-missing-braces -fdiagnostics-color=always -faligned-new -Wno-unused-but-set-variable -Wno-maybe-uninitialized -fno-math-errno -fno-trapping-math -Werror=format -Werror=cast-function-type -Wno-stringop-overflow
--   Build type            : Release
--   Compile definitions   : ONNX_ML=1;ONNXIFI_ENABLE_EXT=1;ONNX_NAMESPACE=onnx_torch;IDEEP_USE_MKL;HAVE_MMAP=1;_FILE_OFFSET_BITS=64;HAVE_SHM_OPEN=1;HAVE_SHM_UNLINK=1;HAVE_MALLOC_USABLE_SIZE=1;USE_EXTERNAL_MZCRC;MINIZ_DISABLE_ZIP_READER_CRC32_CHECKS;BUILD_NVFUSER
--   CMAKE_PREFIX_PATH     : /home/notooth/miniconda3/lib/python3.11/site-packages;/usr/local/cuda-12.2;/usr/local/cuda-12.2
--   CMAKE_INSTALL_PREFIX  : /mnt/Archive/Downloads/1/pytorch/torch
--   USE_GOLD_LINKER       : OFF
-- 
--   TORCH_VERSION         : 2.1.0
--   BUILD_CAFFE2          : OFF
--   BUILD_CAFFE2_OPS      : OFF
--   BUILD_STATIC_RUNTIME_BENCHMARK: OFF
--   BUILD_TENSOREXPR_BENCHMARK: OFF
--   BUILD_NVFUSER_BENCHMARK: OFF
--   BUILD_BINARY          : OFF
--   BUILD_CUSTOM_PROTOBUF : ON
--     Link local protobuf : ON
--   BUILD_DOCS            : OFF
--   BUILD_PYTHON          : True
--     Python version      : 3.11.4
--     Python executable   : /home/notooth/miniconda3/bin/python
--     Pythonlibs version  : 3.11.4
--     Python library      : /home/notooth/miniconda3/lib/libpython3.11.a
--     Python includes     : /home/notooth/miniconda3/include/python3.11
--     Python site-packages: lib/python3.11/site-packages
--   BUILD_SHARED_LIBS     : ON
--   CAFFE2_USE_MSVC_STATIC_RUNTIME     : OFF
--   BUILD_TEST            : True
--   BUILD_JNI             : OFF
--   BUILD_MOBILE_AUTOGRAD : OFF
--   BUILD_LITE_INTERPRETER: OFF
--   INTERN_BUILD_MOBILE   : 
--   TRACING_BASED         : OFF
--   USE_BLAS              : 1
--     BLAS                : mkl
--     BLAS_HAS_SBGEMM     : 
--   USE_LAPACK            : 1
--     LAPACK              : mkl
--   USE_ASAN              : OFF
--   USE_TSAN              : OFF
--   USE_CPP_CODE_COVERAGE : OFF
--   USE_CUDA              : ON
--     Split CUDA          : 
--     CUDA static link    : OFF
--     USE_CUDNN           : OFF
--     USE_EXPERIMENTAL_CUDNN_V8_API: ON
--     USE_CUSPARSELT      : OFF
--     CUDA version        : 12.2
--     USE_FLASH_ATTENTION : ON
--     USE_MEM_EFF_ATTENTION : ON
--     CUDA root directory : /usr/local/cuda-12.2
--     CUDA library        : /usr/lib64/libcuda.so
--     cudart library      : /usr/local/cuda-12.2/lib64/libcudart.so
--     cublas library      : /usr/local/cuda-12.2/lib64/libcublas.so
--     cufft library       : /usr/local/cuda-12.2/lib64/libcufft.so
--     curand library      : /usr/local/cuda-12.2/lib64/libcurand.so
--     cusparse library    : /usr/local/cuda-12.2/lib64/libcusparse.so
--     nvrtc               : /usr/local/cuda-12.2/lib64/libnvrtc.so
--     CUDA include path   : /usr/local/cuda-12.2/include
--     NVCC executable     : /usr/local/cuda-12.2/bin/nvcc
--     CUDA compiler       : /usr/local/cuda-12.2/bin/nvcc
--     CUDA flags          :  -D_GLIBCXX_USE_CXX11_ABI=1 -Xfatbin -compress-all -DONNX_NAMESPACE=onnx_torch -gencode arch=compute_35,code=sm_35 -gencode arch=compute_50,code=sm_50 -gencode arch=compute_80,code=sm_80 -gencode arch=compute_86,code=sm_86 -gencode arch=compute_89,code=sm_89 -gencode arch=compute_90,code=sm_90 -Xcudafe --diag_suppress=cc_clobber_ignored,--diag_suppress=set_but_not_used,--diag_suppress=field_without_dll_interface,--diag_suppress=base_class_has_different_dll_interface,--diag_suppress=dll_interface_conflict_none_assumed,--diag_suppress=dll_interface_conflict_dllexport_assumed,--diag_suppress=bad_friend_decl --expt-relaxed-constexpr --expt-extended-lambda  -Wno-deprecated-gpu-targets --expt-extended-lambda -DCUB_WRAPPED_NAMESPACE=at_cuda_detail -DCUDA_HAS_FP16=1 -D__CUDA_NO_HALF_OPERATORS__ -D__CUDA_NO_HALF_CONVERSIONS__ -D__CUDA_NO_HALF2_OPERATORS__ -D__CUDA_NO_BFLOAT16_CONVERSIONS__
--     CUDA host compiler  : 
--     CUDA --device-c     : OFF
--     USE_TENSORRT        : OFF
--   USE_ROCM              : OFF
--   BUILD_NVFUSER         : ON
--   USE_EIGEN_FOR_BLAS    : 
--   USE_FBGEMM            : ON
--     USE_FAKELOWP          : OFF
--   USE_KINETO            : ON
--   USE_FFMPEG            : OFF
--   USE_GFLAGS            : OFF
--   USE_GLOG              : OFF
--   USE_LEVELDB           : OFF
--   USE_LITE_PROTO        : OFF
--   USE_LMDB              : OFF
--   USE_METAL             : OFF
--   USE_PYTORCH_METAL     : OFF
--   USE_PYTORCH_METAL_EXPORT     : OFF
--   USE_MPS               : OFF
--   USE_FFTW              : OFF
--   USE_MKL               : ON
--   USE_MKLDNN            : ON
--   USE_MKLDNN_ACL        : OFF
--   USE_MKLDNN_CBLAS      : OFF
--   USE_UCC               : OFF
--   USE_ITT               : ON
--   USE_NCCL              : ON
--     USE_SYSTEM_NCCL     : OFF
--     USE_NCCL_WITH_UCC   : OFF
--   USE_NNPACK            : ON
--   USE_NUMPY             : ON
--   USE_OBSERVERS         : ON
--   USE_OPENCL            : OFF
--   USE_OPENCV            : OFF
--   USE_OPENMP            : ON
--   USE_TBB               : OFF
--   USE_MIMALLOC          : OFF
--   USE_VULKAN            : OFF
--   USE_PROF              : OFF
--   USE_QNNPACK           : ON
--   USE_PYTORCH_QNNPACK   : ON
--   USE_XNNPACK           : ON
--   USE_REDIS             : OFF
--   USE_ROCKSDB           : OFF
--   USE_ZMQ               : OFF
--   USE_DISTRIBUTED       : ON
--     USE_MPI               : OFF
--     USE_GLOO              : ON
--     USE_GLOO_WITH_OPENSSL : OFF
--     USE_TENSORPIPE        : ON
--   Public Dependencies  : caffe2::mkl
--   Private Dependencies : Threads::Threads;pthreadpool;cpuinfo;qnnpack;pytorch_qnnpack;nnpack;XNNPACK;fbgemm;ittnotify;fp16;caffe2::openmp;tensorpipe;gloo;foxi_loader;rt;fmt::fmt-header-only;kineto;gcc_s;gcc;dl
--   Public CUDA Deps.    : caffe2::cufft;caffe2::curand;caffe2::cublas
--   Private CUDA Deps.   : __caffe2_nccl;tensorpipe_cuda;gloo_cuda;/usr/local/cuda-12.2/lib64/libcudart.so;CUDA::cusparse;CUDA::curand;CUDA::cufft;ATEN_CUDA_FILES_GEN_LIB
--   USE_COREML_DELEGATE     : OFF
--   BUILD_LAZY_TS_BACKEND   : ON
--   TORCH_DISABLE_GPU_ASSERTS : ON
-- Performing Test HAS_WMISSING_PROTOTYPES
-- Performing Test HAS_WMISSING_PROTOTYPES - Failed
-- Performing Test HAS_WERROR_MISSING_PROTOTYPES
-- Performing Test HAS_WERROR_MISSING_PROTOTYPES - Failed
-- Configuring done (74.4s)
CMake Warning at caffe2/CMakeLists.txt:813 (add_library):
  Cannot generate a safe runtime search path for target torch_cpu because
  files in some directories may conflict with libraries in implicit
  directories:

    runtime library [libgomp.so.1] in /usr/lib/gcc/x86_64-redhat-linux/8 may be hidden by files in:
      /home/notooth/miniconda3/lib

  Some of these libraries may not be found correctly.


-- Generating done (4.9s)
-- Build files have been written to: /mnt/Archive/Downloads/1/pytorch/build
cmake --build . --target install --config Release
[3/4] Generating ATen sources
[4046/6950] Running gen_proto.py on onnx/onnx.in.proto
Processing /mnt/Archive/Downloads/1/pytorch/third_party/onnx/onnx/onnx.in.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_onnx_torch-ml.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_onnx_torch-ml.proto3
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-ml.pb.h
generating /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_pb.py
[4054/6950] Running gen_proto.py on onnx/onnx-operators.in.proto
Processing /mnt/Archive/Downloads/1/pytorch/third_party/onnx/onnx/onnx-operators.in.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-operators_onnx_torch-ml.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-operators_onnx_torch-ml.proto3
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-operators-ml.pb.h
generating /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_operators_pb.py
[4055/6950] Running gen_proto.py on onnx/onnx-data.in.proto
Processing /mnt/Archive/Downloads/1/pytorch/third_party/onnx/onnx/onnx-data.in.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-data_onnx_torch.proto
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-data_onnx_torch.proto3
Writing /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx-data.pb.h
generating /mnt/Archive/Downloads/1/pytorch/build/third_party/onnx/onnx/onnx_data_pb.py
[4287/6950] Performing build step for 'nccl_external'
FAILED: nccl_external-prefix/src/nccl_external-stamp/nccl_external-build nccl/lib/libnccl_static.a /mnt/Archive/Downloads/1/pytorch/build/nccl_external-prefix/src/nccl_external-stamp/nccl_external-build /mnt/Archive/Downloads/1/pytorch/build/nccl/lib/libnccl_static.a 
cd /mnt/Archive/Downloads/1/pytorch/third_party/nccl/nccl && make -j12 -l12 CXX=/usr/bin/c++ CUDA_HOME=/usr/local/cuda-12.2 NVCC=/usr/local/cuda-12.2/bin/nvcc "NVCC_GENCODE=-gencode=arch=compute_35,code=sm_35 -gencode=arch=compute_50,code=sm_50 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 -gencode=arch=compute_89,code=sm_89 -gencode=arch=compute_90,code=sm_90" BUILDDIR=/mnt/Archive/Downloads/1/pytorch/build/nccl VERBOSE=0 DEBUG=0 && /home/notooth/miniconda3/bin/cmake -E touch /mnt/Archive/Downloads/1/pytorch/build/nccl_external-prefix/src/nccl_external-stamp/nccl_external-build
make -C src build BUILDDIR=/mnt/Archive/Downloads/1/pytorch/build/nccl
make[1]: Entering directory '/mnt/Archive/Downloads/1/pytorch/third_party/nccl/nccl/src'
NVCC_GENCODE is -gencode=arch=compute_35,code=sm_35 -gencode=arch=compute_50,code=sm_50 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 -gencode=arch=compute_89,code=sm_89 -gencode=arch=compute_90,code=sm_90
Generating nccl.h.in                           > /mnt/Archive/Downloads/1/pytorch/build/nccl/include/nccl.h
Generating nccl.pc.in                          > /mnt/Archive/Downloads/1/pytorch/build/nccl/lib/pkgconfig/nccl.pc
Grabbing   include/nccl_net.h                  > /mnt/Archive/Downloads/1/pytorch/build/nccl/include/nccl_net.h
Compiling  init.cc                             > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/init.o
Compiling  misc/nvmlwrap.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/nvmlwrap.o
Compiling  misc/ibvsymbols.cc                  > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/ibvsymbols.o
Compiling  misc/ibvwrap.cc                     > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/ibvwrap.o
Compiling  misc/gdrwrap.cc                     > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/gdrwrap.o
Compiling  misc/utils.cc                       > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/utils.o
Compiling  misc/argcheck.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/argcheck.o
Compiling  misc/socket.cc                      > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/socket.o
Compiling  misc/shmutils.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/shmutils.o
Compiling  misc/profiler.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/profiler.o
Compiling  misc/param.cc                       > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/param.o
Compiling  misc/strongstream.cc                > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/strongstream.o
Compiling  misc/ipcsocket.cc                   > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/misc/ipcsocket.o
Compiling  transport/p2p.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/p2p.o
Compiling  transport/shm.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/shm.o
Compiling  transport/net.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/net.o
Compiling  transport/net_socket.cc             > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/net_socket.o
Compiling  transport/net_ib.cc                 > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/net_ib.o
Compiling  transport/coll_net.cc               > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/coll_net.o
Compiling  transport/nvls.cc                   > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/transport/nvls.o
Compiling  collectives/sendrecv.cc             > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/sendrecv.o
Compiling  collectives/all_reduce.cc           > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/all_reduce.o
Compiling  collectives/all_gather.cc           > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/all_gather.o
Compiling  collectives/broadcast.cc            > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/broadcast.o
Compiling  collectives/reduce.cc               > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/reduce.o
Compiling  collectives/reduce_scatter.cc       > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/reduce_scatter.o
Compiling  graph/topo.cc                       > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/topo.o
Compiling  graph/paths.cc                      > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/paths.o
Compiling  graph/search.cc                     > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/search.o
Compiling  graph/connect.cc                    > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/connect.o
Compiling  graph/rings.cc                      > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/rings.o
Compiling  graph/trees.cc                      > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/trees.o
Compiling  graph/tuning.cc                     > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/tuning.o
Compiling  graph/xml.cc                        > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/graph/xml.o
Compiling  enhcompat.cc                        > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/enhcompat.o
make[2]: Entering directory '/mnt/Archive/Downloads/1/pytorch/third_party/nccl/nccl/src/collectives/device'
NVCC_GENCODE is -gencode=arch=compute_35,code=sm_35 -gencode=arch=compute_50,code=sm_50 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 -gencode=arch=compute_89,code=sm_89 -gencode=arch=compute_90,code=sm_90
Generating rules                               > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/device/Makefile.rules
NVCC_GENCODE is -gencode=arch=compute_35,code=sm_35 -gencode=arch=compute_50,code=sm_50 -gencode=arch=compute_80,code=sm_80 -gencode=arch=compute_86,code=sm_86 -gencode=arch=compute_89,code=sm_89 -gencode=arch=compute_90,code=sm_90
Copying    sendrecv.cu                         > /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/device/sendrecv_sum_i8.cu
nvcc fatal   : Unsupported gpu architecture 'compute_35'
make[2]: *** [Makefile:53: /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/device/sendrecv.dep] Error 1
make[2]: Leaving directory '/mnt/Archive/Downloads/1/pytorch/third_party/nccl/nccl/src/collectives/device'
make[1]: *** [Makefile:58: /mnt/Archive/Downloads/1/pytorch/build/nccl/obj/collectives/device/colldevice.a] Error 2
make[1]: Leaving directory '/mnt/Archive/Downloads/1/pytorch/third_party/nccl/nccl/src'
make: *** [Makefile:25: src.build] Error 2
[4312/6950] Building CXX object third_party/fbgemm/CMakeFiles/fbgemm_avx2.dir/src/FbgemmI8DepthwiseAvx2.cc.o
ninja: build stopped: subcommand failed.