Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

compile error:token ""__CUDACC_VER__ is no longer supported. #13785

Open
squirrelim opened this issue Jan 7, 2019 · 5 comments
Open

compile error:token ""__CUDACC_VER__ is no longer supported. #13785

squirrelim opened this issue Jan 7, 2019 · 5 comments

Comments

@squirrelim
Copy link

Description

compile error
what to do:
git clone --recursive https://github.com/dmlc/mxnet.git
cd mxnet
git checkout 998378a
git submodule update --init --recursive
make -j $(nproc) USE_OPENCV=1 USE_BLAS=openblas USE_CUDA=1 USE_CUDA_PATH=/usr/local/cuda USE_CUDNN=1

Error Message

cd /root/Deformable-ConvNets/mxnet/dmlc-core; make libdmlc.a USE_SSE=1 config=/root/Deformable-ConvNets/mxnet/config.mk; cd /root/Deformable-ConvNets/mxnet
make[1]: Entering directory '/root/Deformable-ConvNets/mxnet/dmlc-core'
make[1]: 'libdmlc.a' is up to date.
make[1]: Leaving directory '/root/Deformable-ConvNets/mxnet/dmlc-core'
/usr/local/cuda/bin/nvcc -std=c++11 -Xcompiler -D_FORCE_INLINES -O3 -ccbin g++ -gencode arch=compute_30,code=sm_30 -gencode arch=compute_35,code=sm_35 -gencode arch=compute_50,code=sm_50 -gencode arch=compute_52,code=sm_52 -gencode arch=compute_60,code=sm_60 -gencode arch=compute_61,code=[sm_61,compute_61] --fatbin-options -compress-all -Xcompiler "-DMSHADOW_FORCE_STREAM -Wall -Wsign-compare -O3 -DNDEBUG=1 -I/root/Deformable-ConvNets/mxnet/mshadow/ -I/root/Deformable-ConvNets/mxnet/dmlc-core/include -fPIC -I/root/Deformable-ConvNets/mxnet/nnvm/include -I/root/Deformable-ConvNets/mxnet/dlpack/include -Iinclude -funroll-loops -Wno-unused-variable -Wno-unused-parameter -Wno-unknown-pragmas -Wno-unused-local-typedefs -msse3 -I/usr/local/cuda/include -DMSHADOW_USE_CBLAS=1 -DMSHADOW_USE_MKL=0 -DMSHADOW_RABIT_PS=0 -DMSHADOW_DIST_PS=0 -DMSHADOW_USE_PASCAL=0 -DMXNET_USE_OPENCV=1 -I/usr/local/include/opencv -I/usr/local/include -fopenmp -DMXNET_USE_LAPACK -DMSHADOW_USE_CUDNN=1 -I/root/Deformable-ConvNets/mxnet/cub -DMXNET_USE_NVRTC=0" -M -MT build/src/operator/tensor/indexing_op_gpu.o src/operator/tensor/indexing_op.cu >build/src/operator/tensor/indexing_op_gpu.d
In file included from /usr/local/cuda/include/common_functions.h:50:0,
from /usr/local/cuda/include/cuda_runtime.h:115,
from :0:
/usr/local/cuda/include/crt/common_functions.h:64:24: error: token ""CUDACC_VER is no longer supported. Use CUDACC_VER_MAJOR, CUDACC_VER_MINOR, and CUDACC_VER_BUILD instead."" is not valid in preprocessor expressions
#define CUDACC_VER "CUDACC_VER is no longer supported. Use CUDACC_VER_MAJOR, CUDACC_VER_MINOR, and CUDACC_VER_BUILD instead."
^
/root/Deformable-ConvNets/mxnet/cub/cub/device/dispatch/../../agent/../thread/../thread/../util_arch.cuh:46:6: note: in expansion of macro 'CUDACC_VER'
#if (CUDACC_VER >= 80500)
^
Makefile:274: recipe for target 'build/src/operator/tensor/indexing_op_gpu.o' failed
make: *** [build/src/operator/tensor/indexing_op_gpu.o] Error 1
make: *** Waiting for unfinished jobs....

Environment:

Ubuntu 16.04
cuda9.0
cudnn7.05

@mxnet-label-bot
Copy link
Contributor

Hey, this is the MXNet Label Bot.
Thank you for submitting the issue! I will try and suggest some labels so that the appropriate MXNet community members can help resolve it.
Here are my recommended labels: Build, Installation

@zachgk
Copy link
Contributor

zachgk commented Jan 7, 2019

Thank you for submitting the issue! I'm labeling it so the MXNet community members can help resolve it.

@mxnet-label-bot add [Build, CUDA]

@Lin-Zhipeng
Copy link

I also encounter the same question. Is there any solution now?

@zachgk
Copy link
Contributor

zachgk commented Aug 15, 2019

@squirrelim What is the reason behind using commit 998378a? It is currently over 2 years old.

@LZP4GitHub What commit are you using to build?

@qoopen0815
Copy link

qoopen0815 commented Nov 15, 2019

@zachgk Here is the reason behind using commit 998378a, maybe

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

6 participants