Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support cuda 7.5 and cudnn 7.0 #54

Closed
amolchanov86 opened this issue Nov 10, 2015 · 5 comments
Closed

Support cuda 7.5 and cudnn 7.0 #54

amolchanov86 opened this issue Nov 10, 2015 · 5 comments

Comments

@amolchanov86
Copy link

Please, add support of cuda 7.5 and cudnn 7.0.
The best way would be to introduce it as configuration option. Right now, I have to correct everything manually, since only cuda 7.0 and cudnn 6.5 are supported

@zheng-xq
Copy link
Contributor

There are two short-term options for now.

  1. Install Cuda 7.0 and Cudnn 6.5, in additional to your existing Cuda 7.5 and Cudnn 7.0 libraries. They are the officially supported versions. And it is okay to have multiple Cuda toolkits with different paths on your system. This is the recommended way.
  2. Manually correct source code to pick up Cuda 7.5. We are discussing options to make it easier to configure in the future. However, given Cuda toolkits and libraries are not necessarily backward compatible, you may or may not hit subtle issues down the road.

@shenxudeu
Copy link

I have Cuda 7.0 and Cudnn 7.0 installed on my machine. This message pop up when I import tensor flow:

I tensorflow/stream_executor/dso_loader.cc:77] LD_LIBRARY_PATH: :/usr/local/cuda-7.0/lib64
I tensorflow/stream_executor/cuda/cuda_dnn.cc:1062] Unable to load cuDNN DSO.

Does it mean it does NOT support Cudnn 7.0?

@zheng-xq
Copy link
Contributor

At this moment, the officially supported CUDNN version is 6.5 V2. This may change soon.

You can download the previous version from NVIDIA website, and use it in the configure script. This is okay even if you have CUDNN 7.0 V3 installed in the system.

https://developer.nvidia.com/rdp/cudnn-archive

@shenxudeu
Copy link

Thanks! It helps. I got another problem when I tries to build from source
on another machine (Nvidia Dev Box, Ubuntu 14.04.1).

When I follow the Installation for linux from tensorflow offical site,
everything is fine until the last step:Train your first TensorFlow neural
net model.

When I execute the command, python
tensorflow/models/image/mnist/convolutional.py, An ImportError promps.

Traceback (most recent call last): File
"tensorflow/models/image/mnist/convolutional.py", line 13, in
import tensorflow.python.platform File "tensorflow/tensorflow/init.py",
line 4, in from tensorflow.python import * File
"tensorflow/tensorflow/python/init.py", line 13, in from
tensorflow.core.framework.graph_pb2 import * ImportError: No module named
core.framework.graph_pb2

Would you shed some light on this?

Thanks,

Shen

On Mon, Nov 9, 2015 at 11:27 PM, zheng-xq notifications@github.com wrote:

At this moment, the officially supported CUDNN version is 6.5 V2. This may
change soon.

You can download the previous version from NVIDIA website, and use it in
the configure script. This is okay even if you have CUDNN 7.0 V3 installed
in the system.

https://developer.nvidia.com/rdp/cudnn-archive


Reply to this email directly or view it on GitHub
#54 (comment)
.

@vrv
Copy link

vrv commented Nov 24, 2015

De-duping with #20

@vrv vrv closed this as completed Nov 24, 2015
ilblackdragon added a commit to ilblackdragon/tensorflow that referenced this issue Mar 9, 2016
teamdandelion pushed a commit to teamdandelion/tensorflow that referenced this issue Aug 19, 2016
benoitsteiner pushed a commit that referenced this issue May 31, 2017
* OpenCL Improvements

* Registers Scatter and ScatterNd Ops for SYCL

* Registers Stack op for SYCL

* Fixes No sycl buffer found error for debug ops

* Registers MatMul and Transpose Ops to SYCL device for double

* Extends analyzer_cli_test.py test to cover SYCL

* Fixes Transpose Op for double when on SYCL

* Bumps Eigen version to fix double precision issue on SYCL

* Extends SessionDebugTestBase to cover SYCL

* Register SYCL implementations for random ops

* Avoid functions that might not be defined on SYCL device (#51)

* Avoid functions that might not be defined on SYCL device

* Simplify by using Eigen math functions

* OpenCL improvements

 - Bumps Eigen Version
 - Refactors Ops registration
 - Introduces workaround for Const Op related to the difference between
   CUDA which uses pointers and OpenCL that uses buffers/accessors
 - Extends memory types to cover DEVICE_SYCL as well
 - Introduces  GetSYCLDevice() method that returns list of supported devices
   with GPU device having the highest priority ( doesn't include blacklisted devices )
 - ::internal::Transpose -> tensorflow::internal::Transpose in order to
   avoid compilation reported error
 - re-introduces fix for bugged string replacement causing a lot of compilation
   warnings -c -> --include
 - Adds sycl_runtime to bazels ARRAY_DEPS
 - Replicates TF_CALL_GPU_PROXY_TYPES for SYCL

* [OpenCL] Fixes an issue caused by switch to aligned allocator for sycl buffer (#53)

* [Build] Use gcc/g++ as a host compiler to avoid #8394 (#54)

* [OpenCL] Fixes Scatter Op

* Fix testSimple and testConst in stack_op_test (#3)

* Fix testSimple and testConst in stack_op_test

* Create a specialisation of DoParallelConcatUpdate for SyclDevice and
register it

* Guard all code in TENSORFLOW_USE_SYCL

* Do not use sycl device for int32

* Registration of the Sycl version is now looking like the one for the GPU

* Remove added empty line

* Register batch normalization kernels for OpenCL (#61)

* [OpenCL] RandomGamma has no GPU friendly implementation (#57)

* [OpenCL] Compatibility fixes for TensorFlow 1.1.0-rc1

* [OpenCL] Implements BatchMatmul Op for SYCL

* Lowercase the device name when GPU or SYCL returned

* [OpenCL] kernel_estimator_test.py assertEqual-> assertAlmostEqual due to floating point representation on the device

* [Eigen] Version bump

* GPU device name string manipulation is not needed anymore

* [OpenCL] Adds SYCL to device backwards compatibility

* [OpenCL] Extends core_rnn_test.py to run for SYCL device

* [OpenCL] Minor optimizations for build script

* [OpenCL] Enables skip folder list in build script

* [OpenCL] Fixes ApplyAdamOp for Sycl device

* [OpenCL] SYCL device improvements

* [OpenCL] Fixes debug_ops's SEGFAULT for SYCL device

* [Build] Adds hexagon to skipped folders list

* [OpenCL] Removes EnterLameDuckMode from SYCL device and allocator

* [OpenCL] Registers Unique Op for SYCL device

* [OpenCL][Temporary] Disables tests for SYCL target due to features not being implemented yet

  Tests affected:
    - tensorflow/contrib/memory_stats/python/kernel_tests/memory_stats_ops_test.py
    - tensorflow/contrib/rnn/python/kernel_tests/core_rnn_test.py
    - tensorflow/python/kernel_tests/conv_ops_test.py
    - tensorflow/python/kernel_tests/depthwise_conv_op_test.py
    - tensorflow/python/kernel_tests/pooling_ops_3d_test.py
    - tensorflow/python/kernel_tests/pooling_ops_test.py
    - tensorflow/python/kernel_tests/scatter_nd_ops_test.py
    - tensorflow/python/training/adam_test.py
    - tensorflow/python/training/localhost_cluster_performance_test.py
    - tensorflow/python/training/training_ops_test.py

* [OpenCL][Temporary] Disables failing tests for SYCL in order to establish regression baseline

  Tests affected:
    - tensorflow/python/debug/cli/analyzer_cli_test.py
    - tensorflow/python/debug/lib/session_debug_testlib.py
    - tensorflow/python/debug/lib/stepper_test.py
    - tensorflow/python/kernel_tests/unstack_op_test.py
    - tensorflow/python/ops/image_ops_test.py

* [OpenCL] Take options.config.device_count() into consideration

* [OpenCL] Fixes compilation warning

* [OpenCL] device:SYCL:0 -> sycl:0

* [OpenCL] Removes unwanted flags in building script

Removes flags given to computecpp that enable SIMD instructions
Removes duplicate flags

* bool -> const bool

* [OpenCL] sycl in test_util.gpu_device_name() -> is_sycl_enabled()

* [OpenCL][Temporary] Disables failing tests for SYCL in order to establish regression baseline

  Test affected:
    - tensorflow/contrib/stateless/python/kernel_tests/stateless_random_ops_test.py

* Imports test_util from tensorflow.python.framework

* [OpenCL] Fixes formatting in Python code

* [OpenCL] Extends session_test.py to cover SYCL device

* [OpenCL] Cleans singleton class

* [OpenCL] Keeping CUDA happy

* [OpenCL][Temporary] Disables failing tests for SYCL in order to establish regression baseline

  Test affected:
   - tensorflow/contrib/rnn/python/kernel_tests/core_rnn_cell_test.py
   - tensorflow/contrib/seq2seq/python/kernel_tests/beam_search_ops_test.py

* Added support for building with SYCL on ARM.

* Acts on the review feedback from:
 - #9117 (comment)
 - #9117 (comment)

* [OpenCL] Fixes scatter_nd_op_test

* Fixes auto-merge mistake

* [OpenCL] struct SyclDevice -> class SyclDevice

* Revert "[OpenCL] struct SyclDevice -> class SyclDevice"

This reverts commit addd433.

* [OpenCL] Reverting refactoring commit.

  As requested in the review #9117 (comment)
  This change set will be re-introduced in smaller chunks.

* Revert "[OpenCL] device:SYCL:0 -> sycl:0"

This reverts commit cf16e60.

* Revert "[OpenCL] Adds SYCL to device backwards compatibility"

This reverts commit b8401b5.

* Acts on the feedback from #9117 (comment)

* control_flow_ops_py_test.py expects device name to be lower cased

* Acts on the feedback from #9117 (comment)

* Removes debug print

* Removes not needed partial specialisation

* [OpenCL] Registers ScatterNdFunctor for SYCL device

* [OpenCL] Make it compile

* [OpenCL] Follow gpu_device changes

* [OpenCL] Adds cxx_builtin_include_directory for python lib

  Fixes bazels missing undeclared inclusions that appeared after
  merge with TensorFlow upstream

* [OpenCL] Fixes Constant Op

* [OpenCL] gXX-4.8 -> gXX

* [OpenCL] Removes -D_GLIBCXX_USE_CXX11_ABI=0 as it breaks default compiler setup for Ubuntu 16.04

* Revert "[OpenCL] kernel_estimator_test.py assertEqual-> assertAlmostEqual due to floating point representation on the device"

This reverts commit 06c50c0.

* [OpenCL] CPU allocator is a singleton we should not delete it
lissyx added a commit to lissyx/tensorflow that referenced this issue Feb 16, 2018
Add GraphDef Memmapped conversion tool
tensorflow-copybara pushed a commit that referenced this issue Jul 30, 2019
Add binary operations such as: OpIAdd, OpFAdd, OpISub, OpFSub, OpIMul,
OpFDiv, OpFRem, OpFMod.

Closes #54

COPYBARA_INTEGRATE_REVIEW=tensorflow/mlir#54 from denis0x0D:sandbox/bin_ops ff3689c441a6a181dbe52c410562346993163013
PiperOrigin-RevId: 260734166
cjolivier01 pushed a commit to Cerebras/tensorflow that referenced this issue Dec 6, 2019
…_test_fixes_180626

Special casing GpuAtomicMin / GpuAtomicMax for ROCm
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants