-
Notifications
You must be signed in to change notification settings - Fork 2k
gpu not found in a docker container #92
Comments
Which command did you use to start tensorflow? You should use the prebuilt image available on GCR:
If it doesn't work, try starting with the following:
|
Thank you, it worked. But I need to customize the nvidia-docker image for my research. Can I build the image using Dockerfile as Docker build does? More generally, can I use nvidia-docker like docker such as flags or any other commands? Thank you in advance! |
Your custom image must be based on one of the image we provide on the DockerHub: Yes, you can use all the flags and commands supported by |
Thank you! It works perfect for me. :-) |
Great to know! Note that you can also inherit from the tensorflow image since it itself uses one of our image. |
I have a question, Will i be able to use GPU of another system connected to the same network using docker?. If yes, how can I do that?. Will i be able to use these GPU's to train model zoo present in the Tensorflow Pre-trained models |
With nvidia-docker 2.0 + using a remote docker daemon (e.g. Locking, please don't resurrect old issues. |
Hello,
I recently built a cuda-7.5 with cudnn4 devel docker image with tensorflow. But inside the docker container gpus were not recognized.
The log is as below:
import tensorflow as tf
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcublas.so.7.5 locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcudnn.so.4 locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcufft.so.7.5 locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcuda.so.1 locally
I tensorflow/stream_executor/dso_loader.cc:105] successfully opened CUDA library libcurand.so.7.5 locally
sess = tf.Session()
E tensorflow/stream_executor/cuda/cuda_driver.cc:481] failed call to cuInit: CUresult(-1)
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:114] retrieving CUDA diagnostic information for host: 18a61c37a941
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:121] hostname: 18a61c37a941
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:146] libcuda reported version is: Invalid argument: expected %d.%d form for driver version; got "1"
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:257] driver version file contents: """NVRM version: NVIDIA UNIX x86_64 Kernel Module 352.93 Tue Apr 5 18:18:24 PDT 2016
GCC version: gcc version 4.8.4 (Ubuntu 4.8.4-2ubuntu1~14.04)
"""
I tensorflow/stream_executor/cuda/cuda_diagnostics.cc:150] kernel reported version is: 352.93
I tensorflow/core/common_runtime/gpu/gpu_init.cc:81] No GPU devices available on machine.
In addition, I also tried the flag "--device /dev/nvidia0:~", but the result was the same as above and I could not run "nvidia-smi" inside the container saying that "command not found" error. Is there any way to fix this?
The text was updated successfully, but these errors were encountered: