Sign in
Categories
Migration Mapping Assistant Your Saved List Partners Sell in AWS Marketplace Amazon Web Services Home Help

tensorflow_p36 doesn't work with GPU

  • By emilia
  • on 05/01/2019

You pretty much need to set it up everything yourself. Here is what you get out-of-the-box:

(tensorflow_p36) ubuntu@ip-172-31-6-11:~$ source activate tensorflow_p36
(tensorflow_p36) ubuntu@ip-172-31-6-11:~$ python
Python 3.6.5 |Anaconda, Inc.| (default, Apr 29 2018, 16:14:56)
[GCC 7.2.0] on linux
Type "help", "copyright", "credits" or "license" for more information.
>>> import tensorflow as tf
>>> sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))
2019-05-01 23:37:47.936444: F tensorflow/stream_executor/lib/statusor.cc:34] Attempting to fetch value instead of handling error Internal: no supported devices found for platform CUDA
Aborted (core dumped)


  • By DeepAV
  • on 11/04/2019

I just tried these exact commands on this AMI and here's what I got: ubuntu@ip-172-31-26-72:~$ source activate tensorflow_p36 WARNING: First activation might take some time (1+ min). Installing TensorFlow optimized for your Amazon EC2 instance...... Env where framework will be re-installed: tensorflow_p36 Instance g4dn.xlarge is identified as a GPU instance, removing tensorflow-serving-cpu Installation complete. (tensorflow_p36) ubuntu@ip-172-31-26-72:~$ python Python 3.6.5 |Anaconda, Inc.| (default, Apr 29 2018, 16:14:56) [GCC 7.2.0] on linux Type "help", "copyright", "credits" or "license" for more information. >>> import tensorflow as tf >>> sess = tf.Session(config=tf.ConfigProto(log_device_placement=True)) Device mapping: /job:localhost/replica:0/task:0/device:XLA_GPU:0 -> device: XLA_GPU device /job:localhost/replica:0/task:0/device:XLA_CPU:0 -> device: XLA_CPU device /job:localhost/replica:0/task:0/device:GPU:0 -> device: 0, name: Tesla T4, pci bus id: 0000:00:1e.0, compute capability: 7.5 >>>

  • By aws-bhavin
  • on 05/09/2019

Hello, We do not observe the problem mentioned by you in the current version of the DLAMI v22.0. Can you share the version of the DLAMI used by you and on what EC2 instance-type? Could you also run "nvidia-smi" to double-check that the EC2 instance has GPUs. If nvidia-smi on a GPU instance resulted in the below error, "NVIDIA-SMI has failed because it couldn't communicate with the NVIDIA driver. Make sure that the latest NVIDIA driver is installed and running." it might be related to a recent Kernal update on Ubuntu which broke nvidia driver. In such case we recommend using the latest DLAMI. If you want to keep using the same AMI, please follow this thread to fix the issue: https://forums.aws.amazon.com/thread.jspa?messageID=894021 We would highly recommend that you post such questions on AWS-forum so that we can better assist you in resolving the problem quickly. Thanks