Ubuntu 18.04 + Cuda9.0 + Cudnn7.1 + tensorflow-gpu1.5

Ubuntu 18.04 + Cuda9.0 + Cudnn7.1 + tensorflow-gpu1.5

1.gcc/g++降级

Ubuntu18.04自带的gcc/g++是7.0版本的,但cuda不支持这么高版本,我们需要安装4.8版本。

 

1.下载安装4.8版本的gcc/g++


 
 
  1. sudo apt-get install gcc -4.8
  2. sudo apt- get install g++ -4.8

2.让gcc软连接至4.8版本的gcc,g++软连接至4.5版本的g++

装完后进入到/usr/bin目录下


 
 
  1. sudo mv gcc gcc.bak #备份
  2. sudo ln -s gcc -4.8 gcc #重新链接

对g++也进行同样的操作

 

3.再查看gcc和g++版本号

gcc -v g++ -v
 
 

均显示gcc version 4.8 ,说明gcc 48.8安装成功。

 


2.安装GPU(针对ubuntu18.04)

安装所有推荐的驱动:

sudo ubuntu-drivers autoinstall
 
 

 


3.cuda安装

1.下载cuda安装包,我们安装9.0版本,地址:CUDA

将上面的选项框填好后,会弹出下面的选项框

下载CUDA9.0和全部四个补丁(这几个补丁不安装的话,cuda在使用过程中可能出现各种bug),他们会被保存到该用户目录下的 Downloads 文件夹下。

 

2.安装cuda

进入Downloads/后,输入:


 
 
  1. sh cuda_9 .0 .176_384 .81_linux .run
  2. sh cuda_9 .0 .176 .1_linux .run
  3. sh cuda_9 .0 .176 .2_linux .run
  4. sh cuda_9 .0 .176 .3_linux .run
  5. sh cuda_9 .0 .176 .4_linux .run

安装这五个文件

注意:执行,如果有安装了显卡驱动的,注意在提问是否安装显卡驱动时选择no(因为在前边已经装过了),其他 选择默认路径或者yes即可。


注:我们也可以不在usr/local下安装cuda,而在用户根目录下安装:

然后就可以了

 


3.添加访问cuda的路径

安装完毕之后,将以下两条加入.barshrc文件中.


 
 
  1. sudo vim ~/.barshrc
  2. export PATH=/usr/ local/cuda-9.0/bin ${PATH:+:$PATH}}
  3. export LD_LIBRARY_PATH=/usr/ local/cuda-9.0/lib64 ${LD_LIBRARY_PATH:+:${LD_LIBRARY_PATH}}

下边的这两条和上边的等价(暂时没搞清楚为什么):


 
 
  1. sudo vim ~/.barshrc
  2. export PATH=/usr/ local/cuda-9.0/bin: $PATH
  3. export LD_LIBRARY_PATH=/usr/ local/cuda-9.0/lib64: $LD_LIBRARY_PATH

 


4.CUDNN安装

CUDNN需要注册,我们安装7.1.3版本

注册完以后,出现下面的界面:

 

下载结束后,将压缩包进行解压缩。

然后,在Downloads/文件夹内输入如下命令,将CUDNN拷贝至CUDA的目录中(注,CUDNN无需安装):


 
 
  1. sudo cp cuda/ include/cudnn.h /usr/local/cuda/ include
  2. sudo cp cuda/lib64/libcudnn* /usr/local /cuda/lib64
  3. sudo chmod a+r /usr/local/cuda/ include/cudnn.h /usr/local/cuda/lib64/libcudnn*

 


5.cuda测试(看cuda是否安装好)

进入samples文件夹,一般在home目录下


 
 
  1. cd ~/NVIDIA_CUDA -9.1_Samples/
  2. make

编译完成后,进入:

cd ./bin/x86_64/linux/release 
 
 

使用deviceQuerybandwidthTest测试

 


 
 
  1. $ ./deviceQuery
  2. ./deviceQuery Starting...
  3. CUDA Device Query (Runtime API) version (CUDART static linking)
  4. Detected 1 CUDA Capable device(s)
  5. Device 0: "GeForce GTX 960M"
  6. CUDA Driver Version / Runtime Version 9.1 / 9.1
  7. CUDA Capability Major/Minor version number: 5.0
  8. Total amount of global memory: 2004 MBytes (2101870592 bytes)
  9. ( 5) Multiprocessors, (128) CUDA Cores/MP: 640 CUDA Cores
  10. GPU Max Clock rate: 1176 MHz (1.18 GHz)
  11. Memory Clock rate: 2505 Mhz
  12. Memory Bus Width: 128-bit
  13. L2 Cache Size: 2097152 bytes
  14. Maximum Texture Dimension Size (x,y,z) 1D=( 65536), 2D=( 65536, 65536), 3D=( 4096, 4096, 4096)
  15. Maximum Layered 1D Texture Size, ( num) layers 1D=( 16384), 2048 layers
  16. Maximum Layered 2D Texture Size, ( num) layers 2D=( 16384, 16384), 2048 layers
  17. Total amount of constant memory: 65536 bytes
  18. Total amount of shared memory per block: 49152 bytes
  19. Total number of registers available per block: 65536
  20. Warp size: 32
  21. Maximum number of threads per multiprocessor: 2048
  22. Maximum number of threads per block: 1024
  23. Max dimension size of a thread block (x,y,z): ( 1024, 1024, 64)
  24. Max dimension size of a grid size (x,y,z): ( 2147483647, 65535, 65535)
  25. Maximum memory pitch: 2147483647 bytes
  26. Texture alignment: 512 bytes
  27. Concurrent copy and kernel execution: Yes with 1 copy engine(s)
  28. Run time limit on kernels: Yes
  29. Integrated GPU sharing Host Memory: No
  30. Support host page- locked memory mapping: Yes
  31. Alignment requirement for Surfaces: Yes
  32. Device has ECC support: Disabled
  33. Device supports Unified Addressing (UVA): Yes
  34. Supports Cooperative Kernel Launch: No
  35. Supports MultiDevice Co-op Kernel Launch: No
  36. Device PCI Domain ID / Bus ID / location ID: 0 / 1 / 0
  37. Compute Mode:
  38. < Default (multiple host threads can use ::cudaSetDevice() with device simultaneously) >
  39. deviceQuery, CUDA Driver = CUDART, CUDA Driver Version = 9.1, CUDA Runtime Version = 9.1, NumDevs = 1
  40. Result = PASS

 
 
  1. $ ./bandwidthTest
  2. [CUDA Bandwidth Test] - Starting...
  3. Running on...
  4. Device 0: GeForce GTX 960M
  5. Quick Mode
  6. Host to Device Bandwidth, 1 Device(s)
  7. PINNED Memory Transfers
  8. Transfer Size (Bytes) Bandwidth(MB/s)
  9. 33554432 12339.9
  10. Device to Host Bandwidth, 1 Device(s)
  11. PINNED Memory Transfers
  12. Transfer Size (Bytes) Bandwidth(MB/s)
  13. 33554432 11720.0
  14. Device to Device Bandwidth, 1 Device(s)
  15. PINNED Memory Transfers
  16. Transfer Size (Bytes) Bandwidth(MB/s)
  17. 33554432 65699.6
  18. Result = PASS
  19. NOTE: The CUDA Samples are not meant for performance measurements. Results may vary when GPU Boost is enabled.

 

在usr/local/cuda-9.0中输入nvcc -V

会出现:


 
 
  1. yhao@yhao- X550VB:~$ nvcc -V
  2. nvcc: NVIDIA (R) Cuda compiler driver
  3. Copyright (c) 2005- 2016 NVIDIA Corporation
  4. Built on Tue_Jan_10_13: 22 : 03_CST_2017
  5. Cuda compilation tools, release 9.0, V8. 0. 61

6.tensorflow-gpu安装

在anaconda官网上下载anaconda并安装好后,输入:

pip install tensorflow-gpu==1.5.0
 
 

可以看到pip更新后,anaconda里边的库也会更新(如果是conda install 会出现问题,原因暂时没有搞清楚)

 

安装好后,可以在spyder-python中使用

import tensorflow as tf
 
 

有可能tensorflow在anaconda上不好使,这时候可以安装一个pycharm来运行tensorflow


7.cuda/cudnn版本与tensorflow版本对应关系

cuda/cudnn需要安装对应版本的tensorflow,不然会出现各种bug:

cuda8/cudnn5  --> tensorflow1.2及以下

cuda8/cudnn6  --> tensorflow1.3以及1.4

cuda9/cudnn7  --> tensorflow1.5及以上

REFERENCE:https: //blog.csdn.net/u014513863/article/details/82720866

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值