win10+CUDA11.0+cuDNN8.0+anaconda3+python3.8+tensorflow_gpu 2.4.0+jupyter notebook 踩坑实录

本文详细记录了在Windows 10上安装CUDA 11.0、cuDNN 8.0、Anaconda3、Python 3.8以及TensorFlow-GPU 2.4.0的过程,并遇到的问题,如Python和iPython中导入TensorFlow失败(报错No module named 'gast'),以及在Jupyter Notebook中添加内核的步骤。此外,还介绍了如何删除和重新配置虚拟环境,以及在安装过程中对其他库如keras、sklearn和imblearn的依赖处理。
摘要由CSDN通过智能技术生成


安装cuda,cudnn相对应的版本后,tensorflow-gpu也得对应版本。这里记录配置tensorflow-gpu2.4.0环境和在jupyter notebook调用的踩坑实录。CUDA 和 cuDNN安装后只要版本对了就行。重装啥的还是看tensorflow。

CUDA,cuDNN下载安装

参见 https://blog.csdn.net/Qi__Xi/article/details/127127629
注意【版本匹配】,我这里是CUDA11.0+cuDNN8.0+tensorflow_gpu 2.4.0
在这里插入图片描述

配置tensorflow-gpu

运行anaconda prompt (用管理员)

// 在anaconda下创建虚拟环境,命名为tf240
conda create -n tf240 python=3.8 
//激活tf2.4环境
conda activate tf240
//安装tensorflow-gpu 2.4.0版本
pip install tensorflow-gpu==2.4.0

至此,可以试试能不能成功

// 在虚拟环境 tf240 下,进入python
python 
//进入pyth
自编译tensorflow: 1.python3.5,tensorflow1.12; 2.支持cuda10.0,cudnn7.3.1,TensorRT-5.0.2.6-cuda10.0-cudnn7.3; 3.支持mkl,无MPI; 软硬件硬件环境:Ubuntu16.04,GeForce GTX 1080 配置信息: hp@dla:~/work/ts_compile/tensorflow$ ./configure WARNING: --batch mode is deprecated. Please instead explicitly shut down your Bazel server using the command "bazel shutdown". You have bazel 0.19.1 installed. Please specify the location of python. [Default is /usr/bin/python]: /usr/bin/python3 Found possible Python library paths: /usr/local/lib/python3.5/dist-packages /usr/lib/python3/dist-packages Please input the desired Python library path to use. Default is [/usr/local/lib/python3.5/dist-packages] Do you wish to build TensorFlow with XLA JIT support? [Y/n]: XLA JIT support will be enabled for TensorFlow. Do you wish to build TensorFlow with OpenCL SYCL support? [y/N]: No OpenCL SYCL support will be enabled for TensorFlow. Do you wish to build TensorFlow with ROCm support? [y/N]: No ROCm support will be enabled for TensorFlow. Do you wish to build TensorFlow with CUDA support? [y/N]: y CUDA support will be enabled for TensorFlow. Please specify the CUDA SDK version you want to use. [Leave empty to default to CUDA 10.0]: Please specify the location where CUDA 10.0 toolkit is installed. Refer to README.md for more details. [Default is /usr/local/cuda]: /usr/local/cuda-10.0 Please specify the cuDNN version you want to use. [Leave empty to default to cuDNN 7]: 7.3.1 Please specify the location where cuDNN 7 library is installed. Refer to README.md for more details. [Default is /usr/local/cuda-10.0]: Do you wish to build TensorFlow with TensorRT support? [y/N]: y TensorRT support will be enabled for TensorFlow. Please specify the location where TensorRT is installed. [Default is /usr/lib/x86_64-linux-gnu]:/home/hp/bin/TensorRT-5.0.2.6-cuda10.0-cudnn7.3/targets/x86_64-linux-gnu Please specify the locally installed NCCL version you want to use. [Default is to use https://github.com/nvidia/nccl]: Please specify a list of comma-separated Cuda compute capabilities you want to build with. You can find the compute capability of your device at: https://developer.nvidia.com/cuda-gpus. Please note that each additional compute capability significantly increases your build time and binary size. [Default is: 6.1,6.1,6.1]: Do you want to use clang as CUDA compiler? [y/N]: nvcc will be used as CUDA compiler. Please specify which gcc should be used by nvcc as the host compiler. [Default is /usr/bin/gcc]: Do you wish to build TensorFlow with MPI support? [y/N]: No MPI support will be enabled for TensorFlow. Please specify optimization flags to use during compilation when bazel option "--config=opt" is specified [Default is -march=native -Wno-sign-compare]: Would you like to interactively configure ./WORKSPACE for Android builds? [y/N]: Not configuring the WORKSPACE for Android builds. Preconfigured Bazel build configs. You can use any of the below by adding "--config=" to your build command. See .bazelrc for more details. --config=mkl # Build with MKL support. --config=monolithic # Config for mostly static monolithic build. --config=gdr # Build with GDR support. --config=verbs # Build with libverbs support. --config=ngraph # Build with Intel nGraph support. --config=dynamic_kernels # (Experimental) Build kernels into separate shared objects. Preconfigured Bazel build configs to DISABLE default on features: --config=noaws # Disable AWS S3 filesystem support. --config=nogcp # Disable GCP support. --config=nohdfs # Disable HDFS support. --config=noignite # Disable Apacha Ignite support. --config=nokafka # Disable Apache Kafka support. --config=nonccl # Disable NVIDIA NCCL support. Configuration finished 编译: hp@dla:~/work/ts_compile/tensorflow$ bazel build --config=opt --config=mkl --verbose_failures //tensorflow/tools/pip_package:build_pip_package 卸载已有tensorflow: hp@dla:~/temp$ sudo pip3 uninstall tensorflow 安装自己编译的成果: hp@dla:~/temp$ sudo pip3 install tensorflow-1.12.0-cp35-cp35m-linux_x86_64.whl
NVIDIA® cuDNN is a GPU-accelerated library of primitives for deep neural networks. cuDNN是一个对DNN的GPU加速库。他提供高度可调整的在DNN中的常用的例程实现。 It provides highly tuned implementations of routines arising frequently in DNN applications: 常用语前向后向卷积网络,包括交叉相关。Convolution forward and backward, including cross-correlation 前像后向pooling。Pooling forward and backward 前向后向softmax。Softmax forward and backward 前向后向神经元激活。Neuron activations forward and backward Rectified linear (ReLU) Hyperbolic tangent (TANH) Tensor transformation functions LRN, LCN and batch normalization forward and backward cuDNN’s convolution routines aim for performance competitive with the fastest GEMM (matrix multiply) based implementations of such routines while using significantly less memory. cuDNN突出可定制的数据布局,支持灵活的维数排序,跨步,4D子区域for 4D张量作为输入输出。 cuDNN features customizable data layouts, supporting flexible dimension ordering, striding, and subregions for the 4D tensors used as inputs and outputs to all of its routines. This flexibility allows easy integration into any neural network implementation and avoids the input/output transposition steps sometimes necessary with GEMM-based convolutions. cuDNN offers a context-based API that allows for easy multithreading and (optional) interoperability with CUDA streams. cuDNN提供一种基于上下文的API,允许简单的多线程和CDUA流的互用。
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值