CUDA中使用多个流并行执行数据复制和核函数运算可以进一步提高计算性能。以下程序使用2个流执行运算:
- #include "cuda_runtime.h"
- #include <iostream>
- #include <stdio.h>
- #include <math.h>
- #define N (1024*1024)
- #define FULL_DATA_SIZE N*20
- __global__ void kernel(int* a, int *b, int*c)
- {
- int threadID = blockIdx.x * blockDim.x + threadIdx.x;
- if (threadID < N)
- {
- c[threadID] = (a[threadID] + b[threadID]) / 2;
- }
- }
- int main()
- {
- //获取设备属性
- cudaDeviceProp prop;
- int deviceID;
- cudaGetDevice(&deviceID);
- cudaGetDeviceProperties(&prop, deviceID);
- //检查设备是否支持重叠功能
- if (!prop.deviceOverlap)
- {
- printf("No device will handle overlaps. so no speed up from stream.\n");
- return 0;
- }
- //启动计时器
- cudaEvent_t start, stop;
- float elapsedTime;
- cudaEventCreate(&start);
- cudaEventCreate(&stop);
- cudaEventRecord(start, 0);
- //创建两个CUDA流
- cudaStream_t stream, stream1;
- cudaStreamCreate(&stream);
- cudaStreamCreate(&stream1);
- int *host_a, *host_b, *host_c;
- int *dev_a, *dev_b, *dev_c;
- int *dev_a1, *dev_b1, *dev_c1;
- //在GPU上分配内存
- cudaMalloc((void**)&dev_a, N * sizeof(int));
- cudaMalloc((void**)&dev_b, N * sizeof(int));
- cudaMalloc((void**)&dev_c, N * sizeof(int));
- cudaMalloc((void**)&dev_a1, N * sizeof(int));
- cudaMalloc((void**)&dev_b1, N * sizeof(int));
- cudaMalloc((void**)&dev_c1, N * sizeof(int));
- //在CPU上分配页锁定内存
- cudaHostAlloc((void**)&host_a, FULL_DATA_SIZE * sizeof(int), cudaHostAllocDefault);
- cudaHostAlloc((void**)&host_b, FULL_DATA_SIZE * sizeof(int), cudaHostAllocDefault);
- cudaHostAlloc((void**)&host_c, FULL_DATA_SIZE * sizeof(int), cudaHostAllocDefault);
- //主机上的内存赋值
- for (int i = 0; i < FULL_DATA_SIZE; i++)
- {
- host_a[i] = i;
- host_b[i] = FULL_DATA_SIZE - i;
- }
- for (int i = 0; i < FULL_DATA_SIZE; i += 2 * N)
- {
- cudaMemcpyAsync(dev_a, host_a + i, N * sizeof(int), cudaMemcpyHostToDevice, stream);
- cudaMemcpyAsync(dev_b, host_b + i, N * sizeof(int), cudaMemcpyHostToDevice, stream);
- cudaMemcpyAsync(dev_a1, host_a + i + N, N * sizeof(int), cudaMemcpyHostToDevice, stream1);
- cudaMemcpyAsync(dev_b1, host_b + i + N, N * sizeof(int), cudaMemcpyHostToDevice, stream1);
- kernel << <N / 1024, 1024, 0, stream >> > (dev_a, dev_b, dev_c);
- kernel << <N / 1024, 1024, 0, stream1 >> > (dev_a, dev_b, dev_c1);
- cudaMemcpyAsync(host_c + i, dev_c, N * sizeof(int), cudaMemcpyDeviceToHost, stream);
- cudaMemcpyAsync(host_c + i + N, dev_c1, N * sizeof(int), cudaMemcpyDeviceToHost, stream1);
- }
- // 等待Stream流执行完成
- cudaStreamSynchronize(stream);
- cudaStreamSynchronize(stream1);
- cudaEventRecord(stop, 0);
- cudaEventSynchronize(stop);
- cudaEventElapsedTime(&elapsedTime, start, stop);
- std::cout << "消耗时间: " << elapsedTime << std::endl;
- //输出前10个结果
- for (int i = 0; i < 10; i++)
- {
- std::cout << host_c[i] << std::endl;
- }
- getchar();
- // free stream and mem
- cudaFreeHost(host_a);
- cudaFreeHost(host_b);
- cudaFreeHost(host_c);
- cudaFree(dev_a);
- cudaFree(dev_b);
- cudaFree(dev_c);
- cudaFree(dev_a1);
- cudaFree(dev_b1);
- cudaFree(dev_c1);
- cudaStreamDestroy(stream);
- cudaStreamDestroy(stream1);
- return 0;
- }
使用2个流,执行时间16ms,基本上是使用一个流消耗时间的二分之一。