【模型部署二】C++使用onnxruntime对pix2pix的onnx模型进行推理

1. 环境配置

VS2019/Ubuntu配置Onnxruntime、Opencv:
https://blog.csdn.net/qq_44747572/article/details/121340735?spm=1001.2014.3001.5501

Libtorch + vs 2019安装及配置:
https://blog.csdn.net/qq_44747572/article/details/121510739

2. Pix2Pix的Onnx模型

链接:https://pan.baidu.com/s/1vJ97qj2ieJKpFe-3DvwsSA
提取码:y961

3. Onnx模型推理

#include<opencv2/opencv.hpp>
#include<opencv2/imgcodecs/imgcodecs.hpp>
#include<opencv2/highgui/highgui.hpp>
#include<opencv2/core/core.hpp>
#include<opencv2/imgproc/imgproc_c.h>
#include<iostream>
#include<string>
#include<torch/torch.h>
#include<io.h>
#include<onnxruntime_cxx_api.h>



torch::Tensor to_tensor(cv::Mat img) {
	torch::Tensor img_tensor = torch::from_blob(img.data, { img.rows, img.cols, 3 }, torch::kByte);  // torch::kByte要对应cv::Mat img的数据类型
	img_tensor = img_tensor.permute({ 2, 0, 1 }); //转为C*H*W
	img_tensor = img_tensor.toType(torch::kFloat32);//转为
	img_tensor = img_tensor.div(255);
	return img_tensor;
}

torch::Tensor normalize(torch::Tensor tensor) {
	std::vector<double> mean = { 0.5,0.5,0.5 };
	std::vector<double> std = { 0.5,0.5,0.5 };
	tensor = torch::data::transforms::Normalize<>(mean, std)(tensor);// opencv
	return tensor;
}

int main()
{
	std::cout << "预处理" << std::endl;
	std::clock_t startTime_pre, endTime_pre;
	startTime_pre = clock();
	std::vector<cv::Mat> images = { //读取图片
		cv::imread("F:\\Pycharm\\PyCharm_Study\\Others\\c++_learning\\C++_Master\\Onnx\\pix2pixHD\\cpp-onnxruntime\\A_01425.png") ,
	};
	std::vector<float> inputVec;
	int batchSize = 1, channels, height, width;
	for (cv::Mat img : images) {
		cv::cvtColor(img, img, cv::COLOR_BGR2RGB);
		int w = img.cols;
		int h = img.rows;
		torch::Tensor tensor_A = to_tensor(img);//转为tensor,这样可以直接用归一化的api,方便对输入进行预处理
		torch::Tensor label = normalize(tensor_A);//归一化
		channels = label.sizes()[0], height = label.sizes()[1], width = label.sizes()[2];

		cv::Mat resultImg(h, w, CV_32FC3);
		/*std::memcpy 行中sizeof()中内容,需要修改成c++中内建的数据类型,如果使用torch::kF32或者其他浮点型,会出现数据复制缺失的情况。*/
		std::memcpy((void*)resultImg.data, label.data_ptr(), sizeof(float) * label.numel());
		//构造模型输入std::vector<float> inputVec
		std::vector<cv::Mat> channels;
		cv::split(resultImg, channels);
		cv::Mat blue, green, red;
		blue = channels.at(0);
		green = channels.at(1);
		red = channels.at(2);
		std::vector<float> inputVec_red = (std::vector<float>)(blue.reshape(1, 1));
		std::vector<float> inputVec_green = (std::vector<float>)(green.reshape(1, 1));
		std::vector<float> inputVec_blue = (std::vector<float>)(red.reshape(1, 1));
		inputVec.insert(inputVec.end(), inputVec_red.begin(), inputVec_red.end());
		inputVec.insert(inputVec.end(), inputVec_green.begin(), inputVec_green.end());
		inputVec.insert(inputVec.end(), inputVec_blue.begin(), inputVec_blue.end());
	}

	endTime_pre = clock();
	std::cout << "预处理时间:" << (double)(endTime_pre - startTime_pre) / CLOCKS_PER_SEC << "s" << std::endl;
	std::cout << "预处理结束" << std::endl;

	
	//设置为VERBOSE,方便控制台输出时看到是使用了cpu还是gpu执行
	Ort::Env env(ORT_LOGGING_LEVEL_VERBOSE, "OnnxPix2PixHD");
	Ort::SessionOptions session_options;
	// 使用五个线程执行op,提升速度
	session_options.SetIntraOpNumThreads(1);
	 第二个参数代表GPU device_id = 0,注释这行就是cpu执行
	//OrtSessionOptionsAppendExecutionProvider_CUDA(session_options, 0);     
	// ORT_ENABLE_ALL: To Enable All possible opitmizations
	session_options.SetGraphOptimizationLevel(GraphOptimizationLevel::ORT_ENABLE_ALL);

	Ort::Session session(env, L"F:\\Pycharm\\PyCharm_Study\\Others\\c++_learning\\C++_Master\\Onnx\\pix2pixHD\\pix2pixHD_Cartoon_batch1.onnx", session_options);
	std::cout << "onnxruntime loading onnx model..." << std::endl;
	Ort::AllocatorWithDefaultOptions allocator;


	std::vector<const char*> inputNames = { "inputs" };
	std::vector<const char*> outputNames = { "outputs" };

	std::vector<int64_t> inputSize = { batchSize, channels, height, width };
	rsize_t inputSizeCount = batchSize * channels * height * width;

	auto memoryInfo = Ort::MemoryInfo::CreateCpu(OrtArenaAllocator, OrtMemType::OrtMemTypeDefault);
	std::clock_t startTime, endTime;
	Ort::Value inputTensor = Ort::Value::CreateTensor<float>(memoryInfo, inputVec.data(), inputSizeCount, inputSize.data(), inputSize.size());
	startTime = clock();
	std::vector<Ort::Value>  outputTensors = session.Run(Ort::RunOptions{ nullptr }, inputNames.data(), &inputTensor, inputNames.size(), outputNames.data(), outputNames.size());
	endTime = clock();

	std::cout << "加速时间:" << (double)(endTime - startTime) / CLOCKS_PER_SEC << "s" << std::endl;
	float* output = outputTensors[0].GetTensorMutableData<float>();

	std::vector<cv::Mat> results;
	std::cout << "保存结果" << std::endl;
	std::clock_t startTime_end, endTime_end;
	startTime_end = clock();
	int idx = 0;
	for (int b = 0; b < batchSize; b++) {

		torch::Tensor result_r = torch::from_blob(output, { height,width,1 });
		torch::Tensor result_g = torch::from_blob(&output[height * width - 1], { height,width,1 });
		torch::Tensor result_b = torch::from_blob(&output[height * width * 2 - 1], { height,width,1 });
		torch::Tensor result = torch::cat({ result_r, result_g, result_b }, 2);
		result = result.add(1).div(2).mul(255);
		result = result.clamp(0, 255);
		result = result.to(torch::kU8);
		cv::Mat resultImg(height, width, CV_8UC3);
		std::memcpy((void*)resultImg.data, result.data_ptr(), sizeof(torch::kU8) * result.numel());
		cv::cvtColor(resultImg, resultImg, cv::COLOR_RGB2BGR);
		results.push_back(resultImg);

	}
	cv::imwrite("F:\\Pycharm\\PyCharm_Study\\Others\\c++_learning\\C++_Master\\Onnx\\pix2pixHD\\cpp-onnxruntime\\A_Cartoon.png", results[0]);
	endTime_end = clock();

	std::cout << "后处理时间:" << (double)(endTime_end - startTime_end) / CLOCKS_PER_SEC << "s" << std::endl;
	std::cout << "运行结束" << std::endl;

	printf("Done!\n");
	system("pause");
	return 0;
}
  • 1
    点赞
  • 16
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 21
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论 21
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

只搬烫手的砖

你的鼓励将是我最大的动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值