C++中双线程调用网络摄像头,用多线程调用多个摄像头并同步执行算法

          在调用网络摄像头处理自己的算法时,当解码的速度与算法运行的速度差太多时,会出现类似下面的错误

error while decoding MB 148 4, bytestream

所以需要使用两个线程,一个线程调用摄像头,一个线程用来处理图像。

一、使用双线程调用网络摄像头,并执行算法

方法一:

#include <iostream>
#include <thread>
#include <mutex>
#include <atomic>
#include <opencv2/core/core.hpp>
#include <opencv2/highgui/highgui.hpp>
#include <opencv2/imgproc/imgproc.hpp>

static std::mutex mutex;
static std::atomic_bool isOpen;

static void cameraThreadFunc(std::string camId, cv::Mat* pFrame)
{
    cv::VideoCapture capture(camId);
    //capture.set(cv::CAP_PROP_FOURCC, CV_FOURCC('M', 'J', 'P', 'G'));
    capture.set(cv::CAP_PROP_FPS, 30);

    if (!capture.isOpened()) {
        isOpen = false;
        std::cout << "Failed to open camera with index " << camId << std::endl;
    }

    cv::Mat frame;
    while (isOpen) {
        capture >> frame;
        if (mutex.try_lock()) {
            frame.copyTo(*pFrame);
            mutex.unlock();
        }
        cv::waitKey(5);
    }
    capture.release();
}


int main(int argc, char* argv[])
{
    std::string rtsps = "rtsp://admin:bs123456@192.168.1.64:554/h264/ch1/main/av_stream/1";
    isOpen = true;
    cv::Mat frame(1440, 2560, CV_8UC3);
    //cv::Mat* frame = new cv::Mat(1440, 2560, CV_8UC3);
    std::thread thread(cameraThreadFunc, rtsps, &frame); //调用摄像头
    //std::thread thread(cameraThreadFunc, rtsps, frame);

    
    //*****************************************************
    //调用自己的算法处理图像
    void* p_algorithm;
    p_algorithm = (void*)(new WindNetDetect());
    std::string net_bins = "./models/visdrone_1009-opt.bin";
    std::string net_paras = "./models/visdrone_1009-opt.param";
    int init_res = ((WindNetDetect*)p_algorithm)->init1(net_bins, net_paras);
    WindNetDetect* tmp = (WindNetDetect*)(p_algorithm);
    std::vector<Object> objects;
    //*****************************************************
    
    while (isOpen) {

        //*********************************
        //调用目标检测的算法
        tmp->detect(frame, objects);
        tmp->draw_objects(frame, objects);
        //*********************************

        if (cv::waitKey(1) == 'q') {
            break;
        }
    }

    frame.release();

    isOpen = false;
    thread.join();
    return 0;
}

 方法二:

// video_test.cpp
#include <iostream>
#include <string>
#include <thread>
#include <mutex>
#include <atomic>
#include <opencv2/core/core.hpp>
#include <opencv2/highgui/highgui.hpp>
#include <opencv2/imgproc/imgproc.hpp>

class VideoCaptureMT {
public:
	VideoCaptureMT(int index, int height=480, int width=640);
	VideoCaptureMT(std::string filePath, int height=480, int width=640);
	~VideoCaptureMT();
	
	bool isOpened() {
		return m_IsOpen;
	}
	void release() {
		m_IsOpen = false;
	}
	bool read(cv::Mat& frame);

private:
	void captureInit(int index, std::string filePath, int height, int width);
	void captureFrame();

	cv::VideoCapture* m_pCapture;
	cv::Mat* m_pFrame;
	std::mutex* m_pMutex;
	std::thread* m_pThread;
	std::atomic_bool m_IsOpen;
};

VideoCaptureMT::VideoCaptureMT(int index, int height, int width)
{
	captureInit(index, std::string(), height, width);
}

VideoCaptureMT::VideoCaptureMT(std::string filePath, int height, int width)
{
	captureInit(0, filePath, height, width);
}

VideoCaptureMT::~VideoCaptureMT()
{
	m_IsOpen = false;
	m_pThread->join();
	if (m_pCapture->isOpened()) {
		m_pCapture->release();
	}

	delete m_pThread;
	delete m_pMutex;
	delete m_pCapture;
	delete m_pFrame;
}

void VideoCaptureMT::captureInit(int index, std::string filePath, int height, int width)
{
	if (!filePath.empty()) {
		m_pCapture = new cv::VideoCapture(filePath);
	}
	else {
		m_pCapture = new cv::VideoCapture(index);
	}
	m_pCapture->set(cv::CAP_PROP_FRAME_WIDTH, width);
	m_pCapture->set(cv::CAP_PROP_FRAME_HEIGHT, height);
	m_pCapture->set(cv::CAP_PROP_FPS, 30);
	
	m_IsOpen = true;
	m_pFrame = new cv::Mat(height, width, CV_8UC3);
	m_pMutex = new std::mutex();
	m_pThread = new std::thread(&VideoCaptureMT::captureFrame, this);
}

void VideoCaptureMT::captureFrame()
{
	cv::Mat frameBuff;
	while (m_IsOpen) {
		(*m_pCapture) >> frameBuff;
		if (m_pMutex->try_lock()) {
			frameBuff.copyTo(*m_pFrame);
			m_pMutex->unlock();
		}
		cv::waitKey(5);
	}
}

bool VideoCaptureMT::read(cv::Mat& frame)
{
	if (m_pFrame->empty()) {
		m_IsOpen = false;
	}
	else {
		m_pMutex->lock();
		m_pFrame->copyTo(frame);
		m_pMutex->unlock();
	}
	return m_IsOpen;
}

int main(int argc, char* argv[])
{
	VideoCaptureMT capture(0);
	cv::Mat frame, gray;
	while (capture.isOpened()) {
		if (!capture.read(frame)) {
			break;
		}
		cv::cvtColor(frame, gray, cv::COLOR_BGR2GRAY);
		cv::blur(gray, gray, cv::Size(3, 3));
		cv::Canny(gray, gray, 5 , 38 , 3);
		cv::waitKey(100);

		cv::imshow("image", gray);
        if (cv::waitKey(5) == 'q') {
			break;
		}
	}
	capture.release();
	return 0;
}

 二、使用多线程调用多路摄像头并同步执行多个算法:

        使用两个线程调用两个函数,get_cam1中包含了调用网络摄像头的类VideoCaptureMT (见上面代码),以及调用算法。另外一个函数是一样的。代码如下 :

void get_cam1()
{
	clock_t start, end, end1, end2;


	std::string rtsps = "rtsp://admin:bs123456@192.168.1.64:554/h264/ch1/main/av_stream/1";

	VideoCaptureMT capture(rtsps);
	//VideoCaptureMT capture, captureusb;

	cv::namedWindow("外接摄像头检测", 0);
	cv::Mat frame, gray;


	while (capture.isOpened()) {
		std::lock_guard<std::mutex> mtx_locker(mtx);

		start = clock();
		if (!capture.read(frame)) {
			break;
		}
		//gray = frame;
		end = clock();
		std::vector<Object> objects;
		tmp->detect(frame, objects);


		for (size_t i = 0; i < objects.size(); i++)
		{
			const Object& obj = objects[i];
			fprintf(stderr, "%d = %.5f at %.2f %.2f %.2f x %.2f\n", obj.label, obj.prob,
				obj.rect.x, obj.rect.y, obj.rect.width, obj.rect.height);

			cv::rectangle(frame, obj.rect, cv::Scalar(255, 0, 0));

			char text[256];
			//sprintf(text, "%s %.1f%%", class_names[obj.label], obj.prob * 100);
			sprintf(text, "%s", class_names[obj.label]);

			int baseLine = 0;
			cv::Size label_size = cv::getTextSize(text, cv::FONT_HERSHEY_SIMPLEX, 0.5, 1, &baseLine);

			int x = obj.rect.x;
			int y = obj.rect.y - label_size.height - baseLine;
			if (y < 0)
				y = 0;
			if (x + label_size.width > frame.cols)
				x = frame.cols - label_size.width;

			cv::rectangle(frame, cv::Rect(cv::Point(x, y), cv::Size(label_size.width, label_size.height + baseLine)),
				cv::Scalar(255, 255, 255), -1);

			cv::putText(frame, text, cv::Point(x, y + label_size.height),
				cv::FONT_HERSHEY_SIMPLEX, 0.5, cv::Scalar(0, 0, 0));
		}

		//cv::imwrite("./result20.jpg", image);

		cv::imshow("外接摄像头检测", frame);

		end2 = clock();

		float rumtime = (float)(end2 - start) / CLOCKS_PER_SEC;
		std::stringstream buf;
		buf.precision(3);//����Ĭ�Ͼ���
		buf.setf(std::ios::fixed);//����С��λ
		buf << rumtime;
		std::string strtime;
		strtime = buf.str();
		std::cout << "strtime1111 = " << strtime << std::endl;
		//start = end2;

		if (cv::waitKey(5) == 'q') {
			break;
		}

	}
	capture.release();
	
}

void get_camusb()
{
	clock_t start, end, end1, end2;

	std::string rtsps = "rtsp://wowzaec2demo.streamlock.net/vod/mp4:BigBuckBunny_115k.mp4";
	VideoCaptureMT capture(rtsps);

	cv::Mat frameusb;
	cv::namedWindow("外接摄像头检测1", 0);

	//void* p_algorithmusb;
	//p_algorithmusb = (void*)(new WindNetDetect());
	//std::string net_binsusb = "./models/visdrone_1009-opt.bin";
	//std::string net_parasusb = "./models/visdrone_1009-opt.param";
	//int init_res = ((WindNetDetect*)p_algorithmusb)->init1(net_binsusb, net_parasusb);
	//WindNetDetect* tmp = (WindNetDetect*)(p_algorithmusb);

	while (capture.isOpened()) {
		std::lock_guard<std::mutex> mtx_locker(mtx);

		start = clock();
		if (!capture.read(frameusb)) {
			break;
		}
		//gray = frame;
		end = clock();
		std::vector<Object> objectsusb;
		tmp->detect(frameusb, objectsusb);

		for (size_t i = 0; i < objectsusb.size(); i++)
		{
			const Object& obj = objectsusb[i];
			fprintf(stderr, "%d = %.5f at %.2f %.2f %.2f x %.2f\n", obj.label, obj.prob,
				obj.rect.x, obj.rect.y, obj.rect.width, obj.rect.height);

			cv::rectangle(frameusb, obj.rect, cv::Scalar(255, 0, 0));

			char text[256];
			//sprintf(text, "%s %.1f%%", class_names[obj.label], obj.prob * 100);
			sprintf(text, "%s", class_names[obj.label]);

			int baseLine = 0;
			cv::Size label_size = cv::getTextSize(text, cv::FONT_HERSHEY_SIMPLEX, 0.5, 1, &baseLine);

			int x = obj.rect.x;
			int y = obj.rect.y - label_size.height - baseLine;
			if (y < 0)
				y = 0;
			if (x + label_size.width > frameusb.cols)
				x = frameusb.cols - label_size.width;

			cv::rectangle(frameusb, cv::Rect(cv::Point(x, y), cv::Size(label_size.width, label_size.height + baseLine)),
				cv::Scalar(255, 255, 255), -1);

			cv::putText(frameusb, text, cv::Point(x, y + label_size.height),
				cv::FONT_HERSHEY_SIMPLEX, 0.5, cv::Scalar(0, 0, 0));
		}

		//cv::imwrite("./result20.jpg", image);

		cv::imshow("外接摄像头检测1", frameusb);
		end2 = clock();
		float rumtime = (float)(end2 - start) / CLOCKS_PER_SEC;
		std::stringstream buf;
		buf.precision(3);//����Ĭ�Ͼ���
		buf.setf(std::ios::fixed);//����С��λ
		buf << rumtime;
		std::string strtime;
		strtime = buf.str();
		std::cout << "strtime1111 = " << strtime << std::endl;
		//start = end2;


		if (cv::waitKey(5) == 'q') {
			break;
		}

	}
	capture.release();
	
}

int main() {

	void* p_algorithm;
	p_algorithm = (void*)(new WindNetDetect());
	std::string net_bins = "./models/visdrone_1009-opt.bin";
	std::string net_paras = "./models/visdrone_1009-opt.param";
	int init_res = ((WindNetDetect*)p_algorithm)->init1(net_bins, net_paras);
	tmp = (WindNetDetect*)(p_algorithm);

	//HANDLE hThread1 = CreateThread(NULL, 0, get_cam1, NULL, 0, NULL);
	//CloseHandle(hThread1);
	//HANDLE hThread2 = CreateThread(NULL, 0, get_camusb, NULL, 0, NULL);
	//CloseHandle(hThread2);

	std::thread thrd_1(get_cam1);
	std::thread thrd_2(get_camusb);
	thrd_1.join();
	thrd_2.join();

	while (true)
	{
		std::cout << "Main Thread Display!" << std::endl;
		Sleep(3000);
	}
	return 0;
}

运行上面的代码时,两个函数没有同步运行。如下strtime1111表示一个线程调用的函数,strtime2222表示另一个线程调用的函数,能看出没有同时调用两个函数。

 所以需对代码进行如下改进:

static std::mutex mutexk;
void get_cam1()
{
	clock_t start, end, end1, end2;
	std::string rtsps = "rtsp://admin:bs123456@192.168.1.64:554/h264/ch1/main/av_stream/1";
	VideoCaptureMT capture(rtsps);
	cv::Mat frame;
	cv::namedWindow("ss", 0);

	while (capture.isOpened()) {
		//pthread_rwlock_rdlock(&rwlock);
		//std::lock_guard<std::mutex> mtx_locker(mtx);

		start = clock();
		if (!capture.read(frame)) {
			break;
		}
		if (mutexk.try_lock()) {
			//gray = frame;
			end = clock();
			std::vector<Object> objects;
			tmp->detect(frame, objects);

			//tmp->draw_objects(frame, objects);
			for (size_t i = 0; i < objects.size(); i++)
			{
				const Object& obj = objects[i];
				fprintf(stderr, "%d = %.5f at %.2f %.2f %.2f x %.2f\n", obj.label, obj.prob,
					obj.rect.x, obj.rect.y, obj.rect.width, obj.rect.height);

				cv::rectangle(frame, obj.rect, cv::Scalar(255, 0, 0));

				char text[256];
				//sprintf(text, "%s %.1f%%", class_names[obj.label], obj.prob * 100);
				sprintf(text, "%s", class_names[obj.label]);

				int baseLine = 0;
				cv::Size label_size = cv::getTextSize(text, cv::FONT_HERSHEY_SIMPLEX, 0.5, 1, &baseLine);

				int x = obj.rect.x;
				int y = obj.rect.y - label_size.height - baseLine;
				if (y < 0)
					y = 0;
				if (x + label_size.width > frame.cols)
					x = frame.cols - label_size.width;

				cv::rectangle(frame, cv::Rect(cv::Point(x, y), cv::Size(label_size.width, label_size.height + baseLine)),
					cv::Scalar(255, 255, 255), -1);

				cv::putText(frame, text, cv::Point(x, y + label_size.height),
					cv::FONT_HERSHEY_SIMPLEX, 0.5, cv::Scalar(0, 0, 0));
			}
			//cv::imwrite("./result20.jpg", image);
			cv::imshow("ss", frame);

			end2 = clock();
			float rumtime = (float)(end2 - start) / CLOCKS_PER_SEC;
			std::stringstream buf;
			buf.precision(3);//����Ĭ�Ͼ���
			buf.setf(std::ios::fixed);//����С��λ
			buf << rumtime;
			std::string strtime;
			strtime = buf.str();
			std::cout << "strtime2222 = " << strtime << std::endl;
			//start = end2;

			if (cv::waitKey(5) == 'q') {
				break;
			}
			mutexk.unlock();
		}

	}
	capture.release();
}

void get_camusb()
{
	//std::lock_guard<std::mutex> mtx_locker(mtx);
	clock_t start, end, end1, end2;

	//std::string rtsps = "rtsp://wowzaec2demo.streamlock.net/vod/mp4:BigBuckBunny_115k.mp4";
	std::string rtsps = "rtsp://admin:bs123456@192.168.1.64:554/h264/ch1/main/av_stream/1";

	VideoCaptureMT capture(rtsps);

	cv::Mat frameusb;
	cv::namedWindow("zz1", 0);

	while (capture.isOpened()) {
		std::lock_guard<std::mutex> mtx_locker1(mtx);

		start = clock();
		if (!capture.read(frameusb)) {
			break;
		}
		if (mutexk.try_lock()) {
			//gray = frame;
			end = clock();
			std::vector<Object> objectsusb;
			tmp->detect(frameusb, objectsusb);

			//tmp->draw_objects(frameusb, objectsusb);

			for (size_t i = 0; i < objectsusb.size(); i++)
			{
				const Object& obj = objectsusb[i];
				fprintf(stderr, "%d = %.5f at %.2f %.2f %.2f x %.2f\n", obj.label, obj.prob,
					obj.rect.x, obj.rect.y, obj.rect.width, obj.rect.height);

				cv::rectangle(frameusb, obj.rect, cv::Scalar(255, 0, 0));

				char text[256];
				//sprintf(text, "%s %.1f%%", class_names[obj.label], obj.prob * 100);
				sprintf(text, "%s", class_names[obj.label]);

				int baseLine = 0;
				cv::Size label_size = cv::getTextSize(text, cv::FONT_HERSHEY_SIMPLEX, 0.5, 1, &baseLine);

				int x = obj.rect.x;
				int y = obj.rect.y - label_size.height - baseLine;
				if (y < 0)
					y = 0;
				if (x + label_size.width > frameusb.cols)
					x = frameusb.cols - label_size.width;

				cv::rectangle(frameusb, cv::Rect(cv::Point(x, y), cv::Size(label_size.width, label_size.height + baseLine)),
					cv::Scalar(255, 255, 255), -1);

				cv::putText(frameusb, text, cv::Point(x, y + label_size.height),
					cv::FONT_HERSHEY_SIMPLEX, 0.5, cv::Scalar(0, 0, 0));
			}

			//cv::imwrite("./result20.jpg", image);

			cv::imshow("zz1", frameusb);

			end2 = clock();

			float rumtime = (float)(end2 - start) / CLOCKS_PER_SEC;
			std::stringstream buf;
			buf.precision(3);//����Ĭ�Ͼ���
			buf.setf(std::ios::fixed);//����С��λ
			buf << rumtime;
			std::string strtime;
			strtime = buf.str();
			std::cout << "strtime1111 = " << strtime << std::endl;
			//start = end2;


			if (cv::waitKey(5) == 'q') {
				break;
			}
			mutexk.unlock();
		}
		
	}
	capture.release();

}


int main() {
	

	void* p_algorithm;
	p_algorithm = (void*)(new WindNetDetect());
	std::string net_bins = "./models/visdrone_1009-opt.bin";
	std::string net_paras = "./models/visdrone_1009-opt.param";
	int init_res = ((WindNetDetect*)p_algorithm)->init1(net_bins, net_paras);
	tmp = (WindNetDetect*)(p_algorithm);

	std::thread thrd_1(get_cam1);
	std::thread thrd_2(get_camusb);
	thrd_1.join();
	thrd_2.join();

	while (true)
	{
		std::cout << "Main Thread Display!" << std::endl;
		Sleep(3000);
	}
	return 0;
}

         使用多线线程调用多个函数同步执行模板如下:

#include <thread>
#include <mutex>
static std::mutex mutexk;
void func1() {
	while (capture1.isOpened()) {
		if (mutexk.try_lock()) {
			//我的算法
			my_func1();
            mutexk.unlock();
		}
		
	}
}
void func2() {
	while (capture2.isOpened()) {
		if (mutexk.try_lock()) {
			//我的算法
			my_func2();
            mutexk.unlock();
		}
		
	}
}

int main() {
	std::thread thrd_1(func1);
	std::thread thrd_2(func2);
	thrd_1.join();
	thrd_2.join();
	return 0;
}

在ubuntu中c++opencv多线程显示图片时出现Gdk-WARNING **: xx: Fatal IO error

 解决办法:

1、sudo apt install libx11-dev

2、在代码中添加头文件#include <X11/Xlib.h>

3、在创建线程之前添加XInitThreads()

4、在编译的时候需要添加-lX11 

参考链接:OpenCV视频流的C++多线程处理方式_hlld26的博客-CSDN博客_opencv 多线程

  • 1
    点赞
  • 22
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
要在PyQt中使用多线程打开多个摄像头,可以使用Python线程OpenCV库。以下是一个简单的示例代码,可以同时打开两个USB摄像头: ```python import cv2 import sys from PyQt5.QtCore import QThread, QTimer from PyQt5.QtGui import QImage, QPixmap from PyQt5.QtWidgets import QApplication, QDialog, QLabel class CameraThread(QThread): def __init__(self, camera_id): super().__init__() self.camera_id = camera_id self.image = None def run(self): cap = cv2.VideoCapture(self.camera_id) while True: ret, frame = cap.read() if ret: self.image = frame class CameraWidget(QLabel): def __init__(self, camera_id): super().__init__() self.camera_thread = CameraThread(camera_id) self.camera_thread.start() self.timer = QTimer(self) self.timer.timeout.connect(self.update_frame) self.timer.start(50) def update_frame(self): if self.camera_thread.image is not None: image = QImage(self.camera_thread.image.data, self.camera_thread.image.shape[1], self.camera_thread.image.shape[0], QImage.Format_RGB888) pixmap = QPixmap.fromImage(image.rgbSwapped()) self.setPixmap(pixmap) class MainWindow(QDialog): def __init__(self): super().__init__() self.setWindowTitle("Multiple Cameras") self.camera1 = CameraWidget(0) self.camera2 = CameraWidget(1) layout = QVBoxLayout() layout.addWidget(self.camera1) layout.addWidget(self.camera2) self.setLayout(layout) if __name__ == "__main__": app = QApplication(sys.argv) window = MainWindow() window.show() sys.exit(app.exec_()) ``` 此示例代码中,CameraThread类定义了一个线程,用于从摄像头读取图像。CameraWidget类继承自QLabel,用于显示摄像头图像,并且使用QTimer定时更新图像。MainWindow类则是一个简单的窗口,包含两个CameraWidget实例。 在这个示例中,我们使用OpenCV库读取摄像头图像,然后将其转换为QImage和QPixmap格式,以便在PyQt窗口中显示。每个摄像头都在自己的线程中运行,以避免阻塞主线程

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值