D:\opencv\Videoplay\src\videotortsp
#include <opencv2/highgui.hpp>
#include <iostream>
extern "C"
{
#include <libswscale/swscale.h>
#include <libavcodec/avcodec.h>
#include <libavformat/avformat.h>
}
#pragma comment(lib, "swscale.lib")
#pragma comment(lib, "avcodec.lib")
#pragma comment(lib, "avutil.lib")
#pragma comment(lib, "avformat.lib")
#pragma comment(lib,"opencv_world455d.lib")
using namespace std;
using namespace cv;
int main(int argc, char* argv[])
{
//nginx-rtmp 直播服务器rtmp推流URL
const char* outUrl = "rtsp://127.0.0.1:554/live/camera0";
//注册所有网络协议
avformat_network_init();
VideoCapture cam;
Mat frame;
//namedWindow("video");
//像素格式转换上下文
SwsContext* vsc = NULL;
//输出的数据结构
AVFrame* yuv = NULL;
//编码器上下文
AVCodecContext* vc = NULL;
//rtmp flv 封装器
AVFormatContext* ic = NULL;
try
{
/// 1 使用opencv打开rtsp相机
cam.open(0);
if (!cam.isOpened())
{
throw exception("cam open failed!");
}
cout << " cam open success" << endl;
int inWidth = cam.get(CAP_PROP_FRAME_WIDTH);
int inHeight = cam.get(CAP_PROP_FRAME_HEIGHT);
int fps = cam.get(CAP_PROP_FPS);
///2 初始化格式转换上下文
vsc = sws_getCachedContext(vsc,
inWidth, inHeight, AV_PIX_FMT_BGR24, //源宽、高、像素格式
inWidth, inHeight, AV_PIX_FMT_YUV420P, //目标宽、高、像素格式
SWS_BICUBIC, // 尺寸变化使用算法
0, 0, 0
);
if (!vsc)
{
throw exception("sws_getCachedContext failed!");
}
///3 初始化输出的数据结构
yuv = av_frame_alloc();
yuv->format = AV_PIX_FMT_YUV420P;
yuv->width = inWidth;
yuv->height = inHeight;
yuv->pts = 0;
//分配yuv空间
int ret = av_frame_get_buffer(yuv, 32);
if (ret != 0)
{
char buf[1024] = { 0 };
av_strerror(ret, buf, sizeof(buf) - 1);
throw exception(buf);
}
//4 初始化编码上下文
//a 找到编码器
const AVCodec* codec = avcodec_find_encoder(AV_CODEC_ID_H264);
if (!codec)
{
throw exception("Can`t find h264 encoder!");
}
//b 创建编码器上下文
vc = avcodec_alloc_context3(codec);
if (!vc)
{
throw exception("avcodec_alloc_context3 failed!");
}
//c 配置编码器参数
vc->flags |= AV_CODEC_FLAG_GLOBAL_HEADER; //全局参数
vc->codec_id = codec->id;
vc->thread_count = 8;
vc->bit_rate = 50 * 1024 * 8;//压缩后每秒视频的bit位大小 50kB
vc->width = inWidth;
vc->height = inHeight;
vc->time_base = { 1,fps };
vc->framerate = { fps,1 };
//画面组的大小,多少帧一个关键帧
vc->gop_size = 50;
// B 帧数量设置为0 ,不用考虑解码顺序
vc->max_b_frames = 0;
vc->pix_fmt = AV_PIX_FMT_YUV420P;
AVDictionary* video_codec_options = NULL;
av_dict_set(&video_codec_options, "preset", "superfast", 0);
//av_dict_set(&video_codec_options, "preset", "ultrafast", 0);
av_dict_set(&video_codec_options, "tune", "zerolatency", 0);
//d 打开编码器上下文
ret = avcodec_open2(vc, codec, &video_codec_options);
if (ret != 0)
{
char buf[1024] = { 0 };
av_strerror(ret, buf, sizeof(buf) - 1);
throw exception(buf);
}
cout << "avcodec_open2 success!" << endl;
///5 输出封装器和视频流配置
//a 创建输出封装器上下文
// ret = avformat_alloc_output_context2(&ic, 0, "flv", outUrl);
ret = avformat_alloc_output_context2(&ic, 0, "rtsp", outUrl);
if (ret != 0)
{
char buf[1024] = { 0 };
av_strerror(ret, buf, sizeof(buf) - 1);
throw exception(buf);
}
//b 添加视频流
AVStream* vs = avformat_new_stream(ic, NULL);
if (!vs)
{
throw exception("avformat_new_stream failed");
}
vs->codecpar->codec_tag = 0;
//从编码器复制参数
avcodec_parameters_from_context(vs->codecpar, vc);
av_dump_format(ic, 0, outUrl, 1);
///打开rtmp 的网络输出IO
//ret = avio_open(&ic->pb, outUrl, AVIO_FLAG_WRITE);
//if (ret != 0)
//{
// char buf[1024] = { 0 };
// av_strerror(ret, buf, sizeof(buf) - 1);
// throw exception(buf);
//}
AVDictionary* fmt_options = NULL;
//av_dict_set(&fmt_options, "bufsize", "1024", 0);
av_dict_set(&fmt_options, "rw_timeout", "30000000", 0); //设置rtmp/http-flv连接超时(单位 us)
av_dict_set(&fmt_options, "stimeout", "30000000", 0); //设置rtsp连接超时(单位 us)
av_dict_set(&fmt_options, "rtsp_transport", "tcp", 0);
//av_dict_set(&fmt_options, "rtsp_transport", "udp", 0);
//写入封装头
ret = avformat_write_header(ic, &fmt_options);
if (ret != 0)
{
char buf[1024] = { 0 };
av_strerror(ret, buf, sizeof(buf) - 1);
throw exception(buf);
}
AVPacket pack;
memset(&pack, 0, sizeof(pack));
int vpts = 0;
for (;;)
{
///读取rtsp视频帧,解码视频帧
if (!cam.grab())
{
continue;
}
///yuv转换为rgb
// 视频帧存储在 frame 变量中。这个帧通常是BGR格式的图像帧
if (!cam.retrieve(frame))
{
continue;
}
//imshow("video", frame);
//waitKey(1);
///rgb to yuv
//输入的数据结构
uint8_t* indata[AV_NUM_DATA_POINTERS] = { 0 };
//indata[0] bgrbgrbgr
//plane indata[0] bbbbb indata[1]ggggg indata[2]rrrrr
indata[0] = frame.data;
int insize[AV_NUM_DATA_POINTERS] = { 0 };
//一行(宽)数据的字节数
insize[0] = frame.cols * frame.elemSize();
int h = sws_scale(vsc, indata, insize, 0, frame.rows, //源数据
yuv->data, yuv->linesize);
if (h <= 0)
{
continue;
}
///h264编码
yuv->pts = vpts;
vpts++;
//将帧送入编码器
ret = avcodec_send_frame(vc, yuv);
if (ret != 0)
continue;
//将帧送入编码器
ret = avcodec_receive_packet(vc, &pack);
if (ret != 0 || pack.size > 0)
{
//cout << "*" << pack.size << flush;
}
else
{
continue;
}
//推流
pack.pts = av_rescale_q(pack.pts, vc->time_base, vs->time_base);
pack.dts = av_rescale_q(pack.dts, vc->time_base, vs->time_base);
pack.duration = av_rescale_q(pack.duration, vc->time_base, vs->time_base);
ret = av_interleaved_write_frame(ic, &pack);
}
}
catch (exception& ex)
{
if (cam.isOpened())
cam.release();
if (vsc)
{
sws_freeContext(vsc);
vsc = NULL;
}
if (vc)
{
avio_closep(&ic->pb);
avcodec_free_context(&vc);
}
cerr << ex.what() << endl;
}
getchar();
return 0;
}
通过ffmpeg实现拉流+转码+推流(优化后,延时在900毫秒左右)
ffmpeg -rtsp_transport udp -i srcUrl -fflags nobuffer -max_delay 1 -threads 5 -profile:v high -preset superfast -tune zerolatency -an -c:v h264 -crf 25 -s 1280*720 -f rtsp -bf 0 -g 5 -rtsp_transport udp dstUrl
ffmpeg -rtsp_transport tcp -i rtsp://127.0.0.1:554/live/camera -fflags nobuffer -max_delay 1 -threads 5 -profile:v high -preset superfast -tune zerolatency -an -c:v h264 -crf 25 -s 1280*720 -f rtsp -bf 0 -g 5 -rtsp_transport tcp rtsp://127.0.0.1:554/live/camera2
ffmpeg -i camera_60s.mp4 -vf "scale=1280:720:force_original_aspect_ratio=decrease,pad=1280:720:(ow-iw)/2:(oh-ih)/2" output.mp4
ffmpeg
-rtsp_transport tcp -i srcUrl
-fflags nobuffer
-max_delay 1
-threads 5
-profile:v high
-preset superfast
-tune zerolatency
-an -c:v h264
-crf 25
-s 1280*720
-f rtsp
-bf 0
-g 5
-rtsp_transport udp dstUrl
通过推流本地文件,模拟一个摄像头的RTSP视频流
// rtsp推流(文件推流)
ffmpeg -re -i camera_60s.mp4 -rtsp_transport tcp -c copy -f rtsp rtsp://127.0.0.1:554/live/camera
// rtsp推流(文件循环推流)
ffmpeg -re -stream_loop -1 -i camera_60s.mp4 -rtsp_transport tcp -c copy -f rtsp rtsp://127.0.0.1:554/live/camera
//查看本机USB摄像头列表
ffmpeg -list_devices true -f dshow -i dummy
//播放本机USB播放摄像头,FULL HD webcam 是通过查看列表的命令行获得的名称
ffplay -f dshow -i video="FULL HD webcam"
//本机USB摄像头+转码+推流到RTSP服务器(rtp over tcp),FULL HD webcam 是通过查看列表的命令行获得的名称
ffmpeg -f dshow -i video="FULL HD webcam" -fflags nobuffer -max_delay 1 -threads 5 -profile:v high -preset superfast -tune zerolatency -an -c:v h264 -crf 25 -s 1280*720 -f rtsp -bf 0 -g 5 -f rtsp rtsp://127.0.0.1/live/test