macOS使用ffmpeg与QT进行音视频推拉流

1.先启动流服务器 ./mediamtx

2.开始推流: ffmpeg -re -stream_loop -1 -i /Users/hackerx/Desktop/test.mp4 -c copy -rtsp_transport tcp -f rtsp rtsp://127.0.0.1:8554/stream

  1. 安装ffmpeg 4.4

brew install ffmpeg@4

4.添加ffmpeg头文件目录与库目录

5.链接ffmpeg相关库

6.设计界面

7.拉流

ffmpegmanager.cpp:

cpp 复制代码
#include "ffmpegmananger.h"
#include <QThread>
//构造
ffmpegMananger::ffmpegMananger(QObject *parent) : QObject{parent}
{
    m_pInFmtCtx = nullptr;//输入流格式上下文
    m_pTsFmtCtx  = nullptr;//输出流格式上下文
    m_strInputStreamUrl = "";//输入流地址
    m_strOutputStreamPath = "";//输出流地址

}

//拆构
ffmpegMananger::~ffmpegMananger()
{
    avformat_free_context(m_pInFmtCtx);//释放输入流格式上下文
    avformat_free_context(m_pTsFmtCtx);//释放输出流格式上下文
}

//取输入流地址
void ffmpegMananger::getRtspAddress(QString url)
{
    this->m_strInputStreamUrl = url;
}

//取输出流地址
void ffmpegMananger::getOutputAddress(QString path)
{
    this->m_strOutputStreamPath = path;
    printf("输出流地址: %s\n",m_strOutputStreamPath.toStdString().c_str());
}

void ffmpegMananger::setOutputCtx(AVCodecContext *encCtx, AVFormatContext **pTsFmtCtx,int &nVideoIdx_out)
{
    avformat_alloc_output_context2(pTsFmtCtx , nullptr, nullptr, m_strOutputStreamPath.toStdString().c_str());
    if (!pTsFmtCtx ) {
        printf("创建输出上下文失败: avformat_alloc_output_context2\n");
        return;
    }
    if (avio_open(&((*pTsFmtCtx)->pb), m_strOutputStreamPath.toStdString().c_str(), AVIO_FLAG_READ_WRITE) < 0)
    {
        avformat_free_context(*pTsFmtCtx);
        printf("打开输出流失败: avio_open\n");
        return;
    }
    AVStream *out_stream = avformat_new_stream(*pTsFmtCtx, encCtx->codec);
    nVideoIdx_out = out_stream->index;
    avcodec_parameters_from_context(out_stream->codecpar, encCtx);
    printf("输出流信息:\n");
    av_dump_format(*pTsFmtCtx, 0, m_strOutputStreamPath.toStdString().c_str(), 1);
    printf("----------------------------\n");
}

//拉流并播放
int ffmpegMananger::ffmepgInput()
{
    int nRet = 0;
    AVCodecContext *encCtx = nullptr;
    std::string temp = m_strInputStreamUrl.toStdString();
    const char *pUrl = temp.c_str();
    printf("输入流地址: %s\n",pUrl);

    //设置选项
    AVDictionary *dict = nullptr;
    av_dict_set(&dict,"rtsp_transport", "tcp", 0);
    av_dict_set(&dict,"stimeout","10000000",0);
    av_dict_set(&dict, "buffer_size", "1024000", 0);

    //打开输入流
    nRet = avformat_open_input(&m_pInFmtCtx,pUrl,nullptr,&dict);
    if( nRet < 0)
    {
        printf("打开输入流失败\n");
        return nRet;
    }

    avformat_find_stream_info(m_pInFmtCtx, nullptr);
    printf("输入流信息:\n");
    av_dump_format(m_pInFmtCtx, 0, pUrl, 0);
    printf("---------------------------\n");

    //视频流索引
    int nVideo_indx = av_find_best_stream(m_pInFmtCtx,AVMEDIA_TYPE_VIDEO,-1,-1,nullptr,0);
    if(nVideo_indx < 0)
    {
        avformat_free_context(m_pInFmtCtx);
        printf("查找视频流索引失败: av_find_best_stream\n");
        return -1;
    }

    //查找解码器
    auto pInCodec = avcodec_find_decoder(m_pInFmtCtx->streams[nVideo_indx]->codecpar->codec_id);
    if(nullptr == pInCodec)
    {
        printf("查找解码器失败: avcodec_find_decoder fail.");
        return -1;
    }

    //解码器上下文
    AVCodecContext* pInCodecCtx = avcodec_alloc_context3(pInCodec);

    //设置解码器参数
    nRet = avcodec_parameters_to_context(pInCodecCtx, m_pInFmtCtx->streams[nVideo_indx]->codecpar);
    if(nRet < 0)
    {

        avcodec_free_context(&pInCodecCtx);
        printf("设置解码器参数失败: avcodec_parameters_to_context");
        return -1;
    }

    //打开解码器
    if(avcodec_open2(pInCodecCtx, pInCodec, nullptr) < 0)
    {
        avcodec_free_context(&pInCodecCtx);
        printf("打开解码器失败: avcodec_open2\n");
        return -1;
    }

    //输出视频分辨率
    printf("视频宽:%d\n", pInCodecCtx->width);
    printf("视频高:%d\n", pInCodecCtx->height);

    int frame_index = 0;//帧索引
    int got_picture = 0;//帧解码结果

    //输出输入流
    AVStream *in_stream =nullptr;
    AVStream *out_stream =nullptr;

    //分配内存
    AVFrame *pFrame= av_frame_alloc();
    AVFrame *pFrameRGB = av_frame_alloc();
    AVPacket *newpkt = av_packet_alloc();
    AVPacket *packet = av_packet_alloc();

    //初始化视频包
    av_init_packet(newpkt);
    av_init_packet(packet);


    //图像色彩空间转换、分辨率缩放、前后图像滤波处理
    SwsContext *m_SwsContext = sws_getContext(
                                    pInCodecCtx->width,
                                    pInCodecCtx->height,
                                    pInCodecCtx->pix_fmt,
                                    pInCodecCtx->width,
                                    pInCodecCtx->height,
                                    AV_PIX_FMT_RGB32,
                                    SWS_BICUBIC,
                                    nullptr, nullptr, nullptr);

    int bytes = av_image_get_buffer_size(
                                    AV_PIX_FMT_RGB32,
                                    pInCodecCtx->width,
                                    pInCodecCtx->height,
                                    4);
    uint8_t *m_OutBuffer = (uint8_t *)av_malloc(bytes * sizeof(uint8_t));

    //将分配的内存空间给pFrameRGB使用
    avpicture_fill((AVPicture *)pFrameRGB,
                   m_OutBuffer,
                   AV_PIX_FMT_RGB32,
                   pInCodecCtx->width,
                   pInCodecCtx->height);

    if(encCtx == nullptr)
    {
        //打开编码器
        openEncoder(pInCodecCtx->width, pInCodecCtx->height,&encCtx);
    }
    //视频索引
    int videoindex_out = 0;
    //设置输出文件上下文
    setOutputCtx(encCtx,&m_pTsFmtCtx,videoindex_out);
    //写文件头
    if (avformat_write_header(m_pTsFmtCtx, nullptr) < 0)
    {
        avformat_free_context(m_pTsFmtCtx);
        printf("写文件头失败\n");
        return -1;
    }
    printf("写文件头成功.\n");

    int count = 0;//已解码帧数量
    nRet = 0;//读取帧结果
    //从pInFmtCtx读H264数据到packet;
    while(av_read_frame(m_pInFmtCtx, packet) >= 0)
    {
        if(packet->stream_index != nVideo_indx)//仅保留图像
        {
            continue;
        }

        //送packet中H264数据给解码器码器进行解码,解码好的YUV数据放在pInCodecCtx,
        if(avcodec_send_packet(pInCodecCtx, packet)<0)
        {
            break;
        }
        //释放已解码帧引用
        av_packet_unref(packet);
        //把解码好的YUV数据放到pFrame中
        got_picture = avcodec_receive_frame(pInCodecCtx, pFrame);

        //解码好一帧数据
        if(0 == got_picture)
        {
            //发送显示图像的信号
            // 对解码视频帧进行缩放、格式转换等操作
            sws_scale(m_SwsContext,
                      (uint8_t const * const *)pFrame->data,
                      pFrame->linesize,
                      0,
                      pInCodecCtx->height,
                      pFrameRGB->data,
                      pFrameRGB->linesize);

            // 转换到QImage
            QImage tmmImage((uchar *)m_OutBuffer, pInCodecCtx->width, pInCodecCtx->height, QImage::Format_RGB32);
            //复制图像
            QImage image = tmmImage.copy();
            //发送图像帧解码完成信息
            emit Sig_GetOneFrame(image);
            //设置解码器PTS
            setDecoderPts(newpkt->stream_index,count, pFrame);
            count++;//已解码计数
            //送原始数据给编码器进行编码
            nRet = avcodec_send_frame(encCtx,pFrame);
            if(nRet < 0)
            {
                continue;
            }
            //从编码器获取编号的数据
            while(nRet >= 0)
            {
                //接收已编码包
                nRet = avcodec_receive_packet(encCtx,newpkt);
                if(nRet < 0)
                {
                    break;
                }
                //设置编码包PTS
                setEncoderPts(nVideo_indx,frame_index,videoindex_out,newpkt);
                int _count = 1;
                printf("写%d包,大小:%5d,PTS:%lld\n", _count,newpkt->size, newpkt->pts);

                if (av_interleaved_write_frame(m_pTsFmtCtx, newpkt) < 0)
                {
                    printf("写帧失败: av_interleaved_write_frame\n");
                    goto end;
                }
                _count++;
                av_packet_unref(newpkt);//释放已编码包
            }
        }
    }
    while(1)//从pInFmtCtx读H264数据到packet;
    {
        if(packet->stream_index != nVideo_indx)//仅保留图像
        {
            continue;
        }

        //送packet中H264数据给解码器码器进行解码,解码好的YUV数据放在pInCodecCtx,
        if(avcodec_send_packet(pInCodecCtx, packet)<0)
        {
            continue;
        }

        //释放已解码包
        av_packet_unref(packet);
        //把解码好的YUV数据放到pFrame中
        got_picture = avcodec_receive_frame(pInCodecCtx, pFrame);

        //解码好一帧数据
        if(!got_picture)
        {
            AVRational in_time_base1 = in_stream->time_base;
            in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
            //PTS
            int64_t in_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
            pFrame->pts = (double)(count*in_duration) / (double)(av_q2d(in_time_base1)*AV_TIME_BASE);
            count++;
            //送原始数据给编码器进行编码
            nRet = avcodec_send_frame(encCtx,pFrame);
            if(nRet < 0)
            {
                break;
            }
            //从编码器获取编号的数据
            while(nRet >= 0)
            {
                nRet = avcodec_receive_packet(encCtx,newpkt);
                if(nRet < 0)
                {
                    continue;
                }
                in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
                out_stream = m_pTsFmtCtx->streams[videoindex_out];
                if (newpkt->stream_index == nVideo_indx)
                {
                    if (newpkt->pts == AV_NOPTS_VALUE)
                    {
                        //写入PTS
                        AVRational time_base1 = in_stream->time_base;
                        int64_t calc_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
                        //设置包参数
                        newpkt->pts = (double)(frame_index*calc_duration) / (double)(av_q2d(time_base1)*AV_TIME_BASE);
                        newpkt->dts = newpkt->pts;
                        newpkt->duration = (double)calc_duration / (double)(av_q2d(time_base1)*AV_TIME_BASE);
                        frame_index++;
                    }
                }
                //转换PTS/DTS
                newpkt->pts = av_rescale_q_rnd(newpkt->pts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
                newpkt->dts = av_rescale_q_rnd(newpkt->dts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
                newpkt->duration = av_rescale_q(newpkt->duration, in_stream->time_base, out_stream->time_base);
                newpkt->pos = -1;
                newpkt->stream_index = videoindex_out;
                int count = 1;
                printf("写%d包,大小:%5dPTS:%lld\n", count,newpkt->size, newpkt->pts);

                if (av_interleaved_write_frame(m_pTsFmtCtx, newpkt) < 0)
                {
                    printf("写帧失败: av_interleaved_write_frame\n");
                    goto end;
                }
                count++;
                av_packet_unref(newpkt);
            }
        }
    }
    //Write file trailer
    av_write_trailer(m_pTsFmtCtx);
end:
    av_frame_free(&pFrame);
    av_frame_free(&pFrameRGB);
    av_packet_unref(newpkt);
    av_packet_unref(packet);
    std::cout<<"拉流完成";
    return  0;
}

void ffmpegMananger::setDecoderPts(int idx,int count,AVFrame *pFrame)
{
    AVStream* in_stream = m_pInFmtCtx->streams[idx];
    AVRational in_time_base1 = in_stream->time_base;
    //Duration between 2 frames (us)
    int64_t in_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
    pFrame->pts = (double)(count*in_duration) / (double)(av_q2d(in_time_base1)*AV_TIME_BASE);
}

void ffmpegMananger::setEncoderPts(int nVideo_indx,int frame_index,int videoindex_out,AVPacket *newpkt)
{
    AVStream*in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
    AVStream*out_stream = m_pTsFmtCtx->streams[videoindex_out];
    if (newpkt->stream_index == nVideo_indx)
    {
        //FIX:No PTS (Example: Raw H.264)
        //Simple Write PTS
        if (newpkt->pts == AV_NOPTS_VALUE)
        {
            //Write PTS
            AVRational time_base1 = in_stream->time_base;
            //Duration between 2 frames (us)
            int64_t calc_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
            //Parameters
            newpkt->pts = (double)(frame_index*calc_duration) / (double)(av_q2d(time_base1)*AV_TIME_BASE);
            newpkt->dts = newpkt->pts;
            newpkt->duration = (double)calc_duration / (double)(av_q2d(time_base1)*AV_TIME_BASE);
            frame_index++;
        }
    }
    //Convert PTS/DTS
    newpkt->pts = av_rescale_q_rnd(newpkt->pts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
    newpkt->dts = av_rescale_q_rnd(newpkt->dts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
    newpkt->duration = av_rescale_q(newpkt->duration, in_stream->time_base, out_stream->time_base);
    newpkt->pos = -1;
    newpkt->stream_index = videoindex_out;
}

void ffmpegMananger::writeTail()
{
    //Write file trailer
    av_write_trailer(m_pTsFmtCtx);
}

void ffmpegMananger::openEncoder(int width, int height, AVCodecContext** enc_ctx)
{
    //使用libx264编码器
    auto pCodec = avcodec_find_encoder_by_name("libx264");
    if(nullptr == pCodec)
    {
        printf("avcodec_find_encoder_by_name fail.\n");
        return;
    }
    //获取编码器上下文
    *enc_ctx = avcodec_alloc_context3(pCodec);
    if(nullptr == enc_ctx)
    {
        printf("avcodec_alloc_context3(pCodec) fail.\n");
        return;
    }
    //sps/pps
    (*enc_ctx)->profile = FF_PROFILE_H264_MAIN;
    (*enc_ctx)->level = 30;//表示level是5.0
    //分辨率
    (*enc_ctx)->width = width;
    (*enc_ctx)->height = height;
    //gop
    (*enc_ctx)->gop_size = 25;//i帧间隔
    (*enc_ctx)->keyint_min = 20;//设置最小自动插入i帧的间隔.OPTION
    //B帧
    (*enc_ctx)->max_b_frames = 0;//不要B帧
    (*enc_ctx)->has_b_frames = 0;//
    //参考帧
    (*enc_ctx)->refs = 3;//OPTION
    //设置输入的yuv格式
    (*enc_ctx)->pix_fmt = AV_PIX_FMT_YUV420P;
    //设置码率
    (*enc_ctx)->bit_rate = 3000000;
    //设置帧率
    (*enc_ctx)->time_base = (AVRational){1,25};//帧与帧之间的间隔
    (*enc_ctx)->framerate = (AVRational){25,1};//帧率 25帧每秒
    if(avcodec_open2((*enc_ctx),pCodec,nullptr) < 0)
    {
        printf("avcodec_open2 fail.\n");
    }
    return;
}

ffmpegmanager.h

cpp 复制代码
#ifndef FFMPEGMANANGER_H
#define FFMPEGMANANGER_H
#pragma execution_character_set("utf-8")
//QT头
#include <QObject>
#include <QTimer>
#include <QImage>
//C标准头
#include <stdio.h>
#include <iostream>
//FFmpeg头
extern "C"
{
    #include "libswscale/swscale.h"
    #include "libavdevice/avdevice.h"
    #include "libavcodec/avcodec.h"
    #include "libavcodec/bsf.h"
    #include "libavformat/avformat.h"
    #include "libavutil/avutil.h"
    #include "libavutil/imgutils.h"
    #include "libavutil/log.h"
    #include "libavutil/time.h"
    #include <libswresample/swresample.h>

}

class ffmpegMananger : public QObject
{
    Q_OBJECT
public:
    //构造
    explicit ffmpegMananger(QObject *parent = nullptr);
    //拆构
    ~ffmpegMananger();
    //取输入流地址
    void getRtspAddress(QString url);
    //取输出流地址
    void getOutputAddress(QString path);
    //ffmpeg拉流播放
    int ffmepgInput();
    //打开解码器
    void openEncoder(int width, int height, AVCodecContext** enc_ctx);
    //设置输出上下文
    void setOutputCtx(AVCodecContext *encCtx, AVFormatContext **pTsFmtCtx,int &nVideoIdx_out);
    //写文件尾
    void writeTail();
    //设置解码的pts
    void setDecoderPts(int idx,int count,AVFrame *pFrame);
    //设置编码的pts
    void setEncoderPts(int nVideo_indx,int frame_index,int videoindex_out,AVPacket *newpkt);
signals:
    //取一帧图像信号
    void Sig_GetOneFrame(QImage img);
private:
    //输入流地址
    QString m_strInputStreamUrl;
    //输出流地址
    QString m_strOutputStreamPath;
    //输入流动格式上下文
    AVFormatContext *m_pInFmtCtx;
    //输出流动格式上下文
    AVFormatContext *m_pTsFmtCtx;
    bool m_ifRec;

};

#endif // FFMPEGMANANGER_H
相关推荐
LNTON羚通5 分钟前
算法定制LiteAIServer摄像机实时接入分析平台玩手机打电话检测算法:智能监控的新篇章
安全·目标检测·音视频·监控·视频监控
姆路2 小时前
QT中使用图表之QChart绘制动态折线图
c++·qt
何曾参静谧2 小时前
「QT」文件类 之 QDir 目录类
开发语言·qt
棱角~~2 小时前
10款音频剪辑工具的个人实践体验感受!!
人工智能·经验分享·其他·音视频·学习方法
何曾参静谧2 小时前
「QT」文件类 之 QTemporaryDir 临时目录类
开发语言·qt
EasyCVR3 小时前
GA/T1400视图库平台EasyCVR视频融合平台HLS视频协议是什么?
服务器·网络·人工智能·音视频
老赵的博客4 小时前
QSS 设置bug
前端·bug·音视频
huangjiazhi_5 小时前
QTcpSocket 服务端和客户端
开发语言·qt
EasyNVR5 小时前
NVR录像机汇聚管理EasyNVR多品牌NVR管理工具视频汇聚技术在智慧安防监控中的应用与优势
安全·音视频·监控·视频监控
写bug的小屁孩6 小时前
websocket身份验证
开发语言·网络·c++·qt·websocket·网络协议·qt6.3