使用FFmpeg解码并用swscale将YUV转为RGB

#include <stdio.h>
#include <libavcodec/avcodec.h>
#include <libavformat/avformat.h>
#include <libswscale/swscale.h>
#include <libavutil/imgutils.h>

void SaveFrame(AVFrame *pFrame, int width, int height, int iFrame);

int main(int argc, char *argv[])
{
    // 获取AVFormatContext句柄
    AVFormatContext *pFormatCtx = avformat_alloc_context();

    // 打开一个流媒体文件 open video file
    if (avformat_open_input(&pFormatCtx, argv[1], NULL, NULL)) {
        fprintf(stderr, "open input failed\n");
        return -1;
    }

    // 获取多媒体信息,存至句柄 retrieve stream information
    if (avformat_find_stream_info(pFormatCtx, NULL) < 0) {
        fprintf(stderr, "find stream info failed\n");
        return -1;
    }

    // 打印多媒体信息 dump information about file onto standard error
    av_dump_format(pFormatCtx, 0, argv[1], 0);
    
    // 通过句柄,找出视频流 find the video stream
    int videoStream = -1;
    videoStream = av_find_best_stream(pFormatCtx, AVMEDIA_TYPE_VIDEO, -1, -1, NULL, 0);
    if (-1 == videoStream) {
        fprintf(stderr, "Can not find video stream!\n");
        return -1;
    }
    
    AVCodecParameters *pCodecPar = NULL;
    pCodecPar = pFormatCtx->streams[videoStream]->codecpar;
    
    // 搜索合适的视频解码器 find the decoder for the video stream
    AVCodec *pCodec = NULL;
    pCodec = avcodec_find_decoder(pCodecPar->codec_id);
    if (NULL == pCodec) {
        fprintf(stderr, "Unsupported codec!\n");
        return -1;
    }
    
    AVCodecContext *pCodecCtx = NULL;
    pCodecCtx = avcodec_alloc_context3(pCodec);
    // 因为 AVStream::codec 被弃用,AVCodecContext 需要通过 AVCodecParameters 转换得到
    avcodec_parameters_to_context(pCodecCtx, pCodecPar);
    
    // 打开视频解码器 open Codec
    if (avcodec_open2(pCodecCtx, pCodec, NULL)) {
        fprintf(stderr, "Codec open failed!\n");
        return -1;
    }
    
    AVFrame *pFrame = NULL;
    AVFrame *pFrameRGB = NULL;
    // 分配两个视频帧,pFrame保存原始帧,pFrameRGB存放转换后的RGB帧 Allocate video frame
    pFrame = av_frame_alloc();
    pFrameRGB = av_frame_alloc();
    if (NULL == pFrameRGB || NULL == pFrame) {
        fprintf(stderr, "Alloc frame failed!\n");
        return -1;
    }
    uint8_t *buffer = NULL;
    int numBytes = 0;
    // 计算解码后原始数据所需缓冲区大小,并分配内存空间 Determine required buffer size and allocate buffer
    numBytes = av_image_get_buffer_size(AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height, 1);
    buffer = (uint8_t *)av_malloc(numBytes * sizeof(uint8_t));
    
    av_image_fill_arrays(pFrameRGB->data, pFrameRGB->linesize, buffer, AV_PIX_FMT_RGB24, pCodecCtx->width, pCodecCtx->height, 1);
        
    int frameFinished = 0;
    AVPacket packet = {0};
    int i = 0;
    struct SwsContext *img_convert_ctx = NULL;
    // 获取swscale句柄
    img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt,  
        pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL); 
    while (av_read_frame(pFormatCtx, &packet) >= 0) {
        // Is this a packet from the video stream?
        if (packet.stream_index == videoStream) {
            // Decode video frame
            int ret = avcodec_send_packet(pCodecCtx, &packet);
            if (0 != ret)
                continue;
            while (avcodec_receive_frame(pCodecCtx, pFrame) == 0) {
                // Convert the image from its native format to RGB
                //img_convert((AVPicture *)pFrameRGB, AV_PIX_FMT_RGB24, (AVPicture *)pFrame, 
                //            pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height);
                sws_scale(img_convert_ctx, (const unsigned char* const*)pFrame->data, pFrame->linesize, 0, pCodecCtx->height, pFrameRGB->data, pFrameRGB->linesize);
                // Save the frame to disk
                if (++i <= 5)
                    SaveFrame(pFrameRGB, pCodecCtx->width, pCodecCtx->height, i);
                else
                    break;
            }
#if 0            
            avcodec_decode_video(pCodecCtx, pFrame, &frameFinished, packet.data, packet.size);
            // Did we get a video frame?
            if (frameFinished) {
                // Convert the image from its native format to RGB
                img_convert((AVPicture *)pFrameRGB, AV_PIX_FMT_RGB24, (AVPicture *)pFrame, 
                            pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height);
                
                // Save the frame to disk
                if (++i <= 5)
                    SaveFrame(pFrameRGB, pCodecCtx->width, pCodecCtx->height, i);
            }
#endif            
        }
    }
    // Free the packet that was allocate by av_read_frame
    av_packet_unref(&packet);
    
    // Free the RGB image
    av_free(buffer);
    av_free(pFrameRGB);
    
    // Free the YUV frame
    av_free(pFrame);
    
    // Close the codec
    avcodec_close(pCodecCtx);
    
    // Close the video file
    avformat_close_input(&pFormatCtx);
       
    return 0;
}

void SaveFrame(AVFrame *pFrame, int width, int height, int iFrame)
{
    FILE *pFile = NULL;
    char szFilename[32] = {0};
    int y = 0;
    
    // Open file
    sprintf(szFilename, "./raw/frame%d.ppm", iFrame);
    pFile = fopen(szFilename, "wb");
    if (NULL == pFile)
        return;
    
    // Write header
    fprintf(pFile, "P6\n%d %d\n255\n", width, height);
    
    // Write pixel data
    for (y = 0; y < height; y++)
        fwrite(pFrame->data[0] + y * pFrame->linesize[0], 1, width * 3, pFile);
    
    // Close file
    fclose(pFile);
    
}

Makefile:

all:yuv2rgb
CC=gcc
CLIBSFLAGS=-lavformat -lavcodec -lavutil -lswresample -lswscale -lz -lpthread -lm
FFMPEG=/usr/local
CFLAGS=-I$(FFMPEG)/include/
LDFLAGS = -L$(FFMPEG)/lib/
yuv2rgb:yuv2rgb.o
	$(CC) -o yuv2rgb yuv2rgb.o $(CLIBSFLAGS) $(CFLAGS) $(LDFLAGS)
yuv2rgb.o:yuv2rgb.c
	$(CC) -o yuv2rgb.o -c yuv2rgb.c  $(CLIBSFLAGS) $(CFLAGS) $(LDFLAGS)
clean:
	rm yuv2rgb yuv2rgb.o

posted on 2019-04-26 00:33  陈知许  阅读(1408)  评论(0编辑  收藏  举报

导航