<分区>
所以我有带有这样代码的 VideoEncoder.h
/*
FFmpeg simple Encoder
*/
#ifndef __VIDEO_ENCODER_H__
#define __VIDEO_ENCODER_H__
#include "ffmpegInclude.h"
#include <Windows.h>
#include <string>
class VideoEncoder
{
private:
// output file name
std::string outputFilename;
// output format.
AVOutputFormat *pOutFormat;
// format context
AVFormatContext *pFormatContext;
// video stream context
AVStream * pVideoStream;
// audio streams context
AVStream * pAudioStream;
// convert context context
struct SwsContext *pImgConvertCtx;
// encode buffer and size
uint8_t * pVideoEncodeBuffer;
int nSizeVideoEncodeBuffer;
// audio buffer and size
uint8_t * pAudioEncodeBuffer;
int nSizeAudioEncodeBuffer;
// count of sample
int audioInputSampleSize;
// current picture
AVFrame *pCurrentPicture;
// audio buffer
char* audioBuffer;
int nAudioBufferSize;
int nAudioBufferSizeCurrent;
public:
VideoEncoder()
{
pOutFormat = NULL;
pFormatContext = NULL;
pVideoStream = NULL;
pImgConvertCtx = NULL;
pCurrentPicture = NULL;
pVideoEncodeBuffer = NULL;
nSizeVideoEncodeBuffer = 0;
pAudioEncodeBuffer = NULL;
nSizeAudioEncodeBuffer = 0;
nAudioBufferSize = 1024 * 1024 * 4;
audioBuffer = new char[nAudioBufferSize];
nAudioBufferSizeCurrent = 0;
}
virtual ~VideoEncoder()
{
Finish();
}
// init output file
bool InitFile(std::string& inputFile, std::string& container);
// Add video and audio data
bool AddFrame(AVFrame* frame, const char* soundBuffer, int soundBufferSize);
// end of output
bool Finish();
private:
// Add video stream
AVStream *AddVideoStream(AVFormatContext *pContext, CodecID codec_id);
// Open Video Stream
bool OpenVideo(AVFormatContext *oc, AVStream *pStream);
// Allocate memory
AVFrame * CreateFFmpegPicture(int pix_fmt, int nWidth, int nHeight);
// Close video stream
void CloseVideo(AVFormatContext *pContext, AVStream *pStream);
// Add audio stream
AVStream * AddAudioStream(AVFormatContext *pContext, CodecID codec_id);
// Open audio stream
bool OpenAudio(AVFormatContext *pContext, AVStream *pStream);
// close audio stream
void CloseAudio(AVFormatContext *pContext, AVStream *pStream);
// Add video frame
bool AddVideoFrame(AVFrame * frame, AVCodecContext *pVideoCodec);
// Add audio samples
bool AddAudioSample(AVFormatContext *pFormatContext,
AVStream *pStream, const char* soundBuffer, int soundBufferSize);
// Free resourses.
void Free();
bool NeedConvert();
};
#endif // __VIDEO_ENCODER_H__
所以我在这里看到了 InitFile、AddFrame 和 Finish。
在 VideoEncoder.cpp 中我看到了这个
#include <stdio.h>
#include <stdlib.h>
#include "ffmpegInclude.h"
#include <math.h>
#include "VideoEncoder.h"
#include "Settings.h"
#define MAX_AUDIO_PACKET_SIZE (128 * 1024)
bool VideoEncoder::InitFile(std::string& inputFile, std::string& container)
{
bool res = false;
const char * filename = inputFile.c_str();
outputFilename = inputFile;
// Initialize libavcodec
av_register_all();
if (container == std::string("auto"))
{
// Create format
pOutFormat = guess_format(NULL, filename, NULL);
}
else
{
// use contanier
pOutFormat = guess_format(container.c_str(), NULL, NULL);
}
if (pOutFormat)
{
// allocate context
pFormatContext = avformat_alloc_context();
if (pFormatContext)
{
pFormatContext->oformat = pOutFormat;
memcpy(pFormatContext->filename, filename, min(strlen(filename),
sizeof(pFormatContext->filename)));
// Add video and audio stream
pVideoStream = AddVideoStream(pFormatContext, pOutFormat->video_codec);
pAudioStream = AddAudioStream(pFormatContext, pOutFormat->audio_codec);
// Set the output parameters (must be done even if no
// parameters).
if (av_set_parameters(pFormatContext, NULL) >=0)
{
dump_format(pFormatContext, 0, filename, 1);
// Open Video and Audio stream
res = false;
if (pVideoStream)
{
res = OpenVideo(pFormatContext, pVideoStream);
}
res = OpenAudio(pFormatContext, pAudioStream);
if (res && !(pOutFormat->flags & AVFMT_NOFILE))
{
if (url_fopen(&pFormatContext->pb, filename, URL_WRONLY) < 0)
{
res = false;
printf("Cannot open file\n");
}
}
if (res)
{
av_write_header(pFormatContext);
res = true;
}
}
}
}
if (!res)
{
Free();
printf("Cannot init file\n");
}
return res;
}
bool VideoEncoder::AddFrame(AVFrame* frame, const char* soundBuffer, int soundBufferSize)
{
bool res = true;
int nOutputSize = 0;
AVCodecContext *pVideoCodec = NULL;
if (pVideoStream && frame && frame->data[0])
{
pVideoCodec = pVideoStream->codec;
if (NeedConvert())
{
// RGB to YUV420P.
if (!pImgConvertCtx)
{
pImgConvertCtx = sws_getContext(pVideoCodec->width, pVideoCodec->height,
PIX_FMT_RGB24,
pVideoCodec->width, pVideoCodec->height,
pVideoCodec->pix_fmt,
SWS_BICUBLIN, NULL, NULL, NULL);
}
}
// Allocate picture.
pCurrentPicture = CreateFFmpegPicture(pVideoCodec->pix_fmt, pVideoCodec->width,
pVideoCodec->height);
if (NeedConvert() && pImgConvertCtx)
{
// Convert RGB to YUV.
sws_scale(pImgConvertCtx, frame->data, frame->linesize,
0, pVideoCodec->height, pCurrentPicture->data, pCurrentPicture->linesize);
}
res = AddVideoFrame(pCurrentPicture, pVideoStream->codec);
// Free temp frame
av_free(pCurrentPicture->data[0]);
av_free(pCurrentPicture);
pCurrentPicture = NULL;
}
// Add sound
if (soundBuffer && soundBufferSize > 0)
{
res = AddAudioSample(pFormatContext, pAudioStream, soundBuffer, soundBufferSize);
}
return res;
}
bool VideoEncoder::Finish()
{
bool res = true;
if (pFormatContext)
{
av_write_trailer(pFormatContext);
Free();
}
if (audioBuffer)
{
delete[] audioBuffer;
audioBuffer = NULL;
}
return res;
}
void VideoEncoder::Free()
{
bool res = true;
if (pFormatContext)
{
// close video stream
if (pVideoStream)
{
CloseVideo(pFormatContext, pVideoStream);
}
// close audio stream.
if (pAudioStream)
{
CloseAudio(pFormatContext, pAudioStream);
}
// Free the streams.
for(size_t i = 0; i < pFormatContext->nb_streams; i++)
{
av_freep(&pFormatContext->streams[i]->codec);
av_freep(&pFormatContext->streams[i]);
}
if (!(pFormatContext->flags & AVFMT_NOFILE) && pFormatContext->pb)
{
url_fclose(pFormatContext->pb);
}
// Free the stream.
av_free(pFormatContext);
pFormatContext = NULL;
}
}
AVFrame * VideoEncoder::CreateFFmpegPicture(int pix_fmt, int nWidth, int nHeight)
{
AVFrame *picture = NULL;
uint8_t *picture_buf = NULL;
int size;
picture = avcodec_alloc_frame();
if ( !picture)
{
printf("Cannot create frame\n");
return NULL;
}
size = avpicture_get_size(pix_fmt, nWidth, nHeight);
picture_buf = (uint8_t *) av_malloc(size);
if (!picture_buf)
{
av_free(picture);
printf("Cannot allocate buffer\n");
return NULL;
}
avpicture_fill((AVPicture *)picture, picture_buf,
pix_fmt, nWidth, nHeight);
return picture;
}
bool VideoEncoder::OpenVideo(AVFormatContext *oc, AVStream *pStream)
{
AVCodec *pCodec;
AVCodecContext *pContext;
pContext = pStream->codec;
// Find the video encoder.
pCodec = avcodec_find_encoder(pContext->codec_id);
if (!pCodec)
{
printf("Cannot found video codec\n");
return false;
}
// Open the codec.
if (avcodec_open(pContext, pCodec) < 0)
{
printf("Cannot open video codec\n");
return false;
}
pVideoEncodeBuffer = NULL;
if (!(pFormatContext->oformat->flags & AVFMT_RAWPICTURE))
{
/* allocate output buffer */
nSizeVideoEncodeBuffer = 10000000;
pVideoEncodeBuffer = (uint8_t *)av_malloc(nSizeVideoEncodeBuffer);
}
return true;
}
void VideoEncoder::CloseVideo(AVFormatContext *pContext, AVStream *pStream)
{
avcodec_close(pStream->codec);
if (pCurrentPicture)
{
if (pCurrentPicture->data)
{
av_free(pCurrentPicture->data[0]);
pCurrentPicture->data[0] = NULL;
}
av_free(pCurrentPicture);
pCurrentPicture = NULL;
}
if (pVideoEncodeBuffer)
{
av_free(pVideoEncodeBuffer);
pVideoEncodeBuffer = NULL;
}
nSizeVideoEncodeBuffer = 0;
}
bool VideoEncoder::NeedConvert()
{
bool res = false;
if (pVideoStream && pVideoStream->codec)
{
res = (pVideoStream->codec->pix_fmt != PIX_FMT_RGB24);
}
return res;
}
AVStream *VideoEncoder::AddVideoStream(AVFormatContext *pContext, CodecID codec_id)
{
AVCodecContext *pCodecCxt = NULL;
AVStream *st = NULL;
st = av_new_stream(pContext, 0);
if (!st)
{
printf("Cannot add new vidoe stream\n");
return NULL;
}
pCodecCxt = st->codec;
pCodecCxt->codec_id = (CodecID)codec_id;
pCodecCxt->codec_type = CODEC_TYPE_VIDEO;
pCodecCxt->frame_number = 0;
// Put sample parameters.
pCodecCxt->bit_rate = 2000000;
// Resolution must be a multiple of two.
pCodecCxt->width = W_VIDEO;
pCodecCxt->height = H_VIDEO;
/* time base: this is the fundamental unit of time (in seconds) in terms
of which frame timestamps are represented. for fixed-fps content,
timebase should be 1/framerate and timestamp increments should be
identically 1. */
pCodecCxt->time_base.den = 25;
pCodecCxt->time_base.num = 1;
pCodecCxt->gop_size = 12; // emit one intra frame every twelve frames at most
pCodecCxt->pix_fmt = PIX_FMT_YUV420P;
if (pCodecCxt->codec_id == CODEC_ID_MPEG2VIDEO)
{
// Just for testing, we also add B frames
pCodecCxt->max_b_frames = 2;
}
if (pCodecCxt->codec_id == CODEC_ID_MPEG1VIDEO)
{
/* Needed to avoid using macroblocks in which some coeffs overflow.
This does not happen with normal video, it just happens here as
the motion of the chroma plane does not match the luma plane. */
pCodecCxt->mb_decision = 2;
}
// Some formats want stream headers to be separate.
if(pContext->oformat->flags & AVFMT_GLOBALHEADER)
{
pCodecCxt->flags |= CODEC_FLAG_GLOBAL_HEADER;
}
return st;
}
AVStream * VideoEncoder::AddAudioStream(AVFormatContext *pContext, CodecID codec_id)
{
AVCodecContext *pCodecCxt = NULL;
AVStream *pStream = NULL;
// Try create stream.
pStream = av_new_stream(pContext, 1);
if (!pStream)
{
printf("Cannot add new audio stream\n");
return NULL;
}
// Codec.
pCodecCxt = pStream->codec;
pCodecCxt->codec_id = codec_id;
pCodecCxt->codec_type = CODEC_TYPE_AUDIO;
// Set format
pCodecCxt->bit_rate = 128000;
pCodecCxt->sample_rate = 44100;
pCodecCxt->channels = 1;
pCodecCxt->sample_fmt = SAMPLE_FMT_S16;
nSizeAudioEncodeBuffer = 4 * MAX_AUDIO_PACKET_SIZE;
if (pAudioEncodeBuffer == NULL)
{
pAudioEncodeBuffer = (uint8_t * )av_malloc(nSizeAudioEncodeBuffer);
}
// Some formats want stream headers to be separate.
if(pContext->oformat->flags & AVFMT_GLOBALHEADER)
{
pCodecCxt->flags |= CODEC_FLAG_GLOBAL_HEADER;
}
return pStream;
}
bool VideoEncoder::OpenAudio(AVFormatContext *pContext, AVStream *pStream)
{
AVCodecContext *pCodecCxt = NULL;
AVCodec *pCodec = NULL;
pCodecCxt = pStream->codec;
// Find the audio encoder.
pCodec = avcodec_find_encoder(pCodecCxt->codec_id);
if (!pCodec)
{
printf("Cannot open audio codec\n");
return false;
}
// Open it.
if (avcodec_open(pCodecCxt, pCodec) < 0)
{
printf("Cannot open audio codec\n");
return false;
}
if (pCodecCxt->frame_size <= 1)
{
// Ugly hack for PCM codecs (will be removed ASAP with new PCM
// support to compute the input frame size in samples.
audioInputSampleSize = nSizeAudioEncodeBuffer / pCodecCxt->channels;
switch (pStream->codec->codec_id)
{
case CODEC_ID_PCM_S16LE:
case CODEC_ID_PCM_S16BE:
case CODEC_ID_PCM_U16LE:
case CODEC_ID_PCM_U16BE:
audioInputSampleSize >>= 1;
break;
default:
break;
}
pCodecCxt->frame_size = audioInputSampleSize;
}
else
{
audioInputSampleSize = pCodecCxt->frame_size;
}
return true;
}
void VideoEncoder::CloseAudio(AVFormatContext *pContext, AVStream *pStream)
{
avcodec_close(pStream->codec);
if (pAudioEncodeBuffer)
{
av_free(pAudioEncodeBuffer);
pAudioEncodeBuffer = NULL;
}
nSizeAudioEncodeBuffer = 0;
}
bool VideoEncoder::AddVideoFrame(AVFrame * pOutputFrame, AVCodecContext *pVideoCodec)
{
bool res = false;
if (pFormatContext->oformat->flags & AVFMT_RAWPICTURE)
{
// Raw video case. The API will change slightly in the near
// futur for that.
AVPacket pkt;
av_init_packet(&pkt);
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index = pVideoStream->index;
pkt.data= (uint8_t *) pOutputFrame;
pkt.size= sizeof(AVPicture);
res = av_interleaved_write_frame(pFormatContext, &pkt);
res = true;
}
else
{
// Encode
int nOutputSize = avcodec_encode_video(pVideoCodec, pVideoEncodeBuffer,
nSizeVideoEncodeBuffer, pOutputFrame);
if (nOutputSize > 0)
{
AVPacket pkt;
av_init_packet(&pkt);
if (pVideoCodec->coded_frame->pts != AV_NOPTS_VALUE)
{
pkt.pts = av_rescale_q(pVideoCodec->coded_frame->pts,
pVideoCodec->time_base, pVideoStream->time_base);
}
if(pVideoCodec->coded_frame->key_frame)
{
pkt.flags |= PKT_FLAG_KEY;
}
pkt.stream_index = pVideoStream->index;
pkt.data = pVideoEncodeBuffer;
pkt.size = nOutputSize;
// Write frame
res = (av_interleaved_write_frame(pFormatContext, &pkt) == 0);
}
else
{
res = false;
}
}
return res;
}
bool VideoEncoder::AddAudioSample(AVFormatContext *pFormatContext, AVStream *pStream,
const char* soundBuffer, int soundBufferSize)
{
AVCodecContext *pCodecCxt;
bool res = true;
pCodecCxt = pStream->codec;
memcpy(audioBuffer + nAudioBufferSizeCurrent, soundBuffer, soundBufferSize);
nAudioBufferSizeCurrent += soundBufferSize;
BYTE * pSoundBuffer = (BYTE *)audioBuffer;
int nCurrentSize = nAudioBufferSizeCurrent;
// Size of packet on bytes.
// FORMAT s16
DWORD packSizeInSize = 2 * audioInputSampleSize;
while(nCurrentSize >= packSizeInSize)
{
AVPacket pkt;
av_init_packet(&pkt);
pkt.size = avcodec_encode_audio(pCodecCxt, pAudioEncodeBuffer,
nSizeAudioEncodeBuffer, (const short *)pSoundBuffer);
if (pCodecCxt->coded_frame && pCodecCxt->coded_frame->pts != AV_NOPTS_VALUE)
{
pkt.pts = av_rescale_q(pCodecCxt->coded_frame->pts, pCodecCxt->time_base, pStream->time_base);
}
pkt.flags |= PKT_FLAG_KEY;
pkt.stream_index = pStream->index;
pkt.data = pAudioEncodeBuffer;
// Write the compressed frame in the media file.
if (av_interleaved_write_frame(pFormatContext, &pkt) != 0)
{
res = false;
break;
}
nCurrentSize -= packSizeInSize;
pSoundBuffer += packSizeInSize;
}
// save excess
memcpy(audioBuffer, audioBuffer + nAudioBufferSizeCurrent - nCurrentSize, nCurrentSize);
nAudioBufferSizeCurrent = nCurrentSize;
return res;
}
(示例中的代码示例 in this article in russian)
这一切 VideoEncoder::
是什么意思?为什么要在 H 文件中声明而不是在 cpp 文件中创建所有类?
为什么进一步在代码中仅使用#include "VideoEncoder.h"来声明 VE 类,以及为什么在 VE.h 中声明的函数即使没有引用 VE.cpp 文件也能完美运行?