#include "../plugin.h" #include "../engine.h" static plugfsfuncs_t *filefuncs; static plugaudiofuncs_t *audiofuncs; #include "libavcodec/avcodec.h" #include "libavformat/avformat.h" #include "libswscale/swscale.h" #include "libavutil/imgutils.h" #define TARGET_FFMPEG (LIBAVFORMAT_VERSION_MICRO >= 100) #define HAVE_DECOUPLED_API (LIBAVCODEC_VERSION_MAJOR>57 || (LIBAVCODEC_VERSION_MAJOR==57&&LIBAVCODEC_VERSION_MINOR>=36)) //between av 52.31 and 54.35, lots of constants etc got renamed to gain an extra AV_ prefix. /* #define AV_PIX_FMT_BGRA PIX_FMT_BGRA #define AVMEDIA_TYPE_AUDIO CODEC_TYPE_AUDIO #define AVMEDIA_TYPE_VIDEO CODEC_TYPE_VIDEO #define AV_PIX_FMT_BGRA PIX_FMT_BGRA #define AV_SAMPLE_FMT_U8 SAMPLE_FMT_U8 #define AV_SAMPLE_FMT_S16 SAMPLE_FMT_S16 #define AV_SAMPLE_FMT_FLT SAMPLE_FMT_FLT #define AVIOContext ByteIOContext #define avio_alloc_context av_alloc_put_byte */ #define DECODERNAME "ffmpeg" /*should probably try threading this, though I suppose it should be the engine doing that.*/ /*timing is based upon the start time. this means overflow issues with rtsp etc*/ struct decctx { unsigned int width, height; vfsfile_t *file; int64_t fileofs; int64_t filelen; AVFormatContext *pFormatCtx; int audioStream; AVCodecContext *pACodecCtx; AVFrame *pAFrame; int videoStream; AVCodecContext *pVCodecCtx; AVFrame *pVFrame; int64_t num, denum; int64_t lasttime; uint8_t *rgb_data; int rgb_linesize; struct SwsContext *pScaleCtx; }; static qboolean AVDec_SetSize (void *vctx, int width, int height) { struct decctx *ctx = (struct decctx*)vctx; uint8_t *rgb_data[4]; //av_image_alloc requires at least 4 entries for certain pix formats (libav (but not ffmpeg) zero-fills, so this is important). int rgb_linesize[4]; //colourspace conversions will be fastest if we // if (width > ctx->pCodecCtx->width) width = ctx->pVCodecCtx->width; // if (height > ctx->pCodecCtx->height) height = ctx->pVCodecCtx->height; //is this a no-op? if (width == ctx->width && height == ctx->height && ctx->pScaleCtx) return true; if (av_image_alloc(rgb_data, rgb_linesize, width, height, AV_PIX_FMT_BGRA, 1) >= 0) { //update the scale context as required //clear the old stuff out av_free(ctx->rgb_data); ctx->width = width; ctx->height = height; ctx->rgb_data = rgb_data[0]; ctx->rgb_linesize = rgb_linesize[0]; return qtrue; } return qfalse; //unsupported } static int AVIO_Read(void *opaque, uint8_t *buf, int buf_size) { struct decctx *ctx = opaque; int ammount; ammount = VFS_READ(ctx->file, buf, buf_size); if (ammount > 0) ctx->fileofs += ammount; return ammount; } static int64_t AVIO_Seek(void *opaque, int64_t offset, int whence) { struct decctx *ctx = opaque; whence &= ~AVSEEK_FORCE; switch(whence) { default: return -1; case SEEK_SET: ctx->fileofs = offset; break; case SEEK_CUR: ctx->fileofs += offset; break; case SEEK_END: ctx->fileofs = ctx->filelen + offset; break; case AVSEEK_SIZE: return ctx->filelen; } VFS_SEEK(ctx->file, ctx->fileofs); return ctx->fileofs; } static void AVDec_Destroy(void *vctx) { struct decctx *ctx = (struct decctx*)vctx; // Free the video stuff av_free(ctx->rgb_data); if (ctx->pVCodecCtx) avcodec_close(ctx->pVCodecCtx); av_free(ctx->pVFrame); // Free the audio decoder if (ctx->pACodecCtx) avcodec_close(ctx->pACodecCtx); av_free(ctx->pAFrame); // Close the video file avformat_close_input(&ctx->pFormatCtx); if (ctx->file) VFS_CLOSE(ctx->file); free(ctx); } static void *AVDec_Create(const char *medianame) { struct decctx *ctx; unsigned int i; AVCodec *pCodec; qboolean useioctx = false; // const char *extension = strrchr(medianame, '.'); /*always respond to av: media prefixes*/ if (!strncmp(medianame, "av:", 3) || !strncmp(medianame, "ff:", 3)) { medianame = medianame + 3; useioctx = true; } else if (!strncmp(medianame, "avs:", 4) || !strncmp(medianame, "ffs:", 4)) { medianame = medianame + 4; //let avformat do its own avio context stuff } else if (strchr(medianame, ':')) //block other types of url/prefix. return NULL; // else if (!strcasecmp(extension, ".roq") || !strcasecmp(extension, ".roq") || !strcasecmp(extension, ".cin")) // return NULL; //roq+cin should be played back via the engine instead... else useioctx = true; ctx = malloc(sizeof(*ctx)); memset(ctx, 0, sizeof(*ctx)); ctx->lasttime = -1; ctx->file = NULL; if (useioctx) { // Create internal Buffer for FFmpeg: const int iBufSize = 32 * 1024; char *pBuffer = av_malloc(iBufSize); AVIOContext *ioctx; ctx->file = filefuncs->OpenVFS(medianame, "rb", FS_GAME); if (!ctx->file) ctx->file = filefuncs->OpenVFS(va("video/%s", medianame), "rb", FS_GAME); if (!ctx->file) { Con_Printf("Unable to open %s\n", medianame); free(ctx); av_free(pBuffer); return NULL; } ctx->filelen = VFS_GETLEN(ctx->file); ioctx = avio_alloc_context(pBuffer, iBufSize, 0, ctx, AVIO_Read, 0, AVIO_Seek); ctx->pFormatCtx = avformat_alloc_context(); ctx->pFormatCtx->pb = ioctx; } /* small how-to note for if I ever try to add support for voice-and-video rtp decoding. this stuff is presumably needed to handle ICE+stun+ports etc. I prolly need to hack around with adding rtcp too. :s rtsp: Add support for depacketizing RTP data via custom IO To use this, set sdpflags=custom_io to the sdp demuxer. During the avformat_open_input call, the SDP is read from the AVFormatContext AVIOContext (ctx->pb) - after the avformat_open_input call, during the av_read_frame() calls, the same ctx->pb is used for reading packets (and sending back RTCP RR packets). Normally, one would use this with a read-only AVIOContext for the SDP during the avformat_open_input call, then close that one and replace it with a read-write one for the packets after the avformat_open_input call has returned. This allows using the RTP depacketizers as "pure" demuxers, without having them tied to the libavformat network IO. */ // Open video file if(avformat_open_input(&ctx->pFormatCtx, medianame, NULL, NULL)==0) { // Retrieve stream information if(avformat_find_stream_info(ctx->pFormatCtx, NULL)>=0) { ctx->audioStream=-1; for(i=0; ipFormatCtx->nb_streams && ctx->audioStream==-1; i++) { #if LIBAVFORMAT_VERSION_MAJOR >= 57 if(ctx->pFormatCtx->streams[i]->codecpar->codec_type==AVMEDIA_TYPE_AUDIO) #else if(ctx->pFormatCtx->streams[i]->codec->codec_type==AVMEDIA_TYPE_AUDIO) #endif ctx->audioStream=i; } if(ctx->audioStream!=-1) { #if LIBAVFORMAT_VERSION_MAJOR >= 57 pCodec=avcodec_find_decoder(ctx->pFormatCtx->streams[ctx->audioStream]->codecpar->codec_id); ctx->pACodecCtx = avcodec_alloc_context3(pCodec); if (avcodec_parameters_to_context(ctx->pACodecCtx, ctx->pFormatCtx->streams[ctx->audioStream]->codecpar) < 0) { avcodec_free_context(&ctx->pACodecCtx); pCodec = NULL; } #else ctx->pACodecCtx=ctx->pFormatCtx->streams[ctx->audioStream]->codec; pCodec=avcodec_find_decoder(ctx->pACodecCtx->codec_id); #endif ctx->pAFrame=av_frame_alloc(); if(pCodec!=NULL && ctx->pAFrame && avcodec_open2(ctx->pACodecCtx, pCodec, NULL) >= 0) { } else ctx->audioStream = -1; } ctx->videoStream=-1; for(i=0; ipFormatCtx->nb_streams && ctx->videoStream==-1; i++) { #if LIBAVFORMAT_VERSION_MAJOR >= 57 if(ctx->pFormatCtx->streams[i]->codecpar->codec_type==AVMEDIA_TYPE_VIDEO) #else if(ctx->pFormatCtx->streams[i]->codec->codec_type==AVMEDIA_TYPE_VIDEO) #endif ctx->videoStream=i; } if(ctx->videoStream!=-1) { #if LIBAVFORMAT_VERSION_MAJOR >= 57 pCodec=avcodec_find_decoder(ctx->pFormatCtx->streams[ctx->videoStream]->codecpar->codec_id); ctx->pVCodecCtx = avcodec_alloc_context3(pCodec); if (avcodec_parameters_to_context(ctx->pVCodecCtx, ctx->pFormatCtx->streams[ctx->videoStream]->codecpar) < 0) { avcodec_free_context(&ctx->pVCodecCtx); pCodec = NULL; } #else ctx->pVCodecCtx=ctx->pFormatCtx->streams[ctx->videoStream]->codec; pCodec=avcodec_find_decoder(ctx->pVCodecCtx->codec_id); #endif ctx->num = ctx->pFormatCtx->streams[ctx->videoStream]->time_base.num; ctx->denum = ctx->pFormatCtx->streams[ctx->videoStream]->time_base.den; // Open codec if(pCodec!=NULL && avcodec_open2(ctx->pVCodecCtx, pCodec, NULL) >= 0) { // Allocate video frame ctx->pVFrame=av_frame_alloc(); if(ctx->pVFrame!=NULL) { if (AVDec_SetSize(ctx, ctx->pVCodecCtx->width, ctx->pVCodecCtx->height)) { return ctx; } } } } } } AVDec_Destroy(ctx); return NULL; } static qboolean VARGS AVDec_DisplayFrame(void *vctx, qboolean nosound, qboolean forcevideo, double mediatime, void (QDECL *uploadtexture)(void *ectx, uploadfmt_t fmt, int width, int height, void *data, void *palette), void *ectx) { struct decctx *ctx = (struct decctx*)vctx; AVPacket packet; #if HAVE_DECOUPLED_API #else int frameFinished; #endif qboolean repainted = false; int64_t curtime; curtime = (mediatime * ctx->denum) / ctx->num; nosound |= !audiofuncs; while (1) { if (ctx->lasttime > curtime) break; // We're ahead of the previous frame. try and read the next. if (av_read_frame(ctx->pFormatCtx, &packet) < 0) { if (repainted) break; return false; } #if HAVE_DECOUPLED_API if(packet.stream_index==ctx->videoStream) { avcodec_send_packet(ctx->pVCodecCtx, &packet); while(0==avcodec_receive_frame(ctx->pVCodecCtx, ctx->pVFrame)) { //rescale+convert it to what we're rendering (no more yuv) ctx->pScaleCtx = sws_getCachedContext(ctx->pScaleCtx, ctx->pVCodecCtx->width, ctx->pVCodecCtx->height, ctx->pVCodecCtx->pix_fmt, ctx->width, ctx->height, AV_PIX_FMT_BGRA, SWS_POINT, 0, 0, 0); sws_scale(ctx->pScaleCtx, (void*)ctx->pVFrame->data, ctx->pVFrame->linesize, 0, ctx->pVCodecCtx->height, &ctx->rgb_data, &ctx->rgb_linesize); ctx->lasttime = ctx->pVFrame->best_effort_timestamp; repainted = true; } } else if(packet.stream_index==ctx->audioStream && !nosound) { avcodec_send_packet(ctx->pACodecCtx, &packet); while(0==avcodec_receive_frame(ctx->pACodecCtx, ctx->pAFrame)) { int width = 2; int channels = ctx->pACodecCtx->channels; unsigned int auddatasize = av_samples_get_buffer_size(NULL, ctx->pACodecCtx->channels, ctx->pAFrame->nb_samples, ctx->pACodecCtx->sample_fmt, 1); void *auddata = ctx->pAFrame->data[0]; switch(ctx->pACodecCtx->sample_fmt) { default: auddatasize = 0; break; case AV_SAMPLE_FMT_U8P: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_U8: width = 1; break; case AV_SAMPLE_FMT_S16P: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_S16: width = 2; break; case AV_SAMPLE_FMT_FLTP: if (channels == 2) { #ifdef MIXER_F32 float *l = (float*)ctx->pAFrame->data[0], *r = (float*)ctx->pAFrame->data[1], *t; unsigned int i; unsigned int frames = ctx->pAFrame->nb_samples; width = sizeof(*t); auddatasize = frames*width*channels; t = malloc(auddatasize); for (i = 0; i < frames; i++) { t[2*i+0] = l[i]; t[2*i+1] = r[i]; } audiofuncs->RawAudio(-1, t, ctx->pACodecCtx->sample_rate, auddatasize/(channels*width), channels, width, 1); free(t); continue; #else //note that we can only reformat in place because we are NOT outputting floats. float *in[2] = {(float*)ctx->pAFrame->data[0],(float*)ctx->pAFrame->data[1]}; signed short *out = (void*)auddata; int v; unsigned int i, c; unsigned int frames = ctx->pAFrame->nb_samples; for (i = 0; i < frames; i++) { for (c = 0; c < 2; c++) { v = (short)(in[c][i]*32767); if (v < -32767) v = -32767; else if (v > 32767) v = 32767; *out++ = v; } } width = sizeof(*out); auddatasize = frames*width*channels; break; #endif } auddatasize /= channels; channels = 1; //fallthrough, using just the first channel as mono case AV_SAMPLE_FMT_FLT: #ifdef MIXER_F32 width = 4; #else { float *in = (void*)auddata; signed short *out = (void*)auddata; int v; unsigned int i; for (i = 0; i < auddatasize/sizeof(*in); i++) { v = (short)(in[i]*32767); if (v < -32767) v = -32767; else if (v > 32767) v = 32767; out[i] = v; } auddatasize/=2; width = 2; } #endif break; case AV_SAMPLE_FMT_DBLP: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_DBL: { double *in = (double*)auddata; signed short *out = (void*)auddata; int v; unsigned int i; for (i = 0; i < auddatasize/sizeof(*in); i++) { v = (short)(in[i]*32767); if (v < -32767) v = -32767; else if (v > 32767) v = 32767; out[i] = v; } auddatasize/=4; width = 2; } break; } audiofuncs->RawAudio(-1, auddata, ctx->pACodecCtx->sample_rate, auddatasize/(channels*width), channels, width, 1); } } av_packet_unref(&packet); #else // Is this a packet from the video stream? if(packet.stream_index==ctx->videoStream) { // Decode video frame avcodec_decode_video2(ctx->pVCodecCtx, ctx->pVFrame, &frameFinished, &packet); // Did we get a video frame? if(frameFinished) { ctx->pScaleCtx = sws_getCachedContext(ctx->pScaleCtx, ctx->pVCodecCtx->width, ctx->pVCodecCtx->height, ctx->pVCodecCtx->pix_fmt, ctx->width, ctx->height, AV_PIX_FMT_BGRA, SWS_POINT, 0, 0, 0); // Convert the image from its native format to RGB sws_scale(ctx->pScaleCtx, (void*)ctx->pVFrame->data, ctx->pVFrame->linesize, 0, ctx->pVCodecCtx->height, &ctx->rgb_data, &ctx->rgb_linesize); repainted = true; } #if TARGET_FFMPEG ctx->lasttime = av_frame_get_best_effort_timestamp(ctx->pVFrame); #else if(frameFinished) { if (ctx->pVFrame->pkt_pts != AV_NOPTS_VALUE) ctx->lasttime = ctx->pVFrame->pkt_pts; else ctx->lasttime = ctx->pVFrame->pkt_dts; } #endif } else if(packet.stream_index==ctx->audioStream && !nosound) { int okay; int len; void *odata = packet.data; while (packet.size > 0) { okay = false; len = avcodec_decode_audio4(ctx->pACodecCtx, ctx->pAFrame, &okay, &packet); if (len < 0) break; packet.size -= len; packet.data += len; if (okay) { int width = 2; int channels = ctx->pACodecCtx->channels; unsigned int auddatasize = av_samples_get_buffer_size(NULL, ctx->pACodecCtx->channels, ctx->pAFrame->nb_samples, ctx->pACodecCtx->sample_fmt, 1); void *auddata = ctx->pAFrame->data[0]; switch(ctx->pACodecCtx->sample_fmt) { default: auddatasize = 0; break; case AV_SAMPLE_FMT_U8P: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_U8: width = 1; break; case AV_SAMPLE_FMT_S16P: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_S16: width = 2; break; case AV_SAMPLE_FMT_FLTP: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_FLT: //FIXME: support float audio internally. { float *in = (void*)auddata; signed short *out = (void*)auddata; int v; unsigned int i; for (i = 0; i < auddatasize/sizeof(*in); i++) { v = (short)(in[i]*32767); if (v < -32767) v = -32767; else if (v > 32767) v = 32767; out[i] = v; } auddatasize/=2; width = 2; } case AV_SAMPLE_FMT_DBLP: auddatasize /= channels; channels = 1; case AV_SAMPLE_FMT_DBL: { double *in = (double*)auddata; signed short *out = (void*)auddata; int v; unsigned int i; for (i = 0; i < auddatasize/sizeof(*in); i++) { v = (short)(in[i]*32767); if (v < -32767) v = -32767; else if (v > 32767) v = 32767; out[i] = v; } auddatasize/=4; width = 2; } break; } audiofuncs->RawAudio(-1, auddata, ctx->pACodecCtx->sample_rate, auddatasize/(channels*width), channels, width, 1); } } packet.data = odata; } // Free the packet that was allocated by av_read_frame av_packet_unref(&packet); #endif } if (forcevideo || repainted) uploadtexture(ectx, TF_BGRA32, ctx->width, ctx->height, ctx->rgb_data, NULL); return true; } static void AVDec_GetSize (void *vctx, int *width, int *height) { struct decctx *ctx = (struct decctx*)vctx; *width = ctx->width; *height = ctx->height; } /*static void AVDec_CursorMove (void *vctx, float posx, float posy) { //its a video, dumbass } static void AVDec_Key (void *vctx, int code, int unicode, int isup) { //its a video, dumbass } static void AVDec_ChangeStream(void *vctx, char *newstream) { } */ static void AVDec_Rewind(void *vctx) { struct decctx *ctx = (struct decctx*)vctx; if (ctx->lasttime != -1) { av_seek_frame(ctx->pFormatCtx, -1, 0, AVSEEK_FLAG_FRAME|AVSEEK_FLAG_BACKWARD); avcodec_flush_buffers(ctx->pVCodecCtx); } ctx->lasttime = -1; } /* //avcodec has no way to shut down properly. static qboolean AVDec_Shutdown(void) { return 0; } */ static media_decoder_funcs_t decoderfuncs = { sizeof(media_decoder_funcs_t), DECODERNAME, AVDec_Create, AVDec_DisplayFrame, AVDec_Destroy, AVDec_Rewind, NULL,//AVDec_CursorMove, NULL,//AVDec_Key, NULL,//AVDec_SetSize, AVDec_GetSize, NULL,//AVDec_ChangeStream }; qboolean AVDec_Init(void) { filefuncs = plugfuncs->GetEngineInterface(plugfsfuncs_name, sizeof(*filefuncs)); audiofuncs = plugfuncs->GetEngineInterface(plugaudiofuncs_name, sizeof(*audiofuncs)); if (!filefuncs || !plugfuncs->ExportInterface("Media_VideoDecoder", &decoderfuncs, sizeof(decoderfuncs))) { Con_Printf(DECODERNAME": Engine doesn't support media decoder plugins\n"); return false; } return true; }