// Copyright (c) 2007-2008 Fredrik Mellbin // // Permission is hereby granted, free of charge, to any person obtaining a copy // of this software and associated documentation files (the "Software"), to deal // in the Software without restriction, including without limitation the rights // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell // copies of the Software, and to permit persons to whom the Software is // furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN // THE SOFTWARE. #include "ffmpegsource.h" int FFBase::FrameFromDTS(int64_t ADTS) { for (int i = 0; i < (int)Frames.size(); i++) if (Frames[i].DTS == ADTS) return i; return -1; } int FFBase::ClosestFrameFromDTS(int64_t ADTS) { int Frame = 0; int64_t BestDiff = 0xFFFFFFFFFFFFFFLL; // big number for (int i = 0; i < (int)Frames.size(); i++) { int64_t CurrentDiff = FFABS(Frames[i].DTS - ADTS); if (CurrentDiff < BestDiff) { BestDiff = CurrentDiff; Frame = i; } } return Frame; } int FFBase::FindClosestKeyFrame(int AFrame) { for (int i = AFrame; i > 0; i--) if (Frames[i].KeyFrame) return i; return 0; } bool FFBase::LoadFrameInfoFromFile(const char *AVideoCacheFile, const char *ASource, int AVideoTrack) { char DefaultCacheFilename[1024]; sprintf(DefaultCacheFilename, "%s.ffv%dcache", ASource, AVideoTrack); if (!strcmp(AVideoCacheFile, "")) AVideoCacheFile = DefaultCacheFilename; FILE *CacheFile = fopen(AVideoCacheFile, "r"); if (!CacheFile) return false; if (fscanf(CacheFile, "%d\r\n", &VI.num_frames) <= 0 || VI.num_frames <= 0) { VI.num_frames = 0; fclose(CacheFile); return false; } for (int i = 0; i < VI.num_frames; i++) { int64_t DTSTemp; int KFTemp; fscanf(CacheFile, "%lld %d\r\n", &DTSTemp, &KFTemp); Frames.push_back(FrameInfo(DTSTemp, KFTemp != 0)); } fclose(CacheFile); return true; } bool FFBase::SaveFrameInfoToFile(const char *AVideoCacheFile, const char *ASource, int AVideoTrack) { char DefaultCacheFilename[1024]; sprintf(DefaultCacheFilename, "%s.ffv%dcache", ASource, AVideoTrack); if (!strcmp(AVideoCacheFile, "")) AVideoCacheFile = DefaultCacheFilename; FILE *CacheFile = fopen(AVideoCacheFile, "wb"); if (!CacheFile) return false; fprintf(CacheFile, "%d\r\n", VI.num_frames); for (int i = 0; i < VI.num_frames; i++) fprintf(CacheFile, "%lld %d\r\n", Frames[i].DTS, (int)(Frames[i].KeyFrame ? 1 : 0)); fclose(CacheFile); return true; } bool FFBase::SaveTimecodesToFile(const char *ATimecodeFile, int64_t ScaleD, int64_t ScaleN) { if (!strcmp(ATimecodeFile, "")) return true; FILE *TimecodeFile = fopen(ATimecodeFile, "wb"); if (!TimecodeFile) return false; std::set Timecodes; for (int i = 0; i < VI.num_frames; i++) Timecodes.insert(Frames[i].DTS); fprintf(TimecodeFile, "# timecode format v2\r\n"); for (std::set::iterator Cur=Timecodes.begin(); Cur!=Timecodes.end(); Cur++) fprintf(TimecodeFile, "%f\r\n", (*Cur * ScaleD) / (double)ScaleN); fclose(TimecodeFile); return true; } bool FFBase::OpenAudioCache(const char *AAudioCacheFile, const char *ASource, int AAudioTrack, IScriptEnvironment *Env) { char DefaultCacheFilename[1024]; sprintf(DefaultCacheFilename, "%s.ffa%dcache", ASource, AAudioTrack); if (!strcmp(AAudioCacheFile, "")) AAudioCacheFile = DefaultCacheFilename; // Is an empty file? FILE *FCFile = fopen(AAudioCacheFile, "rb"); int64_t CacheSize; if (FCFile) { _fseeki64(FCFile, 0, SEEK_END); CacheSize = _ftelli64(FCFile); _fseeki64(FCFile, 0, SEEK_SET); if (CacheSize <= 0) { fclose(FCFile); FCFile = NULL; return false; } } else { return false; } // Raw audio VI.num_audio_samples = VI.AudioSamplesFromBytes(CacheSize); AudioCacheType = acRaw; RawAudioCache = FCFile; FCFile = NULL; return true; } FILE *FFBase::NewRawCacheWriter(const char *AAudioCacheFile, const char *ASource, int AAudioTrack, IScriptEnvironment *Env) { char DefaultCacheFilename[1024]; sprintf(DefaultCacheFilename, "%s.ffa%dcache", ASource, AAudioTrack); if (!strcmp(AAudioCacheFile, "")) AAudioCacheFile = DefaultCacheFilename; FILE *RCF = fopen(AAudioCacheFile, "wb"); if (RCF == NULL) Env->ThrowError("FFmpegSource: Failed to open '%s' for writing", AAudioCacheFile); return RCF; } void FFBase::CloseRawCacheWriter(FILE *ARawCache) { fclose(ARawCache); } void FFBase::InitPP(int AWidth, int AHeight, const char *APPString, int AQuality, int APixelFormat, IScriptEnvironment *Env) { if (!strcmp(APPString, "")) return; if (AQuality < 0 || AQuality > PP_QUALITY_MAX) Env->ThrowError("FFmpegSource: Quality is out of range"); // Unsafe? PPMode = pp_get_mode_by_name_and_quality(APPString, AQuality); if (!PPMode) Env->ThrowError("FFmpegSource: Invalid postprocesing settings"); int Flags = GetPPCPUFlags(Env); switch (APixelFormat) { case PIX_FMT_YUV420P: Flags |= PP_FORMAT_420; break; case PIX_FMT_YUV422P: Flags |= PP_FORMAT_422; break; case PIX_FMT_YUV411P: Flags |= PP_FORMAT_411; break; case PIX_FMT_YUV444P: Flags |= PP_FORMAT_444; break; default: Env->ThrowError("FFmpegSource: Input format is not supported for postprocessing"); } PPContext = pp_get_context(VI.width, VI.height, Flags); if (avpicture_alloc(&PPPicture, APixelFormat, AWidth, AHeight) < 0) Env->ThrowError("FFmpegSource: Failed to allocate picture"); } void FFBase::SetOutputFormat(int ACurrentFormat, IScriptEnvironment *Env) { int Loss; int BestFormat = avcodec_find_best_pix_fmt((1 << PIX_FMT_YUVJ420P) | (1 << PIX_FMT_YUV420P) | (1 << PIX_FMT_YUYV422) | (1 << PIX_FMT_RGB32) | (1 << PIX_FMT_BGR24), ACurrentFormat, 1 /* Required to prevent pointless RGB32 => RGB24 conversion */, &Loss); switch (BestFormat) { case PIX_FMT_YUVJ420P: // stupid yv12 distinctions, also inexplicably completely undeniably incompatible with all other supported output formats case PIX_FMT_YUV420P: VI.pixel_type = VideoInfo::CS_I420; break; case PIX_FMT_YUYV422: VI.pixel_type = VideoInfo::CS_YUY2; break; case PIX_FMT_RGB32: VI.pixel_type = VideoInfo::CS_BGR32; break; case PIX_FMT_BGR24: VI.pixel_type = VideoInfo::CS_BGR24; break; default: Env->ThrowError("FFmpegSource: No suitable output format found"); } if (BestFormat != ACurrentFormat) { ConvertToFormat = BestFormat; SWS = sws_getContext(VI.width, VI.height, ACurrentFormat, VI.width, VI.height, ConvertToFormat, GetSWSCPUFlags(Env) | SWS_BICUBIC, NULL, NULL, NULL); } if (BestFormat == PIX_FMT_YUVJ420P || BestFormat == PIX_FMT_YUV420P) { VI.height -= VI.height & 1; VI.width -= VI.width & 1; } if (BestFormat == PIX_FMT_YUYV422) { VI.width -= VI.width & 1; } } PVideoFrame FFBase::OutputFrame(AVFrame *AFrame, IScriptEnvironment *Env) { AVPicture *SrcPicture = (AVPicture *)AFrame; if (PPContext) { pp_postprocess(const_cast(AFrame->data), AFrame->linesize, PPPicture.data, PPPicture.linesize, VI.width, VI.height, AFrame->qscale_table, AFrame->qstride, PPMode, PPContext, AFrame->pict_type | (AFrame->qscale_type ? PP_PICT_TYPE_QP2 : 0)); SrcPicture = &PPPicture; } PVideoFrame Dst = Env->NewVideoFrame(VI); if (ConvertToFormat != PIX_FMT_NONE && VI.pixel_type == VideoInfo::CS_I420) { uint8_t *DstData[3] = {Dst->GetWritePtr(PLANAR_Y), Dst->GetWritePtr(PLANAR_U), Dst->GetWritePtr(PLANAR_V)}; int DstStride[3] = {Dst->GetPitch(PLANAR_Y), Dst->GetPitch(PLANAR_U), Dst->GetPitch(PLANAR_V)}; sws_scale(SWS, SrcPicture->data, SrcPicture->linesize, 0, VI.height, DstData, DstStride); } else if (ConvertToFormat != PIX_FMT_NONE) { if (VI.IsRGB()) { uint8_t *DstData[1] = {Dst->GetWritePtr() + Dst->GetPitch() * (Dst->GetHeight() - 1)}; int DstStride[1] = {-Dst->GetPitch()}; sws_scale(SWS, SrcPicture->data, SrcPicture->linesize, 0, VI.height, DstData, DstStride); } else { uint8_t *DstData[1] = {Dst->GetWritePtr()}; int DstStride[1] = {Dst->GetPitch()}; sws_scale(SWS, SrcPicture->data, SrcPicture->linesize, 0, VI.height, DstData, DstStride); } } else if (VI.pixel_type == VideoInfo::CS_I420) { Env->BitBlt(Dst->GetWritePtr(PLANAR_Y), Dst->GetPitch(PLANAR_Y), SrcPicture->data[0], SrcPicture->linesize[0], Dst->GetRowSize(PLANAR_Y), Dst->GetHeight(PLANAR_Y)); Env->BitBlt(Dst->GetWritePtr(PLANAR_U), Dst->GetPitch(PLANAR_U), SrcPicture->data[1], SrcPicture->linesize[1], Dst->GetRowSize(PLANAR_U), Dst->GetHeight(PLANAR_U)); Env->BitBlt(Dst->GetWritePtr(PLANAR_V), Dst->GetPitch(PLANAR_V), SrcPicture->data[2], SrcPicture->linesize[2], Dst->GetRowSize(PLANAR_V), Dst->GetHeight(PLANAR_V)); } else { if (VI.IsRGB()) Env->BitBlt(Dst->GetWritePtr() + Dst->GetPitch() * (Dst->GetHeight() - 1), -Dst->GetPitch(), SrcPicture->data[0], SrcPicture->linesize[0], Dst->GetRowSize(), Dst->GetHeight()); else Env->BitBlt(Dst->GetWritePtr(), Dst->GetPitch(), SrcPicture->data[0], SrcPicture->linesize[0], Dst->GetRowSize(), Dst->GetHeight()); } return Dst; } void FFBase::GetAudio(void* Buf, __int64 Start, __int64 Count, IScriptEnvironment* Env) { if (AudioCacheType == acRaw) { _fseeki64(RawAudioCache, VI.BytesFromAudioSamples(Start), SEEK_SET); fread(Buf, 1, static_cast(VI.BytesFromAudioSamples(Count)), RawAudioCache); } else { Env->ThrowError("FFmpegSource: Audio requested but none available"); } } FFBase::FFBase() { memset(&VI, 0, sizeof(VI)); AudioCacheType = acNone; RawAudioCache = NULL; PPContext = NULL; PPMode = NULL; SWS = NULL; LastFrameNum = -1; DecodingBuffer = new uint8_t[AVCODEC_MAX_AUDIO_FRAME_SIZE]; ConvertToFormat = PIX_FMT_NONE; memset(&PPPicture, 0, sizeof(PPPicture)); DecodeFrame = avcodec_alloc_frame(); } FFBase::~FFBase() { delete [] DecodingBuffer; if (RawAudioCache) fclose(RawAudioCache); if (SWS) sws_freeContext(SWS); if (PPMode) pp_free_mode(PPMode); if (PPContext) pp_free_context(PPContext); avpicture_free(&PPPicture); av_free(DecodeFrame); }