// Copyright (c) 2008, Karl Blomster // All rights reserved. // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are met: // // * Redistributions of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // * Neither the name of the Aegisub Group nor the names of its contributors // may be used to endorse or promote products derived from this software // without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE // ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE // LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR // CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN // CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) // ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE // POSSIBILITY OF SUCH DAMAGE. // // ----------------------------------------------------------------------------- // // AEGISUB // // Website: http://aegisub.cellosoft.com // Contact: mailto:zeratul@cellosoft.com // #ifdef WITH_FFMPEGSOURCE /////////// // Headers #include "video_provider_ffmpegsource.h" #include #include "video_context.h" #include "options.h" /////////////// // Constructor FFmpegSourceVideoProvider::FFmpegSourceVideoProvider(Aegisub::String filename, double fps) { // initialize ffmpegsource FFMS_Init(); // clean up variables VideoSource = NULL; SWSContext = NULL; BufferRGB = NULL; KeyFramesLoaded = false; FrameAllocated = false; FrameNumber = -1; MessageSize = sizeof(FFMSErrorMessage); // and here we go LoadVideo(filename, fps); } /////////////// // Destructor FFmpegSourceVideoProvider::~FFmpegSourceVideoProvider() { Close(); } /////////////// // Open video void FFmpegSourceVideoProvider::LoadVideo(Aegisub::String filename, double fps) { // make sure we don't have anything messy lying around Close(); wxString FileNameWX(filename.c_str(), wxConvUTF8); // generate a name for the cache file wxString CacheName(filename.c_str()); CacheName.append(_T(".ffindex")); FrameIndex *Index; // try to read index Index = FFMS_ReadIndex(CacheName.char_str(), FFMSErrorMessage, MessageSize); if (Index == NULL) { // reading failed, create index // prepare stuff for callback IndexingProgressDialog Progress; Progress.IndexingCanceled = false; Progress.ProgressDialog = new DialogProgress(NULL, _("Indexing"), &Progress.IndexingCanceled, _("Reading keyframes and timecodes from video"), 0, 1); Progress.ProgressDialog->Show(); Progress.ProgressDialog->SetProgress(0,1); Index = FFMS_MakeIndex(FileNameWX.char_str(), 0, "", FFmpegSourceVideoProvider::UpdateIndexingProgress, &Progress, FFMSErrorMessage, MessageSize); if (Index == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } Progress.ProgressDialog->Destroy(); // write it to disk if (FFMS_WriteIndex(CacheName.char_str(), Index, FFMSErrorMessage, MessageSize)) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } } // set thread count int Threads = Options.AsInt(_T("FFmpegSource decoding threads")); if (Threads < 1) throw _T("FFmpegSource video provider: invalid decoding thread count"); // set seekmode // TODO: give this its own option? int SeekMode; if (Options.AsBool(_T("FFmpeg allow unsafe seeking"))) SeekMode = 2; else SeekMode = 1; // finally create the actual video source VideoSource = FFMS_CreateVideoSource(FileNameWX.char_str(), -1, Index, "", Threads, SeekMode, FFMSErrorMessage, MessageSize); if (VideoSource == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } // load video properties VideoInfo = FFMS_GetVideoProperties(VideoSource); // get frame info data FrameInfoVector *FrameData = FFMS_GetVSTrackIndex(VideoSource, FFMSErrorMessage, MessageSize); if (FrameData == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } const TrackTimeBase *TimeBase = FFMS_GetTimeBase(FrameData, FFMSErrorMessage, MessageSize); if (TimeBase == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } const FrameInfo *CurFrameData; // build list of keyframes and timecodes for (int CurFrameNum = 0; CurFrameNum < VideoInfo->NumFrames; CurFrameNum++) { CurFrameData = FFMS_GetFrameInfo(FrameData, CurFrameNum, FFMSErrorMessage, MessageSize); if (CurFrameData == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } // keyframe? if (CurFrameData->KeyFrame) KeyFramesList.Add(CurFrameNum); // calculate timestamp and add to timecodes vector int64_t Timestamp = (int64_t)((CurFrameData->DTS * TimeBase->Num) / (double)TimeBase->Den); // dumb to cast this to an int but the vfr functions all want it as that TimecodesVector.push_back(int(Timestamp)); } KeyFramesLoaded = true; // override already loaded timecodes? Timecodes.SetVFR(TimecodesVector); int OverrideTC = wxYES; if (VFR_Output.IsLoaded()) { OverrideTC = wxMessageBox(_("You already have timecodes loaded. Would you like to replace them with timecodes from the video file?"), _("Replace timecodes?"), wxYES_NO | wxICON_QUESTION); if (OverrideTC == wxYES) { VFR_Input.SetVFR(TimecodesVector); VFR_Output.SetVFR(TimecodesVector); } } else { // no timecodes loaded, go ahead and apply VFR_Input.SetVFR(TimecodesVector); VFR_Output.SetVFR(TimecodesVector); } // we don't need this anymore FFMS_DestroyFrameIndex(Index); FrameNumber = 0; } /////////////// // Close video void FFmpegSourceVideoProvider::Close() { if (SWSContext) sws_freeContext(SWSContext); SWSContext = NULL; if (VideoSource) FFMS_DestroyVideoSource(VideoSource); VideoSource = NULL; if (FrameAllocated) avpicture_free(&FrameRGB); FrameAllocated = false; if (BufferRGB) delete BufferRGB; KeyFramesLoaded = false; KeyFramesList.clear(); TimecodesVector.clear(); FrameNumber = -1; } /////////////// // Update indexing progress int __stdcall FFmpegSourceVideoProvider::UpdateIndexingProgress(int State, int64_t Current, int64_t Total, void *Private) { IndexingProgressDialog *Progress = (IndexingProgressDialog *)Private; Progress->ProgressDialog->SetProgress(Current, Total); if (Progress->IndexingCanceled) { // Close(); return 1; } else { return 0; } } /////////////// // Get frame const AegiVideoFrame FFmpegSourceVideoProvider::GetFrame(int _n, int FormatType) { // don't try to seek to insane places int n = _n; if (n < 0) n = 0; if (n >= GetFrameCount()) n = GetFrameCount()-1; // set position FrameNumber = n; const AVFrameLite *SrcFrame = FFMS_GetFrame(VideoSource, n, FFMSErrorMessage, MessageSize); if (SrcFrame == NULL) { ErrorMsg.Printf(_T("FFmpegSource video provider: %s"), FFMSErrorMessage); throw ErrorMsg; } AVPicture *SrcPicture = reinterpret_cast(const_cast(SrcFrame)); // prepare stuff for conversion to RGB32 int w = VideoInfo->Width; int h = VideoInfo->Height; PixelFormat DstFormat; switch (FormatType) { case FORMAT_RGB32: DstFormat = PIX_FMT_RGB32; break; case FORMAT_RGB24: DstFormat = PIX_FMT_RGB24; break; case FORMAT_YV12: DstFormat = PIX_FMT_YUV420P; break; // may or may not work case FORMAT_YUY2: DstFormat = PIX_FMT_YUYV422; break; default: throw _T("FFmpegSource video provider: upstream provider requested unknown or unsupported pixel format"); } if (!SWSContext) { if (avpicture_alloc(&FrameRGB, DstFormat, w, h) != 0) throw _T("FFmpegSource video provider: could not allocate output picture buffer"); FrameAllocated = true; unsigned int DstSize = avpicture_get_size(DstFormat,w,h); BufferRGB = new uint8_t[DstSize]; // initialize swscaler context SWSContext = sws_getContext(w, h, VideoInfo->PixelFormat, w, h, DstFormat, SWS_BICUBIC, NULL, NULL, NULL); if (SWSContext == NULL) throw _T("FFmpegSource video provider: failed to initialize SWScale colorspace conversion"); } avpicture_fill(&FrameRGB, BufferRGB, DstFormat, w, h); // this is what we'll return eventually AegiVideoFrame &DstFrame = CurFrame; // set some properties DstFrame.w = w; DstFrame.h = h; DstFrame.flipped = false; if (FormatType == FORMAT_RGB32 || FormatType == FORMAT_RGB24) { DstFrame.invertChannels = true; } else { DstFrame.invertChannels = false; } DstFrame.format = (VideoFrameFormat)FormatType; // allocate destination frame for (int i=0;i<4;i++) DstFrame.pitch[i] = FrameRGB.linesize[i]; DstFrame.Allocate(); // let swscale do the conversion to RGB and write directly to the output frame sws_scale(SWSContext, SrcPicture->data, SrcPicture->linesize, 0, h, DstFrame.data, FrameRGB.linesize); return DstFrame; } /////////////// // Utility functions int FFmpegSourceVideoProvider::GetWidth() { return VideoInfo->Width; } int FFmpegSourceVideoProvider::GetHeight() { return VideoInfo->Height; } int FFmpegSourceVideoProvider::GetFrameCount() { return VideoInfo->NumFrames; } int FFmpegSourceVideoProvider::GetPosition() { return FrameNumber; } double FFmpegSourceVideoProvider::GetFPS() { return double(VideoInfo->FPSNumerator) / double(VideoInfo->FPSDenominator); } #endif /* WITH_FFMPEGSOURCE */