// Copyright (c) 2008-2009, Karl Blomster // All rights reserved. // // Redistribution and use in source and binary forms, with or without // modification, are permitted provided that the following conditions are met: // // * Redistributions of source code must retain the above copyright notice, // this list of conditions and the following disclaimer. // * Redistributions in binary form must reproduce the above copyright notice, // this list of conditions and the following disclaimer in the documentation // and/or other materials provided with the distribution. // * Neither the name of the Aegisub Group nor the names of its contributors // may be used to endorse or promote products derived from this software // without specific prior written permission. // // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" // AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE // IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE // ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE // LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR // CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF // SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS // INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN // CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) // ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE // POSSIBILITY OF SUCH DAMAGE. // // Aegisub Project http://www.aegisub.org/ /// @file video_provider_ffmpegsource.cpp /// @brief FFmpegSource2-based video provider /// @ingroup video_input ffms /// #ifdef WITH_FFMS2 #include "ffmpegsource_common.h" #include "include/aegisub/video_provider.h" #include "options.h" #include "utils.h" #include "video_frame.h" #include #include #include namespace { typedef enum AGI_ColorSpaces { AGI_CS_RGB = 0, AGI_CS_BT709 = 1, AGI_CS_UNSPECIFIED = 2, AGI_CS_FCC = 4, AGI_CS_BT470BG = 5, AGI_CS_SMPTE170M = 6, AGI_CS_SMPTE240M = 7, AGI_CS_YCOCG = 8, AGI_CS_BT2020_NCL = 9, AGI_CS_BT2020_CL = 10, AGI_CS_SMPTE2085 = 11, AGI_CS_CHROMATICITY_DERIVED_NCL = 12, AGI_CS_CHROMATICITY_DERIVED_CL = 13, AGI_CS_ICTCP = 14 } AGI_ColorSpaces; /// @class FFmpegSourceVideoProvider /// @brief Implements video loading through the FFMS library. class FFmpegSourceVideoProvider final : public VideoProvider, FFmpegSourceProvider { /// video source object agi::scoped_holder VideoSource; const FFMS_VideoProperties *VideoInfo = nullptr; ///< video properties int Width = -1; ///< width in pixels int Height = -1; ///< height in pixels int CS = -1; ///< Reported colorspace of first frame int CR = -1; ///< Reported colorrange of first frame double DAR; ///< display aspect ratio std::vector KeyFramesList; ///< list of keyframes agi::vfr::Framerate Timecodes; ///< vfr object std::string ColorSpace; ///< Colorspace name std::string RealColorSpace; ///< Colorspace name char FFMSErrMsg[1024]; ///< FFMS error message FFMS_ErrorInfo ErrInfo; ///< FFMS error codes/messages bool has_audio = false; void LoadVideo(agi::fs::path const& filename, std::string const& colormatrix); public: FFmpegSourceVideoProvider(agi::fs::path const& filename, std::string const& colormatrix, agi::BackgroundRunner *br); void GetFrame(int n, VideoFrame &out) override; void SetColorSpace(std::string const& matrix) override { if (matrix == ColorSpace) return; if (matrix == RealColorSpace) FFMS_SetInputFormatV(VideoSource, CS, CR, FFMS_GetPixFmt(""), nullptr); else if (matrix == "TV.601") FFMS_SetInputFormatV(VideoSource, AGI_CS_BT470BG, CR, FFMS_GetPixFmt(""), nullptr); else return; ColorSpace = matrix; } int GetFrameCount() const override { return VideoInfo->NumFrames; } int GetWidth() const override { return Width; } int GetHeight() const override { return Height; } double GetDAR() const override { return DAR; } agi::vfr::Framerate GetFPS() const override { return Timecodes; } std::string GetColorSpace() const override { return ColorSpace; } std::string GetRealColorSpace() const override { return RealColorSpace; } std::vector GetKeyFrames() const override { return KeyFramesList; }; std::string GetDecoderName() const override { return "FFmpegSource"; } bool WantsCaching() const override { return true; } bool HasAudio() const override { return has_audio; } }; std::string colormatrix_description(int cs, int cr) { // Assuming TV for unspecified std::string str = cr == FFMS_CR_JPEG ? "PC" : "TV"; switch (cs) { case AGI_CS_RGB: return "None"; case AGI_CS_BT709: return str + ".709"; case AGI_CS_FCC: return str + ".FCC"; case AGI_CS_BT470BG: case AGI_CS_SMPTE170M: return str + ".601"; case AGI_CS_SMPTE240M: return str + ".240M"; default: throw VideoOpenError("Unknown video color space"); } } FFmpegSourceVideoProvider::FFmpegSourceVideoProvider(agi::fs::path const& filename, std::string const& colormatrix, agi::BackgroundRunner *br) try : FFmpegSourceProvider(br) , VideoSource(nullptr, FFMS_DestroyVideoSource) { ErrInfo.Buffer = FFMSErrMsg; ErrInfo.BufferSize = sizeof(FFMSErrMsg); ErrInfo.ErrorType = FFMS_ERROR_SUCCESS; ErrInfo.SubType = FFMS_ERROR_SUCCESS; SetLogLevel(); LoadVideo(filename, colormatrix); } catch (agi::EnvironmentError const& err) { throw VideoOpenError(err.GetMessage()); } void FFmpegSourceVideoProvider::LoadVideo(agi::fs::path const& filename, std::string const& colormatrix) { FFMS_Indexer *Indexer = FFMS_CreateIndexer(filename.string().c_str(), &ErrInfo); if (!Indexer) { if (ErrInfo.SubType == FFMS_ERROR_FILE_READ) throw agi::fs::FileNotFound(std::string(ErrInfo.Buffer)); else throw VideoNotSupported(ErrInfo.Buffer); } const char* container = FFMS_GetFormatNameI(Indexer); std::map TrackList = GetTracksOfType(Indexer, FFMS_TYPE_VIDEO); if (TrackList.size() <= 0) throw VideoNotSupported("no video tracks found"); int TrackNumber = -1; if (TrackList.size() > 1) { auto Selection = AskForTrackSelection(TrackList, FFMS_TYPE_VIDEO); if (Selection == TrackSelection::None) throw agi::UserCancelException("video loading cancelled by user"); TrackNumber = static_cast(Selection); } // generate a name for the cache file auto CacheName = GetCacheFilename(filename); // try to read index agi::scoped_holder Index(FFMS_ReadIndex(CacheName.string().c_str(), &ErrInfo), FFMS_DestroyIndex); if (Index && FFMS_IndexBelongsToFile(Index, filename.string().c_str(), &ErrInfo)) Index = nullptr; // time to examine the index and check if the track we want is indexed // technically this isn't really needed since all video tracks should always be indexed, // but a bit of sanity checking never hurt anyone if (Index && TrackNumber >= 0) { FFMS_Track *TempTrackData = FFMS_GetTrackFromIndex(Index, TrackNumber); if (FFMS_GetNumFrames(TempTrackData) <= 0) Index = nullptr; } // moment of truth if (!Index) { auto TrackMask = TrackSelection::None; if (OPT_GET("Provider/FFmpegSource/Index All Tracks")->GetBool() || OPT_GET("Video/Open Audio")->GetBool()) TrackMask = TrackSelection::All; Index = DoIndexing(Indexer, CacheName, TrackMask, GetErrorHandlingMode()); } else { FFMS_CancelIndexing(Indexer); } // update access time of index file so it won't get cleaned away agi::fs::Touch(CacheName); // we have now read the index and may proceed with cleaning the index cache CleanCache(); // track number still not set? if (TrackNumber < 0) { // just grab the first track TrackNumber = FFMS_GetFirstIndexedTrackOfType(Index, FFMS_TYPE_VIDEO, &ErrInfo); if (TrackNumber < 0) throw VideoNotSupported(std::string("Couldn't find any video tracks: ") + ErrInfo.Buffer); } // Check if there's an audio track has_audio = FFMS_GetFirstTrackOfType(Index, FFMS_TYPE_AUDIO, nullptr) != -1; // set thread count int Threads = OPT_GET("Provider/Video/FFmpegSource/Decoding Threads")->GetInt(); // set seekmode // TODO: give this its own option? int SeekMode; if (OPT_GET("Provider/Video/FFmpegSource/Unsafe Seeking")->GetBool()) SeekMode = FFMS_SEEK_UNSAFE; else SeekMode = FFMS_SEEK_NORMAL; VideoSource = FFMS_CreateVideoSource(filename.string().c_str(), TrackNumber, Index, Threads, SeekMode, &ErrInfo); if (!VideoSource) throw VideoOpenError(std::string("Failed to open video track: ") + ErrInfo.Buffer); // load video properties VideoInfo = FFMS_GetVideoProperties(VideoSource); const FFMS_Frame *TempFrame = FFMS_GetFrame(VideoSource, 0, &ErrInfo); if (!TempFrame) throw VideoOpenError(std::string("Failed to decode first frame: ") + ErrInfo.Buffer); Width = TempFrame->EncodedWidth; Height = TempFrame->EncodedHeight; if (VideoInfo->SARDen > 0 && VideoInfo->SARNum > 0) DAR = double(Width) * VideoInfo->SARNum / ((double)Height * VideoInfo->SARDen); else DAR = double(Width) / Height; int VideoCS = CS = TempFrame->ColorSpace; CR = TempFrame->ColorRange; if (CS == AGI_CS_UNSPECIFIED) CS = Width > 1024 || Height >= 600 ? AGI_CS_BT709 : AGI_CS_BT470BG; RealColorSpace = ColorSpace = colormatrix_description(CS, CR); if (CS != AGI_CS_RGB && CS != AGI_CS_BT470BG && ColorSpace != colormatrix && colormatrix == "TV.601") { CS = AGI_CS_BT470BG; ColorSpace = colormatrix_description(CS, CR); } if (CS != VideoCS) { if (FFMS_SetInputFormatV(VideoSource, CS, CR, FFMS_GetPixFmt(""), &ErrInfo)) throw VideoOpenError(std::string("Failed to set input format: ") + ErrInfo.Buffer); } const int TargetFormat[] = { FFMS_GetPixFmt("bgra"), -1 }; if (FFMS_SetOutputFormatV2(VideoSource, TargetFormat, Width, Height, FFMS_RESIZER_BICUBIC, &ErrInfo)) throw VideoOpenError(std::string("Failed to set output format: ") + ErrInfo.Buffer); // get frame info data FFMS_Track *FrameData = FFMS_GetTrackFromVideo(VideoSource); if (FrameData == nullptr) throw VideoOpenError("failed to get frame data"); const FFMS_TrackTimeBase *TimeBase = FFMS_GetTimeBase(FrameData); if (TimeBase == nullptr) throw VideoOpenError("failed to get track time base"); // build list of keyframes and timecodes std::vector TimecodesVector; for (int CurFrameNum = 0; CurFrameNum < VideoInfo->NumFrames; CurFrameNum++) { const FFMS_FrameInfo *CurFrameData = FFMS_GetFrameInfo(FrameData, CurFrameNum); if (!CurFrameData) throw VideoOpenError("Couldn't get info about frame " + std::to_string(CurFrameNum)); // keyframe? if (CurFrameData->KeyFrame) KeyFramesList.push_back(CurFrameNum); // calculate timestamp and add to timecodes vector int Timestamp = (int)((CurFrameData->PTS * TimeBase->Num) / TimeBase->Den); TimecodesVector.push_back(Timestamp); } if (TimecodesVector.size() < 2) Timecodes = 25.0; else { bool normalize = !boost::algorithm::contains(container, "matroska"); Timecodes = agi::vfr::Framerate(TimecodesVector, normalize); } } void FFmpegSourceVideoProvider::GetFrame(int n, VideoFrame &out) { n = mid(0, n, GetFrameCount() - 1); auto frame = FFMS_GetFrame(VideoSource, n, &ErrInfo); if (!frame) throw VideoDecodeError(std::string("Failed to retrieve frame: ") + ErrInfo.Buffer); out.data.assign(frame->Data[0], frame->Data[0] + frame->Linesize[0] * Height); out.flipped = false; out.width = Width; out.height = Height; out.pitch = frame->Linesize[0]; } } std::unique_ptr CreateFFmpegSourceVideoProvider(agi::fs::path const& path, std::string const& colormatrix, agi::BackgroundRunner *br) { return agi::make_unique(path, colormatrix, br); } #endif /* WITH_FFMS2 */