mirror of
				https://github.com/nyanmisaka/ffmpeg-rockchip.git
				synced 2025-11-01 04:53:04 +08:00 
			
		
		
		
	
		
			
				
	
	
		
			2038 lines
		
	
	
		
			65 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			2038 lines
		
	
	
		
			65 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * ffmpeg filter configuration
 | |
|  *
 | |
|  * This file is part of FFmpeg.
 | |
|  *
 | |
|  * FFmpeg is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * FFmpeg is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with FFmpeg; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
|  */
 | |
| 
 | |
| #include <stdint.h>
 | |
| 
 | |
| #include "ffmpeg.h"
 | |
| 
 | |
| #include "libavfilter/avfilter.h"
 | |
| #include "libavfilter/buffersink.h"
 | |
| #include "libavfilter/buffersrc.h"
 | |
| 
 | |
| #include "libavutil/avassert.h"
 | |
| #include "libavutil/avstring.h"
 | |
| #include "libavutil/bprint.h"
 | |
| #include "libavutil/channel_layout.h"
 | |
| #include "libavutil/display.h"
 | |
| #include "libavutil/opt.h"
 | |
| #include "libavutil/pixdesc.h"
 | |
| #include "libavutil/pixfmt.h"
 | |
| #include "libavutil/imgutils.h"
 | |
| #include "libavutil/samplefmt.h"
 | |
| #include "libavutil/timestamp.h"
 | |
| 
 | |
| typedef struct FilterGraphPriv {
 | |
|     FilterGraph fg;
 | |
| 
 | |
|     // name used for logging
 | |
|     char log_name[32];
 | |
| 
 | |
|     int is_simple;
 | |
|     // true when the filtergraph contains only meta filters
 | |
|     // that do not modify the frame data
 | |
|     int is_meta;
 | |
| 
 | |
|     const char *graph_desc;
 | |
| 
 | |
|     // frame for temporarily holding output from the filtergraph
 | |
|     AVFrame *frame;
 | |
| } FilterGraphPriv;
 | |
| 
 | |
| static FilterGraphPriv *fgp_from_fg(FilterGraph *fg)
 | |
| {
 | |
|     return (FilterGraphPriv*)fg;
 | |
| }
 | |
| 
 | |
| static const FilterGraphPriv *cfgp_from_cfg(const FilterGraph *fg)
 | |
| {
 | |
|     return (const FilterGraphPriv*)fg;
 | |
| }
 | |
| 
 | |
| typedef struct InputFilterPriv {
 | |
|     InputFilter ifilter;
 | |
| 
 | |
|     AVFilterContext *filter;
 | |
| 
 | |
|     InputStream *ist;
 | |
| 
 | |
|     // used to hold submitted input
 | |
|     AVFrame *frame;
 | |
| 
 | |
|     /* for filters that are not yet bound to an input stream,
 | |
|      * this stores the input linklabel, if any */
 | |
|     uint8_t *linklabel;
 | |
| 
 | |
|     // filter data type
 | |
|     enum AVMediaType type;
 | |
|     // source data type: AVMEDIA_TYPE_SUBTITLE for sub2video,
 | |
|     // same as type otherwise
 | |
|     enum AVMediaType type_src;
 | |
| 
 | |
|     int eof;
 | |
| 
 | |
|     // parameters configured for this input
 | |
|     int format;
 | |
| 
 | |
|     int width, height;
 | |
|     AVRational sample_aspect_ratio;
 | |
| 
 | |
|     int sample_rate;
 | |
|     AVChannelLayout ch_layout;
 | |
| 
 | |
|     AVRational time_base;
 | |
| 
 | |
|     AVFifo *frame_queue;
 | |
| 
 | |
|     AVBufferRef *hw_frames_ctx;
 | |
| 
 | |
|     int     displaymatrix_present;
 | |
|     int32_t displaymatrix[9];
 | |
| 
 | |
|     // fallback parameters to use when no input is ever sent
 | |
|     struct {
 | |
|         int                 format;
 | |
| 
 | |
|         int                 width;
 | |
|         int                 height;
 | |
|         AVRational          sample_aspect_ratio;
 | |
| 
 | |
|         int                 sample_rate;
 | |
|         AVChannelLayout     ch_layout;
 | |
|     } fallback;
 | |
| 
 | |
|     struct {
 | |
|         ///< queue of AVSubtitle* before filter init
 | |
|         AVFifo *queue;
 | |
| 
 | |
|         AVFrame *frame;
 | |
| 
 | |
|         int64_t last_pts;
 | |
|         int64_t end_pts;
 | |
| 
 | |
|         ///< marks if sub2video_update should force an initialization
 | |
|         unsigned int initialize;
 | |
|     } sub2video;
 | |
| } InputFilterPriv;
 | |
| 
 | |
| static InputFilterPriv *ifp_from_ifilter(InputFilter *ifilter)
 | |
| {
 | |
|     return (InputFilterPriv*)ifilter;
 | |
| }
 | |
| 
 | |
| static int configure_filtergraph(FilterGraph *fg);
 | |
| 
 | |
| static int sub2video_get_blank_frame(InputFilterPriv *ifp)
 | |
| {
 | |
|     AVFrame *frame = ifp->sub2video.frame;
 | |
|     int ret;
 | |
| 
 | |
|     av_frame_unref(frame);
 | |
| 
 | |
|     frame->width  = ifp->width;
 | |
|     frame->height = ifp->height;
 | |
|     frame->format = ifp->format;
 | |
| 
 | |
|     ret = av_frame_get_buffer(frame, 0);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     memset(frame->data[0], 0, frame->height * frame->linesize[0]);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void sub2video_copy_rect(uint8_t *dst, int dst_linesize, int w, int h,
 | |
|                                 AVSubtitleRect *r)
 | |
| {
 | |
|     uint32_t *pal, *dst2;
 | |
|     uint8_t *src, *src2;
 | |
|     int x, y;
 | |
| 
 | |
|     if (r->type != SUBTITLE_BITMAP) {
 | |
|         av_log(NULL, AV_LOG_WARNING, "sub2video: non-bitmap subtitle\n");
 | |
|         return;
 | |
|     }
 | |
|     if (r->x < 0 || r->x + r->w > w || r->y < 0 || r->y + r->h > h) {
 | |
|         av_log(NULL, AV_LOG_WARNING, "sub2video: rectangle (%d %d %d %d) overflowing %d %d\n",
 | |
|             r->x, r->y, r->w, r->h, w, h
 | |
|         );
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     dst += r->y * dst_linesize + r->x * 4;
 | |
|     src = r->data[0];
 | |
|     pal = (uint32_t *)r->data[1];
 | |
|     for (y = 0; y < r->h; y++) {
 | |
|         dst2 = (uint32_t *)dst;
 | |
|         src2 = src;
 | |
|         for (x = 0; x < r->w; x++)
 | |
|             *(dst2++) = pal[*(src2++)];
 | |
|         dst += dst_linesize;
 | |
|         src += r->linesize[0];
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void sub2video_push_ref(InputFilterPriv *ifp, int64_t pts)
 | |
| {
 | |
|     AVFrame *frame = ifp->sub2video.frame;
 | |
|     int ret;
 | |
| 
 | |
|     av_assert1(frame->data[0]);
 | |
|     ifp->sub2video.last_pts = frame->pts = pts;
 | |
|     ret = av_buffersrc_add_frame_flags(ifp->filter, frame,
 | |
|                                        AV_BUFFERSRC_FLAG_KEEP_REF |
 | |
|                                        AV_BUFFERSRC_FLAG_PUSH);
 | |
|     if (ret != AVERROR_EOF && ret < 0)
 | |
|         av_log(NULL, AV_LOG_WARNING, "Error while add the frame to buffer source(%s).\n",
 | |
|                av_err2str(ret));
 | |
| }
 | |
| 
 | |
| static void sub2video_update(InputFilterPriv *ifp, int64_t heartbeat_pts,
 | |
|                              const AVSubtitle *sub)
 | |
| {
 | |
|     AVFrame *frame = ifp->sub2video.frame;
 | |
|     int8_t *dst;
 | |
|     int     dst_linesize;
 | |
|     int num_rects, i;
 | |
|     int64_t pts, end_pts;
 | |
| 
 | |
|     if (sub) {
 | |
|         pts       = av_rescale_q(sub->pts + sub->start_display_time * 1000LL,
 | |
|                                  AV_TIME_BASE_Q, ifp->time_base);
 | |
|         end_pts   = av_rescale_q(sub->pts + sub->end_display_time   * 1000LL,
 | |
|                                  AV_TIME_BASE_Q, ifp->time_base);
 | |
|         num_rects = sub->num_rects;
 | |
|     } else {
 | |
|         /* If we are initializing the system, utilize current heartbeat
 | |
|            PTS as the start time, and show until the following subpicture
 | |
|            is received. Otherwise, utilize the previous subpicture's end time
 | |
|            as the fall-back value. */
 | |
|         pts       = ifp->sub2video.initialize ?
 | |
|                     heartbeat_pts : ifp->sub2video.end_pts;
 | |
|         end_pts   = INT64_MAX;
 | |
|         num_rects = 0;
 | |
|     }
 | |
|     if (sub2video_get_blank_frame(ifp) < 0) {
 | |
|         av_log(NULL, AV_LOG_ERROR,
 | |
|                "Impossible to get a blank canvas.\n");
 | |
|         return;
 | |
|     }
 | |
|     dst          = frame->data    [0];
 | |
|     dst_linesize = frame->linesize[0];
 | |
|     for (i = 0; i < num_rects; i++)
 | |
|         sub2video_copy_rect(dst, dst_linesize, frame->width, frame->height, sub->rects[i]);
 | |
|     sub2video_push_ref(ifp, pts);
 | |
|     ifp->sub2video.end_pts = end_pts;
 | |
|     ifp->sub2video.initialize = 0;
 | |
| }
 | |
| 
 | |
| // FIXME: YUV420P etc. are actually supported with full color range,
 | |
| // yet the latter information isn't available here.
 | |
| static const enum AVPixelFormat *get_compliance_normal_pix_fmts(const AVCodec *codec, const enum AVPixelFormat default_formats[])
 | |
| {
 | |
|     static const enum AVPixelFormat mjpeg_formats[] =
 | |
|         { AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P,
 | |
|           AV_PIX_FMT_NONE };
 | |
| 
 | |
|     if (!strcmp(codec->name, "mjpeg")) {
 | |
|         return mjpeg_formats;
 | |
|     } else {
 | |
|         return default_formats;
 | |
|     }
 | |
| }
 | |
| 
 | |
| static enum AVPixelFormat
 | |
| choose_pixel_fmt(const AVCodec *codec, enum AVPixelFormat target,
 | |
|                  int strict_std_compliance)
 | |
| {
 | |
|     if (codec && codec->pix_fmts) {
 | |
|         const enum AVPixelFormat *p = codec->pix_fmts;
 | |
|         const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(target);
 | |
|         //FIXME: This should check for AV_PIX_FMT_FLAG_ALPHA after PAL8 pixel format without alpha is implemented
 | |
|         int has_alpha = desc ? desc->nb_components % 2 == 0 : 0;
 | |
|         enum AVPixelFormat best= AV_PIX_FMT_NONE;
 | |
| 
 | |
|         if (strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL) {
 | |
|             p = get_compliance_normal_pix_fmts(codec, p);
 | |
|         }
 | |
|         for (; *p != AV_PIX_FMT_NONE; p++) {
 | |
|             best = av_find_best_pix_fmt_of_2(best, *p, target, has_alpha, NULL);
 | |
|             if (*p == target)
 | |
|                 break;
 | |
|         }
 | |
|         if (*p == AV_PIX_FMT_NONE) {
 | |
|             if (target != AV_PIX_FMT_NONE)
 | |
|                 av_log(NULL, AV_LOG_WARNING,
 | |
|                        "Incompatible pixel format '%s' for codec '%s', auto-selecting format '%s'\n",
 | |
|                        av_get_pix_fmt_name(target),
 | |
|                        codec->name,
 | |
|                        av_get_pix_fmt_name(best));
 | |
|             return best;
 | |
|         }
 | |
|     }
 | |
|     return target;
 | |
| }
 | |
| 
 | |
| /* May return NULL (no pixel format found), a static string or a string
 | |
|  * backed by the bprint. Nothing has been written to the AVBPrint in case
 | |
|  * NULL is returned. The AVBPrint provided should be clean. */
 | |
| static const char *choose_pix_fmts(OutputFilter *ofilter, AVBPrint *bprint)
 | |
| {
 | |
|     OutputStream *ost = ofilter->ost;
 | |
|     AVCodecContext *enc = ost->enc_ctx;
 | |
|     const AVDictionaryEntry *strict_dict = av_dict_get(ost->encoder_opts, "strict", NULL, 0);
 | |
|     if (strict_dict)
 | |
|         // used by choose_pixel_fmt() and below
 | |
|         av_opt_set(ost->enc_ctx, "strict", strict_dict->value, 0);
 | |
| 
 | |
|      if (ost->keep_pix_fmt) {
 | |
|         avfilter_graph_set_auto_convert(ofilter->graph->graph,
 | |
|                                             AVFILTER_AUTO_CONVERT_NONE);
 | |
|         if (ost->enc_ctx->pix_fmt == AV_PIX_FMT_NONE)
 | |
|             return NULL;
 | |
|         return av_get_pix_fmt_name(ost->enc_ctx->pix_fmt);
 | |
|     }
 | |
|     if (ost->enc_ctx->pix_fmt != AV_PIX_FMT_NONE) {
 | |
|         return av_get_pix_fmt_name(choose_pixel_fmt(enc->codec, enc->pix_fmt,
 | |
|                                                     ost->enc_ctx->strict_std_compliance));
 | |
|     } else if (enc->codec->pix_fmts) {
 | |
|         const enum AVPixelFormat *p;
 | |
| 
 | |
|         p = enc->codec->pix_fmts;
 | |
|         if (ost->enc_ctx->strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL) {
 | |
|             p = get_compliance_normal_pix_fmts(enc->codec, p);
 | |
|         }
 | |
| 
 | |
|         for (; *p != AV_PIX_FMT_NONE; p++) {
 | |
|             const char *name = av_get_pix_fmt_name(*p);
 | |
|             av_bprintf(bprint, "%s%c", name, p[1] == AV_PIX_FMT_NONE ? '\0' : '|');
 | |
|         }
 | |
|         if (!av_bprint_is_complete(bprint))
 | |
|             report_and_exit(AVERROR(ENOMEM));
 | |
|         return bprint->str;
 | |
|     } else
 | |
|         return NULL;
 | |
| }
 | |
| 
 | |
| /* Define a function for appending a list of allowed formats
 | |
|  * to an AVBPrint. If nonempty, the list will have a header. */
 | |
| #define DEF_CHOOSE_FORMAT(name, type, var, supported_list, none, printf_format, get_name) \
 | |
| static void choose_ ## name (OutputFilter *ofilter, AVBPrint *bprint)          \
 | |
| {                                                                              \
 | |
|     if (ofilter->var == none && !ofilter->supported_list)                      \
 | |
|         return;                                                                \
 | |
|     av_bprintf(bprint, #name "=");                                             \
 | |
|     if (ofilter->var != none) {                                                \
 | |
|         av_bprintf(bprint, printf_format, get_name(ofilter->var));             \
 | |
|     } else {                                                                   \
 | |
|         const type *p;                                                         \
 | |
|                                                                                \
 | |
|         for (p = ofilter->supported_list; *p != none; p++) {                   \
 | |
|             av_bprintf(bprint, printf_format "|", get_name(*p));               \
 | |
|         }                                                                      \
 | |
|         if (bprint->len > 0)                                                   \
 | |
|             bprint->str[--bprint->len] = '\0';                                 \
 | |
|     }                                                                          \
 | |
|     av_bprint_chars(bprint, ':', 1);                                           \
 | |
| }
 | |
| 
 | |
| //DEF_CHOOSE_FORMAT(pix_fmts, enum AVPixelFormat, format, formats, AV_PIX_FMT_NONE,
 | |
| //                  GET_PIX_FMT_NAME)
 | |
| 
 | |
| DEF_CHOOSE_FORMAT(sample_fmts, enum AVSampleFormat, format, formats,
 | |
|                   AV_SAMPLE_FMT_NONE, "%s", av_get_sample_fmt_name)
 | |
| 
 | |
| DEF_CHOOSE_FORMAT(sample_rates, int, sample_rate, sample_rates, 0,
 | |
|                   "%d", )
 | |
| 
 | |
| static void choose_channel_layouts(OutputFilter *ofilter, AVBPrint *bprint)
 | |
| {
 | |
|     if (av_channel_layout_check(&ofilter->ch_layout)) {
 | |
|         av_bprintf(bprint, "channel_layouts=");
 | |
|         av_channel_layout_describe_bprint(&ofilter->ch_layout, bprint);
 | |
|     } else if (ofilter->ch_layouts) {
 | |
|         const AVChannelLayout *p;
 | |
| 
 | |
|         av_bprintf(bprint, "channel_layouts=");
 | |
|         for (p = ofilter->ch_layouts; p->nb_channels; p++) {
 | |
|             av_channel_layout_describe_bprint(p, bprint);
 | |
|             av_bprintf(bprint, "|");
 | |
|         }
 | |
|         if (bprint->len > 0)
 | |
|             bprint->str[--bprint->len] = '\0';
 | |
|     } else
 | |
|         return;
 | |
|     av_bprint_chars(bprint, ':', 1);
 | |
| }
 | |
| 
 | |
| static int read_binary(const char *path, uint8_t **data, int *len)
 | |
| {
 | |
|     AVIOContext *io = NULL;
 | |
|     int64_t fsize;
 | |
|     int ret;
 | |
| 
 | |
|     *data = NULL;
 | |
|     *len  = 0;
 | |
| 
 | |
|     ret = avio_open2(&io, path, AVIO_FLAG_READ, &int_cb, NULL);
 | |
|     if (ret < 0) {
 | |
|         av_log(NULL, AV_LOG_ERROR, "Cannot open file '%s': %s\n",
 | |
|                path, av_err2str(ret));
 | |
|         return ret;
 | |
|     }
 | |
| 
 | |
|     fsize = avio_size(io);
 | |
|     if (fsize < 0 || fsize > INT_MAX) {
 | |
|         av_log(NULL, AV_LOG_ERROR, "Cannot obtain size of file %s\n", path);
 | |
|         ret = AVERROR(EIO);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     *data = av_malloc(fsize);
 | |
|     if (!*data) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     ret = avio_read(io, *data, fsize);
 | |
|     if (ret != fsize) {
 | |
|         av_log(NULL, AV_LOG_ERROR, "Error reading file %s\n", path);
 | |
|         ret = ret < 0 ? ret : AVERROR(EIO);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     *len = fsize;
 | |
| 
 | |
|     ret = 0;
 | |
| fail:
 | |
|     avio_close(io);
 | |
|     if (ret < 0) {
 | |
|         av_freep(data);
 | |
|         *len = 0;
 | |
|     }
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int filter_opt_apply(AVFilterContext *f, const char *key, const char *val)
 | |
| {
 | |
|     const AVOption *o = NULL;
 | |
|     int ret;
 | |
| 
 | |
|     ret = av_opt_set(f, key, val, AV_OPT_SEARCH_CHILDREN);
 | |
|     if (ret >= 0)
 | |
|         return 0;
 | |
| 
 | |
|     if (ret == AVERROR_OPTION_NOT_FOUND && key[0] == '/')
 | |
|         o = av_opt_find(f, key + 1, NULL, 0, AV_OPT_SEARCH_CHILDREN);
 | |
|     if (!o)
 | |
|         goto err_apply;
 | |
| 
 | |
|     // key is a valid option name prefixed with '/'
 | |
|     // interpret value as a path from which to load the actual option value
 | |
|     key++;
 | |
| 
 | |
|     if (o->type == AV_OPT_TYPE_BINARY) {
 | |
|         uint8_t *data;
 | |
|         int      len;
 | |
| 
 | |
|         ret = read_binary(val, &data, &len);
 | |
|         if (ret < 0)
 | |
|             goto err_load;
 | |
| 
 | |
|         ret = av_opt_set_bin(f, key, data, len, AV_OPT_SEARCH_CHILDREN);
 | |
|         av_freep(&data);
 | |
|     } else {
 | |
|         char *data = file_read(val);
 | |
|         if (!data) {
 | |
|             ret = AVERROR(EIO);
 | |
|             goto err_load;
 | |
|         }
 | |
| 
 | |
|         ret = av_opt_set(f, key, data, AV_OPT_SEARCH_CHILDREN);
 | |
|         av_freep(&data);
 | |
|     }
 | |
|     if (ret < 0)
 | |
|         goto err_apply;
 | |
| 
 | |
|     return 0;
 | |
| 
 | |
| err_apply:
 | |
|     av_log(NULL, AV_LOG_ERROR,
 | |
|            "Error applying option '%s' to filter '%s': %s\n",
 | |
|            key, f->filter->name, av_err2str(ret));
 | |
|     return ret;
 | |
| err_load:
 | |
|     av_log(NULL, AV_LOG_ERROR,
 | |
|            "Error loading value for option '%s' from file '%s'\n",
 | |
|            key, val);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int graph_opts_apply(AVFilterGraphSegment *seg)
 | |
| {
 | |
|     for (size_t i = 0; i < seg->nb_chains; i++) {
 | |
|         AVFilterChain *ch = seg->chains[i];
 | |
| 
 | |
|         for (size_t j = 0; j < ch->nb_filters; j++) {
 | |
|             AVFilterParams *p = ch->filters[j];
 | |
|             const AVDictionaryEntry *e = NULL;
 | |
| 
 | |
|             av_assert0(p->filter);
 | |
| 
 | |
|             while ((e = av_dict_iterate(p->opts, e))) {
 | |
|                 int ret = filter_opt_apply(p->filter, e->key, e->value);
 | |
|                 if (ret < 0)
 | |
|                     return ret;
 | |
|             }
 | |
| 
 | |
|             av_dict_free(&p->opts);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int graph_parse(AVFilterGraph *graph, const char *desc,
 | |
|                        AVFilterInOut **inputs, AVFilterInOut **outputs,
 | |
|                        AVBufferRef *hw_device)
 | |
| {
 | |
|     AVFilterGraphSegment *seg;
 | |
|     int ret;
 | |
| 
 | |
|     *inputs  = NULL;
 | |
|     *outputs = NULL;
 | |
| 
 | |
|     ret = avfilter_graph_segment_parse(graph, desc, 0, &seg);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     ret = avfilter_graph_segment_create_filters(seg, 0);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     if (hw_device) {
 | |
|         for (int i = 0; i < graph->nb_filters; i++) {
 | |
|             AVFilterContext *f = graph->filters[i];
 | |
| 
 | |
|             if (!(f->filter->flags & AVFILTER_FLAG_HWDEVICE))
 | |
|                 continue;
 | |
|             f->hw_device_ctx = av_buffer_ref(hw_device);
 | |
|             if (!f->hw_device_ctx) {
 | |
|                 ret = AVERROR(ENOMEM);
 | |
|                 goto fail;
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     ret = graph_opts_apply(seg);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     ret = avfilter_graph_segment_apply(seg, 0, inputs, outputs);
 | |
| 
 | |
| fail:
 | |
|     avfilter_graph_segment_free(&seg);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| // Filters can be configured only if the formats of all inputs are known.
 | |
| static int ifilter_has_all_input_formats(FilterGraph *fg)
 | |
| {
 | |
|     int i;
 | |
|     for (i = 0; i < fg->nb_inputs; i++) {
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(fg->inputs[i]);
 | |
|         if (ifp->format < 0)
 | |
|             return 0;
 | |
|     }
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static char *describe_filter_link(FilterGraph *fg, AVFilterInOut *inout, int in)
 | |
| {
 | |
|     AVFilterContext *ctx = inout->filter_ctx;
 | |
|     AVFilterPad *pads = in ? ctx->input_pads  : ctx->output_pads;
 | |
|     int       nb_pads = in ? ctx->nb_inputs   : ctx->nb_outputs;
 | |
|     char *res;
 | |
| 
 | |
|     if (nb_pads > 1)
 | |
|         res = av_strdup(ctx->filter->name);
 | |
|     else
 | |
|         res = av_asprintf("%s:%s", ctx->filter->name,
 | |
|                           avfilter_pad_get_name(pads, inout->pad_idx));
 | |
|     if (!res)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
|     return res;
 | |
| }
 | |
| 
 | |
| static OutputFilter *ofilter_alloc(FilterGraph *fg)
 | |
| {
 | |
|     OutputFilter *ofilter;
 | |
| 
 | |
|     ofilter           = ALLOC_ARRAY_ELEM(fg->outputs, fg->nb_outputs);
 | |
|     ofilter->graph    = fg;
 | |
|     ofilter->format   = -1;
 | |
|     ofilter->last_pts = AV_NOPTS_VALUE;
 | |
| 
 | |
|     return ofilter;
 | |
| }
 | |
| 
 | |
| static int ifilter_bind_ist(InputFilter *ifilter, InputStream *ist)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     int ret;
 | |
| 
 | |
|     ifp->ist             = ist;
 | |
|     ifp->type_src        = ist->st->codecpar->codec_type;
 | |
| 
 | |
|     ret = ist_filter_add(ist, ifilter, filtergraph_is_simple(ifilter->graph));
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     if (ifp->type_src == AVMEDIA_TYPE_SUBTITLE) {
 | |
|         ifp->sub2video.frame = av_frame_alloc();
 | |
|         if (!ifp->sub2video.frame)
 | |
|             return AVERROR(ENOMEM);
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void set_channel_layout(OutputFilter *f, OutputStream *ost)
 | |
| {
 | |
|     const AVCodec *c = ost->enc_ctx->codec;
 | |
|     int i, err;
 | |
| 
 | |
|     if (ost->enc_ctx->ch_layout.order != AV_CHANNEL_ORDER_UNSPEC) {
 | |
|         /* Pass the layout through for all orders but UNSPEC */
 | |
|         err = av_channel_layout_copy(&f->ch_layout, &ost->enc_ctx->ch_layout);
 | |
|         if (err < 0)
 | |
|             report_and_exit(AVERROR(ENOMEM));
 | |
|         return;
 | |
|     }
 | |
| 
 | |
|     /* Requested layout is of order UNSPEC */
 | |
|     if (!c->ch_layouts) {
 | |
|         /* Use the default native layout for the requested amount of channels when the
 | |
|            encoder doesn't have a list of supported layouts */
 | |
|         av_channel_layout_default(&f->ch_layout, ost->enc_ctx->ch_layout.nb_channels);
 | |
|         return;
 | |
|     }
 | |
|     /* Encoder has a list of supported layouts. Pick the first layout in it with the
 | |
|        same amount of channels as the requested layout */
 | |
|     for (i = 0; c->ch_layouts[i].nb_channels; i++) {
 | |
|         if (c->ch_layouts[i].nb_channels == ost->enc_ctx->ch_layout.nb_channels)
 | |
|             break;
 | |
|     }
 | |
|     if (c->ch_layouts[i].nb_channels) {
 | |
|         /* Use it if one is found */
 | |
|         err = av_channel_layout_copy(&f->ch_layout, &c->ch_layouts[i]);
 | |
|         if (err < 0)
 | |
|             report_and_exit(AVERROR(ENOMEM));
 | |
|         return;
 | |
|     }
 | |
|     /* If no layout for the amount of channels requested was found, use the default
 | |
|        native layout for it. */
 | |
|     av_channel_layout_default(&f->ch_layout, ost->enc_ctx->ch_layout.nb_channels);
 | |
| }
 | |
| 
 | |
| void ofilter_bind_ost(OutputFilter *ofilter, OutputStream *ost)
 | |
| {
 | |
|     FilterGraph  *fg = ofilter->graph;
 | |
|     const AVCodec *c = ost->enc_ctx->codec;
 | |
| 
 | |
|     ofilter->ost = ost;
 | |
|     av_freep(&ofilter->linklabel);
 | |
| 
 | |
|     switch (ost->enc_ctx->codec_type) {
 | |
|     case AVMEDIA_TYPE_VIDEO:
 | |
|         ofilter->width      = ost->enc_ctx->width;
 | |
|         ofilter->height     = ost->enc_ctx->height;
 | |
|         if (ost->enc_ctx->pix_fmt != AV_PIX_FMT_NONE) {
 | |
|             ofilter->format = ost->enc_ctx->pix_fmt;
 | |
|         } else {
 | |
|             ofilter->formats = c->pix_fmts;
 | |
|         }
 | |
|         break;
 | |
|     case AVMEDIA_TYPE_AUDIO:
 | |
|         if (ost->enc_ctx->sample_fmt != AV_SAMPLE_FMT_NONE) {
 | |
|             ofilter->format = ost->enc_ctx->sample_fmt;
 | |
|         } else {
 | |
|             ofilter->formats = c->sample_fmts;
 | |
|         }
 | |
|         if (ost->enc_ctx->sample_rate) {
 | |
|             ofilter->sample_rate = ost->enc_ctx->sample_rate;
 | |
|         } else {
 | |
|             ofilter->sample_rates = c->supported_samplerates;
 | |
|         }
 | |
|         if (ost->enc_ctx->ch_layout.nb_channels) {
 | |
|             set_channel_layout(ofilter, ost);
 | |
|         } else if (c->ch_layouts) {
 | |
|             ofilter->ch_layouts = c->ch_layouts;
 | |
|         }
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     // if we have all input parameters and all outputs are bound,
 | |
|     // the graph can now be configured
 | |
|     if (ifilter_has_all_input_formats(fg)) {
 | |
|         int ret;
 | |
| 
 | |
|         for (int i = 0; i < fg->nb_outputs; i++)
 | |
|             if (!fg->outputs[i]->ost)
 | |
|                 return;
 | |
| 
 | |
|         ret = configure_filtergraph(fg);
 | |
|         if (ret < 0) {
 | |
|             av_log(fg, AV_LOG_ERROR, "Error configuring filter graph: %s\n",
 | |
|                    av_err2str(ret));
 | |
|             exit_program(1);
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| static InputFilter *ifilter_alloc(FilterGraph *fg)
 | |
| {
 | |
|     InputFilterPriv *ifp = allocate_array_elem(&fg->inputs, sizeof(*ifp),
 | |
|                                                &fg->nb_inputs);
 | |
|     InputFilter *ifilter = &ifp->ifilter;
 | |
| 
 | |
|     ifilter->graph  = fg;
 | |
| 
 | |
|     ifp->frame = av_frame_alloc();
 | |
|     if (!ifp->frame)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
| 
 | |
|     ifp->format          = -1;
 | |
|     ifp->fallback.format = -1;
 | |
| 
 | |
|     ifp->frame_queue = av_fifo_alloc2(8, sizeof(AVFrame*), AV_FIFO_FLAG_AUTO_GROW);
 | |
|     if (!ifp->frame_queue)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
| 
 | |
|     return ifilter;
 | |
| }
 | |
| 
 | |
| void fg_free(FilterGraph **pfg)
 | |
| {
 | |
|     FilterGraph *fg = *pfg;
 | |
|     FilterGraphPriv *fgp;
 | |
| 
 | |
|     if (!fg)
 | |
|         return;
 | |
|     fgp = fgp_from_fg(fg);
 | |
| 
 | |
|     avfilter_graph_free(&fg->graph);
 | |
|     for (int j = 0; j < fg->nb_inputs; j++) {
 | |
|         InputFilter *ifilter = fg->inputs[j];
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|         if (ifp->frame_queue) {
 | |
|             AVFrame *frame;
 | |
|             while (av_fifo_read(ifp->frame_queue, &frame, 1) >= 0)
 | |
|                 av_frame_free(&frame);
 | |
|             av_fifo_freep2(&ifp->frame_queue);
 | |
|         }
 | |
|         if (ifp->sub2video.queue) {
 | |
|             AVSubtitle sub;
 | |
|             while (av_fifo_read(ifp->sub2video.queue, &sub, 1) >= 0)
 | |
|                 avsubtitle_free(&sub);
 | |
|             av_fifo_freep2(&ifp->sub2video.queue);
 | |
|         }
 | |
|         av_frame_free(&ifp->sub2video.frame);
 | |
| 
 | |
|         av_channel_layout_uninit(&ifp->fallback.ch_layout);
 | |
| 
 | |
|         av_frame_free(&ifp->frame);
 | |
| 
 | |
|         av_buffer_unref(&ifp->hw_frames_ctx);
 | |
|         av_freep(&ifp->linklabel);
 | |
|         av_freep(&ifilter->name);
 | |
|         av_freep(&fg->inputs[j]);
 | |
|     }
 | |
|     av_freep(&fg->inputs);
 | |
|     for (int j = 0; j < fg->nb_outputs; j++) {
 | |
|         OutputFilter *ofilter = fg->outputs[j];
 | |
| 
 | |
|         av_freep(&ofilter->linklabel);
 | |
|         av_freep(&ofilter->name);
 | |
|         av_channel_layout_uninit(&ofilter->ch_layout);
 | |
|         av_freep(&fg->outputs[j]);
 | |
|     }
 | |
|     av_freep(&fg->outputs);
 | |
|     av_freep(&fgp->graph_desc);
 | |
| 
 | |
|     av_frame_free(&fgp->frame);
 | |
| 
 | |
|     av_freep(pfg);
 | |
| }
 | |
| 
 | |
| static const char *fg_item_name(void *obj)
 | |
| {
 | |
|     const FilterGraphPriv *fgp = obj;
 | |
| 
 | |
|     return fgp->log_name;
 | |
| }
 | |
| 
 | |
| static const AVClass fg_class = {
 | |
|     .class_name = "FilterGraph",
 | |
|     .version    = LIBAVUTIL_VERSION_INT,
 | |
|     .item_name  = fg_item_name,
 | |
|     .category   = AV_CLASS_CATEGORY_FILTER,
 | |
| };
 | |
| 
 | |
| FilterGraph *fg_create(char *graph_desc)
 | |
| {
 | |
|     FilterGraphPriv *fgp = allocate_array_elem(&filtergraphs, sizeof(*fgp), &nb_filtergraphs);
 | |
|     FilterGraph      *fg = &fgp->fg;
 | |
| 
 | |
|     AVFilterInOut *inputs, *outputs;
 | |
|     AVFilterGraph *graph;
 | |
|     int ret = 0;
 | |
| 
 | |
|     fg->class       = &fg_class;
 | |
|     fg->index      = nb_filtergraphs - 1;
 | |
|     fgp->graph_desc = graph_desc;
 | |
| 
 | |
|     snprintf(fgp->log_name, sizeof(fgp->log_name), "fc#%d", fg->index);
 | |
| 
 | |
|     fgp->frame = av_frame_alloc();
 | |
|     if (!fgp->frame)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
| 
 | |
|     /* this graph is only used for determining the kinds of inputs
 | |
|      * and outputs we have, and is discarded on exit from this function */
 | |
|     graph = avfilter_graph_alloc();
 | |
|     if (!graph)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
|     graph->nb_threads = 1;
 | |
| 
 | |
|     ret = graph_parse(graph, fgp->graph_desc, &inputs, &outputs, NULL);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     for (AVFilterInOut *cur = inputs; cur; cur = cur->next) {
 | |
|         InputFilter *const ifilter = ifilter_alloc(fg);
 | |
|         InputFilterPriv       *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|         ifp->linklabel = cur->name;
 | |
|         cur->name      = NULL;
 | |
| 
 | |
|         ifp->type      = avfilter_pad_get_type(cur->filter_ctx->input_pads,
 | |
|                                                cur->pad_idx);
 | |
|         ifilter->name  = describe_filter_link(fg, cur, 1);
 | |
|     }
 | |
| 
 | |
|     for (AVFilterInOut *cur = outputs; cur; cur = cur->next) {
 | |
|         OutputFilter *const ofilter = ofilter_alloc(fg);
 | |
| 
 | |
|         ofilter->linklabel = cur->name;
 | |
|         cur->name          = NULL;
 | |
| 
 | |
|         ofilter->type      = avfilter_pad_get_type(cur->filter_ctx->output_pads,
 | |
|                                                    cur->pad_idx);
 | |
|         ofilter->name      = describe_filter_link(fg, cur, 0);
 | |
|     }
 | |
| 
 | |
|     if (!fg->nb_outputs) {
 | |
|         av_log(fg, AV_LOG_FATAL, "A filtergraph has zero outputs, this is not supported\n");
 | |
|         ret = AVERROR(ENOSYS);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
| fail:
 | |
|     avfilter_inout_free(&inputs);
 | |
|     avfilter_inout_free(&outputs);
 | |
|     avfilter_graph_free(&graph);
 | |
| 
 | |
|     if (ret < 0)
 | |
|         report_and_exit(ret);
 | |
| 
 | |
|     return fg;
 | |
| }
 | |
| 
 | |
| int init_simple_filtergraph(InputStream *ist, OutputStream *ost,
 | |
|                             char *graph_desc)
 | |
| {
 | |
|     FilterGraph *fg;
 | |
|     FilterGraphPriv *fgp;
 | |
|     int ret;
 | |
| 
 | |
|     fg = fg_create(graph_desc);
 | |
|     if (!fg)
 | |
|         report_and_exit(AVERROR(ENOMEM));
 | |
|     fgp = fgp_from_fg(fg);
 | |
| 
 | |
|     fgp->is_simple = 1;
 | |
| 
 | |
|     snprintf(fgp->log_name, sizeof(fgp->log_name), "%cf#%d:%d",
 | |
|              av_get_media_type_string(ost->type)[0],
 | |
|              ost->file_index, ost->index);
 | |
| 
 | |
|     if (fg->nb_inputs != 1 || fg->nb_outputs != 1) {
 | |
|         av_log(fg, AV_LOG_ERROR, "Simple filtergraph '%s' was expected "
 | |
|                "to have exactly 1 input and 1 output. "
 | |
|                "However, it had %d input(s) and %d output(s). Please adjust, "
 | |
|                "or use a complex filtergraph (-filter_complex) instead.\n",
 | |
|                graph_desc, fg->nb_inputs, fg->nb_outputs);
 | |
|         return AVERROR(EINVAL);
 | |
|     }
 | |
| 
 | |
|     ost->filter = fg->outputs[0];
 | |
| 
 | |
|     ret = ifilter_bind_ist(fg->inputs[0], ist);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     ofilter_bind_ost(fg->outputs[0], ost);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void init_input_filter(FilterGraph *fg, InputFilter *ifilter)
 | |
| {
 | |
|     FilterGraphPriv *fgp = fgp_from_fg(fg);
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     InputStream *ist = NULL;
 | |
|     enum AVMediaType type = ifp->type;
 | |
|     int i, ret;
 | |
| 
 | |
|     // TODO: support other filter types
 | |
|     if (type != AVMEDIA_TYPE_VIDEO && type != AVMEDIA_TYPE_AUDIO) {
 | |
|         av_log(fg, AV_LOG_FATAL, "Only video and audio filters supported "
 | |
|                "currently.\n");
 | |
|         exit_program(1);
 | |
|     }
 | |
| 
 | |
|     if (ifp->linklabel) {
 | |
|         AVFormatContext *s;
 | |
|         AVStream       *st = NULL;
 | |
|         char *p;
 | |
|         int file_idx = strtol(ifp->linklabel, &p, 0);
 | |
| 
 | |
|         if (file_idx < 0 || file_idx >= nb_input_files) {
 | |
|             av_log(fg, AV_LOG_FATAL, "Invalid file index %d in filtergraph description %s.\n",
 | |
|                    file_idx, fgp->graph_desc);
 | |
|             exit_program(1);
 | |
|         }
 | |
|         s = input_files[file_idx]->ctx;
 | |
| 
 | |
|         for (i = 0; i < s->nb_streams; i++) {
 | |
|             enum AVMediaType stream_type = s->streams[i]->codecpar->codec_type;
 | |
|             if (stream_type != type &&
 | |
|                 !(stream_type == AVMEDIA_TYPE_SUBTITLE &&
 | |
|                   type == AVMEDIA_TYPE_VIDEO /* sub2video hack */))
 | |
|                 continue;
 | |
|             if (check_stream_specifier(s, s->streams[i], *p == ':' ? p + 1 : p) == 1) {
 | |
|                 st = s->streams[i];
 | |
|                 break;
 | |
|             }
 | |
|         }
 | |
|         if (!st) {
 | |
|             av_log(fg, AV_LOG_FATAL, "Stream specifier '%s' in filtergraph description %s "
 | |
|                    "matches no streams.\n", p, fgp->graph_desc);
 | |
|             exit_program(1);
 | |
|         }
 | |
|         ist = input_files[file_idx]->streams[st->index];
 | |
|     } else {
 | |
|         ist = ist_find_unused(type);
 | |
|         if (!ist) {
 | |
|             av_log(fg, AV_LOG_FATAL, "Cannot find a matching stream for "
 | |
|                    "unlabeled input pad %s\n", ifilter->name);
 | |
|             exit_program(1);
 | |
|         }
 | |
|     }
 | |
|     av_assert0(ist);
 | |
| 
 | |
|     ret = ifilter_bind_ist(ifilter, ist);
 | |
|     if (ret < 0) {
 | |
|         av_log(fg, AV_LOG_ERROR,
 | |
|                "Error binding an input stream to complex filtergraph input %s.\n",
 | |
|                ifilter->name);
 | |
|         exit_program(1);
 | |
|     }
 | |
| }
 | |
| 
 | |
| int init_complex_filtergraph(FilterGraph *fg)
 | |
| {
 | |
|     // bind filtergraph inputs to input streams
 | |
|     for (int i = 0; i < fg->nb_inputs; i++)
 | |
|         init_input_filter(fg, fg->inputs[i]);
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int insert_trim(int64_t start_time, int64_t duration,
 | |
|                        AVFilterContext **last_filter, int *pad_idx,
 | |
|                        const char *filter_name)
 | |
| {
 | |
|     AVFilterGraph *graph = (*last_filter)->graph;
 | |
|     AVFilterContext *ctx;
 | |
|     const AVFilter *trim;
 | |
|     enum AVMediaType type = avfilter_pad_get_type((*last_filter)->output_pads, *pad_idx);
 | |
|     const char *name = (type == AVMEDIA_TYPE_VIDEO) ? "trim" : "atrim";
 | |
|     int ret = 0;
 | |
| 
 | |
|     if (duration == INT64_MAX && start_time == AV_NOPTS_VALUE)
 | |
|         return 0;
 | |
| 
 | |
|     trim = avfilter_get_by_name(name);
 | |
|     if (!trim) {
 | |
|         av_log(NULL, AV_LOG_ERROR, "%s filter not present, cannot limit "
 | |
|                "recording time.\n", name);
 | |
|         return AVERROR_FILTER_NOT_FOUND;
 | |
|     }
 | |
| 
 | |
|     ctx = avfilter_graph_alloc_filter(graph, trim, filter_name);
 | |
|     if (!ctx)
 | |
|         return AVERROR(ENOMEM);
 | |
| 
 | |
|     if (duration != INT64_MAX) {
 | |
|         ret = av_opt_set_int(ctx, "durationi", duration,
 | |
|                                 AV_OPT_SEARCH_CHILDREN);
 | |
|     }
 | |
|     if (ret >= 0 && start_time != AV_NOPTS_VALUE) {
 | |
|         ret = av_opt_set_int(ctx, "starti", start_time,
 | |
|                                 AV_OPT_SEARCH_CHILDREN);
 | |
|     }
 | |
|     if (ret < 0) {
 | |
|         av_log(ctx, AV_LOG_ERROR, "Error configuring the %s filter", name);
 | |
|         return ret;
 | |
|     }
 | |
| 
 | |
|     ret = avfilter_init_str(ctx, NULL);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     ret = avfilter_link(*last_filter, *pad_idx, ctx, 0);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     *last_filter = ctx;
 | |
|     *pad_idx     = 0;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int insert_filter(AVFilterContext **last_filter, int *pad_idx,
 | |
|                          const char *filter_name, const char *args)
 | |
| {
 | |
|     AVFilterGraph *graph = (*last_filter)->graph;
 | |
|     AVFilterContext *ctx;
 | |
|     int ret;
 | |
| 
 | |
|     ret = avfilter_graph_create_filter(&ctx,
 | |
|                                        avfilter_get_by_name(filter_name),
 | |
|                                        filter_name, args, NULL, graph);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     ret = avfilter_link(*last_filter, *pad_idx, ctx, 0);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     *last_filter = ctx;
 | |
|     *pad_idx     = 0;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int configure_output_video_filter(FilterGraph *fg, OutputFilter *ofilter, AVFilterInOut *out)
 | |
| {
 | |
|     OutputStream *ost = ofilter->ost;
 | |
|     OutputFile    *of = output_files[ost->file_index];
 | |
|     AVFilterContext *last_filter = out->filter_ctx;
 | |
|     AVBPrint bprint;
 | |
|     int pad_idx = out->pad_idx;
 | |
|     int ret;
 | |
|     const char *pix_fmts;
 | |
|     char name[255];
 | |
| 
 | |
|     snprintf(name, sizeof(name), "out_%d_%d", ost->file_index, ost->index);
 | |
|     ret = avfilter_graph_create_filter(&ofilter->filter,
 | |
|                                        avfilter_get_by_name("buffersink"),
 | |
|                                        name, NULL, NULL, fg->graph);
 | |
| 
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     if ((ofilter->width || ofilter->height) && ofilter->ost->autoscale) {
 | |
|         char args[255];
 | |
|         AVFilterContext *filter;
 | |
|         const AVDictionaryEntry *e = NULL;
 | |
| 
 | |
|         snprintf(args, sizeof(args), "%d:%d",
 | |
|                  ofilter->width, ofilter->height);
 | |
| 
 | |
|         while ((e = av_dict_iterate(ost->sws_dict, e))) {
 | |
|             av_strlcatf(args, sizeof(args), ":%s=%s", e->key, e->value);
 | |
|         }
 | |
| 
 | |
|         snprintf(name, sizeof(name), "scaler_out_%d_%d",
 | |
|                  ost->file_index, ost->index);
 | |
|         if ((ret = avfilter_graph_create_filter(&filter, avfilter_get_by_name("scale"),
 | |
|                                                 name, args, NULL, fg->graph)) < 0)
 | |
|             return ret;
 | |
|         if ((ret = avfilter_link(last_filter, pad_idx, filter, 0)) < 0)
 | |
|             return ret;
 | |
| 
 | |
|         last_filter = filter;
 | |
|         pad_idx = 0;
 | |
|     }
 | |
| 
 | |
|     av_bprint_init(&bprint, 0, AV_BPRINT_SIZE_UNLIMITED);
 | |
|     if ((pix_fmts = choose_pix_fmts(ofilter, &bprint))) {
 | |
|         AVFilterContext *filter;
 | |
| 
 | |
|         ret = avfilter_graph_create_filter(&filter,
 | |
|                                            avfilter_get_by_name("format"),
 | |
|                                            "format", pix_fmts, NULL, fg->graph);
 | |
|         av_bprint_finalize(&bprint, NULL);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|         if ((ret = avfilter_link(last_filter, pad_idx, filter, 0)) < 0)
 | |
|             return ret;
 | |
| 
 | |
|         last_filter = filter;
 | |
|         pad_idx     = 0;
 | |
|     }
 | |
| 
 | |
|     snprintf(name, sizeof(name), "trim_out_%d_%d",
 | |
|              ost->file_index, ost->index);
 | |
|     ret = insert_trim(of->start_time, of->recording_time,
 | |
|                       &last_filter, &pad_idx, name);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
| 
 | |
|     if ((ret = avfilter_link(last_filter, pad_idx, ofilter->filter, 0)) < 0)
 | |
|         return ret;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int configure_output_audio_filter(FilterGraph *fg, OutputFilter *ofilter, AVFilterInOut *out)
 | |
| {
 | |
|     OutputStream *ost = ofilter->ost;
 | |
|     OutputFile    *of = output_files[ost->file_index];
 | |
|     AVFilterContext *last_filter = out->filter_ctx;
 | |
|     int pad_idx = out->pad_idx;
 | |
|     AVBPrint args;
 | |
|     char name[255];
 | |
|     int ret;
 | |
| 
 | |
|     snprintf(name, sizeof(name), "out_%d_%d", ost->file_index, ost->index);
 | |
|     ret = avfilter_graph_create_filter(&ofilter->filter,
 | |
|                                        avfilter_get_by_name("abuffersink"),
 | |
|                                        name, NULL, NULL, fg->graph);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
|     if ((ret = av_opt_set_int(ofilter->filter, "all_channel_counts", 1, AV_OPT_SEARCH_CHILDREN)) < 0)
 | |
|         return ret;
 | |
| 
 | |
| #define AUTO_INSERT_FILTER(opt_name, filter_name, arg) do {                 \
 | |
|     AVFilterContext *filt_ctx;                                              \
 | |
|                                                                             \
 | |
|     av_log(fg, AV_LOG_INFO, opt_name " is forwarded to lavfi "              \
 | |
|            "similarly to -af " filter_name "=%s.\n", arg);                  \
 | |
|                                                                             \
 | |
|     ret = avfilter_graph_create_filter(&filt_ctx,                           \
 | |
|                                        avfilter_get_by_name(filter_name),   \
 | |
|                                        filter_name, arg, NULL, fg->graph);  \
 | |
|     if (ret < 0)                                                            \
 | |
|         goto fail;                                                          \
 | |
|                                                                             \
 | |
|     ret = avfilter_link(last_filter, pad_idx, filt_ctx, 0);                 \
 | |
|     if (ret < 0)                                                            \
 | |
|         goto fail;                                                          \
 | |
|                                                                             \
 | |
|     last_filter = filt_ctx;                                                 \
 | |
|     pad_idx = 0;                                                            \
 | |
| } while (0)
 | |
|     av_bprint_init(&args, 0, AV_BPRINT_SIZE_UNLIMITED);
 | |
| #if FFMPEG_OPT_MAP_CHANNEL
 | |
|     if (ost->audio_channels_mapped) {
 | |
|         AVChannelLayout mapped_layout = { 0 };
 | |
|         int i;
 | |
|         av_channel_layout_default(&mapped_layout, ost->audio_channels_mapped);
 | |
|         av_channel_layout_describe_bprint(&mapped_layout, &args);
 | |
|         for (i = 0; i < ost->audio_channels_mapped; i++)
 | |
|             if (ost->audio_channels_map[i] != -1)
 | |
|                 av_bprintf(&args, "|c%d=c%d", i, ost->audio_channels_map[i]);
 | |
| 
 | |
|         AUTO_INSERT_FILTER("-map_channel", "pan", args.str);
 | |
|         av_bprint_clear(&args);
 | |
|     }
 | |
| #endif
 | |
| 
 | |
|     choose_sample_fmts(ofilter,     &args);
 | |
|     choose_sample_rates(ofilter,    &args);
 | |
|     choose_channel_layouts(ofilter, &args);
 | |
|     if (!av_bprint_is_complete(&args)) {
 | |
|         ret = AVERROR(ENOMEM);
 | |
|         goto fail;
 | |
|     }
 | |
|     if (args.len) {
 | |
|         AVFilterContext *format;
 | |
| 
 | |
|         snprintf(name, sizeof(name), "format_out_%d_%d",
 | |
|                  ost->file_index, ost->index);
 | |
|         ret = avfilter_graph_create_filter(&format,
 | |
|                                            avfilter_get_by_name("aformat"),
 | |
|                                            name, args.str, NULL, fg->graph);
 | |
|         if (ret < 0)
 | |
|             goto fail;
 | |
| 
 | |
|         ret = avfilter_link(last_filter, pad_idx, format, 0);
 | |
|         if (ret < 0)
 | |
|             goto fail;
 | |
| 
 | |
|         last_filter = format;
 | |
|         pad_idx = 0;
 | |
|     }
 | |
| 
 | |
|     if (ost->apad && of->shortest) {
 | |
|         int i;
 | |
| 
 | |
|         for (i = 0; i < of->nb_streams; i++)
 | |
|             if (of->streams[i]->st->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
 | |
|                 break;
 | |
| 
 | |
|         if (i < of->nb_streams) {
 | |
|             AUTO_INSERT_FILTER("-apad", "apad", ost->apad);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     snprintf(name, sizeof(name), "trim for output stream %d:%d",
 | |
|              ost->file_index, ost->index);
 | |
|     ret = insert_trim(of->start_time, of->recording_time,
 | |
|                       &last_filter, &pad_idx, name);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     if ((ret = avfilter_link(last_filter, pad_idx, ofilter->filter, 0)) < 0)
 | |
|         goto fail;
 | |
| fail:
 | |
|     av_bprint_finalize(&args, NULL);
 | |
| 
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int configure_output_filter(FilterGraph *fg, OutputFilter *ofilter,
 | |
|                                    AVFilterInOut *out)
 | |
| {
 | |
|     if (!ofilter->ost) {
 | |
|         av_log(fg, AV_LOG_FATAL, "Filter %s has an unconnected output\n", ofilter->name);
 | |
|         exit_program(1);
 | |
|     }
 | |
| 
 | |
|     switch (avfilter_pad_get_type(out->filter_ctx->output_pads, out->pad_idx)) {
 | |
|     case AVMEDIA_TYPE_VIDEO: return configure_output_video_filter(fg, ofilter, out);
 | |
|     case AVMEDIA_TYPE_AUDIO: return configure_output_audio_filter(fg, ofilter, out);
 | |
|     default: av_assert0(0); return 0;
 | |
|     }
 | |
| }
 | |
| 
 | |
| void check_filter_outputs(void)
 | |
| {
 | |
|     int i;
 | |
|     for (i = 0; i < nb_filtergraphs; i++) {
 | |
|         int n;
 | |
|         for (n = 0; n < filtergraphs[i]->nb_outputs; n++) {
 | |
|             OutputFilter *output = filtergraphs[i]->outputs[n];
 | |
|             if (!output->ost) {
 | |
|                 av_log(filtergraphs[i], AV_LOG_FATAL,
 | |
|                        "Filter %s has an unconnected output\n", output->name);
 | |
|                 exit_program(1);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void sub2video_prepare(InputFilterPriv *ifp)
 | |
| {
 | |
|     ifp->sub2video.last_pts = INT64_MIN;
 | |
|     ifp->sub2video.end_pts  = INT64_MIN;
 | |
| 
 | |
|     /* sub2video structure has been (re-)initialized.
 | |
|        Mark it as such so that the system will be
 | |
|        initialized with the first received heartbeat. */
 | |
|     ifp->sub2video.initialize = 1;
 | |
| }
 | |
| 
 | |
| static int configure_input_video_filter(FilterGraph *fg, InputFilter *ifilter,
 | |
|                                         AVFilterInOut *in)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|     AVFilterContext *last_filter;
 | |
|     const AVFilter *buffer_filt = avfilter_get_by_name("buffer");
 | |
|     const AVPixFmtDescriptor *desc;
 | |
|     InputStream *ist = ifp->ist;
 | |
|     InputFile     *f = input_files[ist->file_index];
 | |
|     AVRational fr = ist->framerate;
 | |
|     AVRational sar;
 | |
|     AVBPrint args;
 | |
|     char name[255];
 | |
|     int ret, pad_idx = 0;
 | |
|     int64_t tsoffset = 0;
 | |
|     AVBufferSrcParameters *par = av_buffersrc_parameters_alloc();
 | |
| 
 | |
|     if (!par)
 | |
|         return AVERROR(ENOMEM);
 | |
|     memset(par, 0, sizeof(*par));
 | |
|     par->format = AV_PIX_FMT_NONE;
 | |
| 
 | |
|     if (ist->dec_ctx->codec_type == AVMEDIA_TYPE_AUDIO) {
 | |
|         av_log(fg, AV_LOG_ERROR, "Cannot connect video filter to audio input\n");
 | |
|         ret = AVERROR(EINVAL);
 | |
|         goto fail;
 | |
|     }
 | |
| 
 | |
|     if (!fr.num)
 | |
|         fr = ist->framerate_guessed;
 | |
| 
 | |
|     if (ifp->type_src == AVMEDIA_TYPE_SUBTITLE)
 | |
|         sub2video_prepare(ifp);
 | |
| 
 | |
|     ifp->time_base =  ist->framerate.num ? av_inv_q(ist->framerate) :
 | |
|                                            ist->st->time_base;
 | |
| 
 | |
|     sar = ifp->sample_aspect_ratio;
 | |
|     if(!sar.den)
 | |
|         sar = (AVRational){0,1};
 | |
|     av_bprint_init(&args, 0, AV_BPRINT_SIZE_AUTOMATIC);
 | |
|     av_bprintf(&args,
 | |
|              "video_size=%dx%d:pix_fmt=%d:time_base=%d/%d:"
 | |
|              "pixel_aspect=%d/%d",
 | |
|              ifp->width, ifp->height, ifp->format,
 | |
|              ifp->time_base.num, ifp->time_base.den, sar.num, sar.den);
 | |
|     if (fr.num && fr.den)
 | |
|         av_bprintf(&args, ":frame_rate=%d/%d", fr.num, fr.den);
 | |
|     snprintf(name, sizeof(name), "graph %d input from stream %d:%d", fg->index,
 | |
|              ist->file_index, ist->index);
 | |
| 
 | |
| 
 | |
|     if ((ret = avfilter_graph_create_filter(&ifp->filter, buffer_filt, name,
 | |
|                                             args.str, NULL, fg->graph)) < 0)
 | |
|         goto fail;
 | |
|     par->hw_frames_ctx = ifp->hw_frames_ctx;
 | |
|     ret = av_buffersrc_parameters_set(ifp->filter, par);
 | |
|     if (ret < 0)
 | |
|         goto fail;
 | |
|     av_freep(&par);
 | |
|     last_filter = ifp->filter;
 | |
| 
 | |
|     desc = av_pix_fmt_desc_get(ifp->format);
 | |
|     av_assert0(desc);
 | |
| 
 | |
|     // TODO: insert hwaccel enabled filters like transpose_vaapi into the graph
 | |
|     if (ist->autorotate && !(desc->flags & AV_PIX_FMT_FLAG_HWACCEL)) {
 | |
|         int32_t *displaymatrix = ifp->displaymatrix;
 | |
|         double theta;
 | |
| 
 | |
|         if (!ifp->displaymatrix_present)
 | |
|             displaymatrix = (int32_t *)av_stream_get_side_data(ist->st, AV_PKT_DATA_DISPLAYMATRIX, NULL);
 | |
|         theta = get_rotation(displaymatrix);
 | |
| 
 | |
|         if (fabs(theta - 90) < 1.0) {
 | |
|             ret = insert_filter(&last_filter, &pad_idx, "transpose",
 | |
|                                 displaymatrix[3] > 0 ? "cclock_flip" : "clock");
 | |
|         } else if (fabs(theta - 180) < 1.0) {
 | |
|             if (displaymatrix[0] < 0) {
 | |
|                 ret = insert_filter(&last_filter, &pad_idx, "hflip", NULL);
 | |
|                 if (ret < 0)
 | |
|                     return ret;
 | |
|             }
 | |
|             if (displaymatrix[4] < 0) {
 | |
|                 ret = insert_filter(&last_filter, &pad_idx, "vflip", NULL);
 | |
|             }
 | |
|         } else if (fabs(theta - 270) < 1.0) {
 | |
|             ret = insert_filter(&last_filter, &pad_idx, "transpose",
 | |
|                                 displaymatrix[3] < 0 ? "clock_flip" : "cclock");
 | |
|         } else if (fabs(theta) > 1.0) {
 | |
|             char rotate_buf[64];
 | |
|             snprintf(rotate_buf, sizeof(rotate_buf), "%f*PI/180", theta);
 | |
|             ret = insert_filter(&last_filter, &pad_idx, "rotate", rotate_buf);
 | |
|         } else if (fabs(theta) < 1.0) {
 | |
|             if (displaymatrix && displaymatrix[4] < 0) {
 | |
|                 ret = insert_filter(&last_filter, &pad_idx, "vflip", NULL);
 | |
|             }
 | |
|         }
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|     }
 | |
| 
 | |
|     snprintf(name, sizeof(name), "trim_in_%d_%d",
 | |
|              ist->file_index, ist->index);
 | |
|     if (copy_ts) {
 | |
|         tsoffset = f->start_time == AV_NOPTS_VALUE ? 0 : f->start_time;
 | |
|         if (!start_at_zero && f->ctx->start_time != AV_NOPTS_VALUE)
 | |
|             tsoffset += f->ctx->start_time;
 | |
|     }
 | |
|     ret = insert_trim(((f->start_time == AV_NOPTS_VALUE) || !f->accurate_seek) ?
 | |
|                       AV_NOPTS_VALUE : tsoffset, f->recording_time,
 | |
|                       &last_filter, &pad_idx, name);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     if ((ret = avfilter_link(last_filter, 0, in->filter_ctx, in->pad_idx)) < 0)
 | |
|         return ret;
 | |
|     return 0;
 | |
| fail:
 | |
|     av_freep(&par);
 | |
| 
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| static int configure_input_audio_filter(FilterGraph *fg, InputFilter *ifilter,
 | |
|                                         AVFilterInOut *in)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     AVFilterContext *last_filter;
 | |
|     const AVFilter *abuffer_filt = avfilter_get_by_name("abuffer");
 | |
|     InputStream *ist = ifp->ist;
 | |
|     InputFile     *f = input_files[ist->file_index];
 | |
|     AVBPrint args;
 | |
|     char name[255];
 | |
|     int ret, pad_idx = 0;
 | |
|     int64_t tsoffset = 0;
 | |
| 
 | |
|     if (ist->dec_ctx->codec_type != AVMEDIA_TYPE_AUDIO) {
 | |
|         av_log(fg, AV_LOG_ERROR, "Cannot connect audio filter to non audio input\n");
 | |
|         return AVERROR(EINVAL);
 | |
|     }
 | |
| 
 | |
|     ifp->time_base = (AVRational){ 1, ifp->sample_rate };
 | |
| 
 | |
|     av_bprint_init(&args, 0, AV_BPRINT_SIZE_AUTOMATIC);
 | |
|     av_bprintf(&args, "time_base=%d/%d:sample_rate=%d:sample_fmt=%s",
 | |
|                ifp->time_base.num, ifp->time_base.den,
 | |
|                ifp->sample_rate,
 | |
|                av_get_sample_fmt_name(ifp->format));
 | |
|     if (av_channel_layout_check(&ifp->ch_layout) &&
 | |
|         ifp->ch_layout.order != AV_CHANNEL_ORDER_UNSPEC) {
 | |
|         av_bprintf(&args, ":channel_layout=");
 | |
|         av_channel_layout_describe_bprint(&ifp->ch_layout, &args);
 | |
|     } else
 | |
|         av_bprintf(&args, ":channels=%d", ifp->ch_layout.nb_channels);
 | |
|     snprintf(name, sizeof(name), "graph_%d_in_%d_%d", fg->index,
 | |
|              ist->file_index, ist->index);
 | |
| 
 | |
|     if ((ret = avfilter_graph_create_filter(&ifp->filter, abuffer_filt,
 | |
|                                             name, args.str, NULL,
 | |
|                                             fg->graph)) < 0)
 | |
|         return ret;
 | |
|     last_filter = ifp->filter;
 | |
| 
 | |
|     snprintf(name, sizeof(name), "trim for input stream %d:%d",
 | |
|              ist->file_index, ist->index);
 | |
|     if (copy_ts) {
 | |
|         tsoffset = f->start_time == AV_NOPTS_VALUE ? 0 : f->start_time;
 | |
|         if (!start_at_zero && f->ctx->start_time != AV_NOPTS_VALUE)
 | |
|             tsoffset += f->ctx->start_time;
 | |
|     }
 | |
|     ret = insert_trim(((f->start_time == AV_NOPTS_VALUE) || !f->accurate_seek) ?
 | |
|                       AV_NOPTS_VALUE : tsoffset, f->recording_time,
 | |
|                       &last_filter, &pad_idx, name);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     if ((ret = avfilter_link(last_filter, 0, in->filter_ctx, in->pad_idx)) < 0)
 | |
|         return ret;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int configure_input_filter(FilterGraph *fg, InputFilter *ifilter,
 | |
|                                   AVFilterInOut *in)
 | |
| {
 | |
|     switch (ifp_from_ifilter(ifilter)->type) {
 | |
|     case AVMEDIA_TYPE_VIDEO: return configure_input_video_filter(fg, ifilter, in);
 | |
|     case AVMEDIA_TYPE_AUDIO: return configure_input_audio_filter(fg, ifilter, in);
 | |
|     default: av_assert0(0); return 0;
 | |
|     }
 | |
| }
 | |
| 
 | |
| static void cleanup_filtergraph(FilterGraph *fg)
 | |
| {
 | |
|     int i;
 | |
|     for (i = 0; i < fg->nb_outputs; i++)
 | |
|         fg->outputs[i]->filter = (AVFilterContext *)NULL;
 | |
|     for (i = 0; i < fg->nb_inputs; i++)
 | |
|         ifp_from_ifilter(fg->inputs[i])->filter = NULL;
 | |
|     avfilter_graph_free(&fg->graph);
 | |
| }
 | |
| 
 | |
| static int filter_is_buffersrc(const AVFilterContext *f)
 | |
| {
 | |
|     return f->nb_inputs == 0 &&
 | |
|            (!strcmp(f->filter->name, "buffer") ||
 | |
|             !strcmp(f->filter->name, "abuffer"));
 | |
| }
 | |
| 
 | |
| static int graph_is_meta(AVFilterGraph *graph)
 | |
| {
 | |
|     for (unsigned i = 0; i < graph->nb_filters; i++) {
 | |
|         const AVFilterContext *f = graph->filters[i];
 | |
| 
 | |
|         /* in addition to filters flagged as meta, also
 | |
|          * disregard sinks and buffersources (but not other sources,
 | |
|          * since they introduce data we are not aware of)
 | |
|          */
 | |
|         if (!((f->filter->flags & AVFILTER_FLAG_METADATA_ONLY) ||
 | |
|               f->nb_outputs == 0                               ||
 | |
|               filter_is_buffersrc(f)))
 | |
|             return 0;
 | |
|     }
 | |
|     return 1;
 | |
| }
 | |
| 
 | |
| static int configure_filtergraph(FilterGraph *fg)
 | |
| {
 | |
|     FilterGraphPriv *fgp = fgp_from_fg(fg);
 | |
|     AVBufferRef *hw_device;
 | |
|     AVFilterInOut *inputs, *outputs, *cur;
 | |
|     int ret, i, simple = filtergraph_is_simple(fg);
 | |
|     const char *graph_desc = fgp->graph_desc;
 | |
| 
 | |
|     cleanup_filtergraph(fg);
 | |
|     if (!(fg->graph = avfilter_graph_alloc()))
 | |
|         return AVERROR(ENOMEM);
 | |
| 
 | |
|     if (simple) {
 | |
|         OutputStream *ost = fg->outputs[0]->ost;
 | |
| 
 | |
|         if (filter_nbthreads) {
 | |
|             ret = av_opt_set(fg->graph, "threads", filter_nbthreads, 0);
 | |
|             if (ret < 0)
 | |
|                 goto fail;
 | |
|         } else {
 | |
|             const AVDictionaryEntry *e = NULL;
 | |
|             e = av_dict_get(ost->encoder_opts, "threads", NULL, 0);
 | |
|             if (e)
 | |
|                 av_opt_set(fg->graph, "threads", e->value, 0);
 | |
|         }
 | |
| 
 | |
|         if (av_dict_count(ost->sws_dict)) {
 | |
|             ret = av_dict_get_string(ost->sws_dict,
 | |
|                                      &fg->graph->scale_sws_opts,
 | |
|                                      '=', ':');
 | |
|             if (ret < 0)
 | |
|                 goto fail;
 | |
|         }
 | |
| 
 | |
|         if (av_dict_count(ost->swr_opts)) {
 | |
|             char *args;
 | |
|             ret = av_dict_get_string(ost->swr_opts, &args, '=', ':');
 | |
|             if (ret < 0)
 | |
|                 goto fail;
 | |
|             av_opt_set(fg->graph, "aresample_swr_opts", args, 0);
 | |
|             av_free(args);
 | |
|         }
 | |
|     } else {
 | |
|         fg->graph->nb_threads = filter_complex_nbthreads;
 | |
|     }
 | |
| 
 | |
|     hw_device = hw_device_for_filter();
 | |
| 
 | |
|     if ((ret = graph_parse(fg->graph, graph_desc, &inputs, &outputs, hw_device)) < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     for (cur = inputs, i = 0; cur; cur = cur->next, i++)
 | |
|         if ((ret = configure_input_filter(fg, fg->inputs[i], cur)) < 0) {
 | |
|             avfilter_inout_free(&inputs);
 | |
|             avfilter_inout_free(&outputs);
 | |
|             goto fail;
 | |
|         }
 | |
|     avfilter_inout_free(&inputs);
 | |
| 
 | |
|     for (cur = outputs, i = 0; cur; cur = cur->next, i++)
 | |
|         configure_output_filter(fg, fg->outputs[i], cur);
 | |
|     avfilter_inout_free(&outputs);
 | |
| 
 | |
|     if (!auto_conversion_filters)
 | |
|         avfilter_graph_set_auto_convert(fg->graph, AVFILTER_AUTO_CONVERT_NONE);
 | |
|     if ((ret = avfilter_graph_config(fg->graph, NULL)) < 0)
 | |
|         goto fail;
 | |
| 
 | |
|     fgp->is_meta = graph_is_meta(fg->graph);
 | |
| 
 | |
|     /* limit the lists of allowed formats to the ones selected, to
 | |
|      * make sure they stay the same if the filtergraph is reconfigured later */
 | |
|     for (i = 0; i < fg->nb_outputs; i++) {
 | |
|         OutputFilter *ofilter = fg->outputs[i];
 | |
|         AVFilterContext *sink = ofilter->filter;
 | |
| 
 | |
|         ofilter->format = av_buffersink_get_format(sink);
 | |
| 
 | |
|         ofilter->width  = av_buffersink_get_w(sink);
 | |
|         ofilter->height = av_buffersink_get_h(sink);
 | |
| 
 | |
|         ofilter->sample_rate    = av_buffersink_get_sample_rate(sink);
 | |
|         av_channel_layout_uninit(&ofilter->ch_layout);
 | |
|         ret = av_buffersink_get_ch_layout(sink, &ofilter->ch_layout);
 | |
|         if (ret < 0)
 | |
|             goto fail;
 | |
|     }
 | |
| 
 | |
|     for (i = 0; i < fg->nb_inputs; i++) {
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(fg->inputs[i]);
 | |
|         AVFrame *tmp;
 | |
|         while (av_fifo_read(ifp->frame_queue, &tmp, 1) >= 0) {
 | |
|             ret = av_buffersrc_add_frame(ifp->filter, tmp);
 | |
|             av_frame_free(&tmp);
 | |
|             if (ret < 0)
 | |
|                 goto fail;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     /* send the EOFs for the finished inputs */
 | |
|     for (i = 0; i < fg->nb_inputs; i++) {
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(fg->inputs[i]);
 | |
|         if (ifp->eof) {
 | |
|             ret = av_buffersrc_add_frame(ifp->filter, NULL);
 | |
|             if (ret < 0)
 | |
|                 goto fail;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     /* process queued up subtitle packets */
 | |
|     for (i = 0; i < fg->nb_inputs; i++) {
 | |
|         InputFilter *ifilter = fg->inputs[i];
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|         if (ifp->type_src == AVMEDIA_TYPE_SUBTITLE && ifp->sub2video.queue) {
 | |
|             AVSubtitle tmp;
 | |
|             while (av_fifo_read(ifp->sub2video.queue, &tmp, 1) >= 0) {
 | |
|                 sub2video_update(ifp, INT64_MIN, &tmp);
 | |
|                 avsubtitle_free(&tmp);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| 
 | |
| fail:
 | |
|     cleanup_filtergraph(fg);
 | |
|     return ret;
 | |
| }
 | |
| 
 | |
| int ifilter_parameters_from_dec(InputFilter *ifilter, const AVCodecContext *dec)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|     if (dec->codec_type == AVMEDIA_TYPE_VIDEO) {
 | |
|         ifp->fallback.format                 = dec->pix_fmt;
 | |
|         ifp->fallback.width                  = dec->width;
 | |
|         ifp->fallback.height                 = dec->height;
 | |
|         ifp->fallback.sample_aspect_ratio    = dec->sample_aspect_ratio;
 | |
|     } else if (dec->codec_type == AVMEDIA_TYPE_AUDIO) {
 | |
|         int ret;
 | |
| 
 | |
|         ifp->fallback.format                 = dec->sample_fmt;
 | |
|         ifp->fallback.sample_rate            = dec->sample_rate;
 | |
| 
 | |
|         ret = av_channel_layout_copy(&ifp->fallback.ch_layout, &dec->ch_layout);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|     } else {
 | |
|         // for subtitles (i.e. sub2video) we set the actual parameters,
 | |
|         // rather than just fallback
 | |
|         ifp->width  = ifp->ist->sub2video.w;
 | |
|         ifp->height = ifp->ist->sub2video.h;
 | |
| 
 | |
|         /* rectangles are AV_PIX_FMT_PAL8, but we have no guarantee that the
 | |
|            palettes for all rectangles are identical or compatible */
 | |
|         ifp->format = AV_PIX_FMT_RGB32;
 | |
| 
 | |
|         av_log(NULL, AV_LOG_VERBOSE, "sub2video: using %dx%d canvas\n", ifp->width, ifp->height);
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int ifilter_parameters_from_frame(InputFilter *ifilter, const AVFrame *frame)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     AVFrameSideData *sd;
 | |
|     int ret;
 | |
| 
 | |
|     ret = av_buffer_replace(&ifp->hw_frames_ctx, frame->hw_frames_ctx);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     ifp->format              = frame->format;
 | |
| 
 | |
|     ifp->width               = frame->width;
 | |
|     ifp->height              = frame->height;
 | |
|     ifp->sample_aspect_ratio = frame->sample_aspect_ratio;
 | |
| 
 | |
|     ifp->sample_rate         = frame->sample_rate;
 | |
|     ret = av_channel_layout_copy(&ifp->ch_layout, &frame->ch_layout);
 | |
|     if (ret < 0)
 | |
|         return ret;
 | |
| 
 | |
|     sd = av_frame_get_side_data(frame, AV_FRAME_DATA_DISPLAYMATRIX);
 | |
|     if (sd)
 | |
|         memcpy(ifp->displaymatrix, sd->data, sizeof(ifp->displaymatrix));
 | |
|     ifp->displaymatrix_present = !!sd;
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int filtergraph_is_simple(const FilterGraph *fg)
 | |
| {
 | |
|     const FilterGraphPriv *fgp = cfgp_from_cfg(fg);
 | |
|     return fgp->is_simple;
 | |
| }
 | |
| 
 | |
| int reap_filters(int flush)
 | |
| {
 | |
|     /* Reap all buffers present in the buffer sinks */
 | |
|     for (OutputStream *ost = ost_iter(NULL); ost; ost = ost_iter(ost)) {
 | |
|         FilterGraphPriv *fgp;
 | |
|         AVFrame *filtered_frame;
 | |
|         AVFilterContext *filter;
 | |
|         int ret = 0;
 | |
| 
 | |
|         if (!ost->filter || !ost->filter->graph->graph)
 | |
|             continue;
 | |
|         filter = ost->filter->filter;
 | |
|         fgp    = fgp_from_fg(ost->filter->graph);
 | |
| 
 | |
|         filtered_frame = fgp->frame;
 | |
| 
 | |
|         while (1) {
 | |
|             FrameData *fd;
 | |
| 
 | |
|             ret = av_buffersink_get_frame_flags(filter, filtered_frame,
 | |
|                                                AV_BUFFERSINK_FLAG_NO_REQUEST);
 | |
|             if (ret < 0) {
 | |
|                 if (ret != AVERROR(EAGAIN) && ret != AVERROR_EOF) {
 | |
|                     av_log(fgp, AV_LOG_WARNING,
 | |
|                            "Error in av_buffersink_get_frame_flags(): %s\n", av_err2str(ret));
 | |
|                 } else if (flush && ret == AVERROR_EOF) {
 | |
|                     if (av_buffersink_get_type(filter) == AVMEDIA_TYPE_VIDEO)
 | |
|                         enc_frame(ost, NULL);
 | |
|                 }
 | |
|                 break;
 | |
|             }
 | |
|             if (ost->finished) {
 | |
|                 av_frame_unref(filtered_frame);
 | |
|                 continue;
 | |
|             }
 | |
| 
 | |
|             if (filtered_frame->pts != AV_NOPTS_VALUE) {
 | |
|                 AVRational tb = av_buffersink_get_time_base(filter);
 | |
|                 ost->filter->last_pts = av_rescale_q(filtered_frame->pts, tb,
 | |
|                                                      AV_TIME_BASE_Q);
 | |
|                 filtered_frame->time_base = tb;
 | |
| 
 | |
|                 if (debug_ts)
 | |
|                     av_log(fgp, AV_LOG_INFO, "filter_raw -> pts:%s pts_time:%s time_base:%d/%d\n",
 | |
|                            av_ts2str(filtered_frame->pts),
 | |
|                            av_ts2timestr(filtered_frame->pts, &tb),
 | |
|                            tb.num, tb.den);
 | |
|             }
 | |
| 
 | |
|             fd = frame_data(filtered_frame);
 | |
|             if (!fd) {
 | |
|                 av_frame_unref(filtered_frame);
 | |
|                 report_and_exit(AVERROR(ENOMEM));
 | |
|             }
 | |
| 
 | |
|             // only use bits_per_raw_sample passed through from the decoder
 | |
|             // if the filtergraph did not touch the frame data
 | |
|             if (!fgp->is_meta)
 | |
|                 fd->bits_per_raw_sample = 0;
 | |
| 
 | |
|             if (ost->type == AVMEDIA_TYPE_VIDEO)
 | |
|                 fd->frame_rate_filter = av_buffersink_get_frame_rate(filter);
 | |
| 
 | |
|             enc_frame(ost, filtered_frame);
 | |
|             av_frame_unref(filtered_frame);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| void ifilter_sub2video_heartbeat(InputFilter *ifilter, int64_t pts, AVRational tb)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     int64_t pts2;
 | |
| 
 | |
|     if (!ifilter->graph->graph)
 | |
|         return;
 | |
| 
 | |
|     /* subtitles seem to be usually muxed ahead of other streams;
 | |
|        if not, subtracting a larger time here is necessary */
 | |
|     pts2 = av_rescale_q(pts, tb, ifp->time_base) - 1;
 | |
| 
 | |
|     /* do not send the heartbeat frame if the subtitle is already ahead */
 | |
|     if (pts2 <= ifp->sub2video.last_pts)
 | |
|         return;
 | |
| 
 | |
|     if (pts2 >= ifp->sub2video.end_pts || ifp->sub2video.initialize)
 | |
|         /* if we have hit the end of the current displayed subpicture,
 | |
|            or if we need to initialize the system, update the
 | |
|            overlayed subpicture and its start/end times */
 | |
|         sub2video_update(ifp, pts2 + 1, NULL);
 | |
| 
 | |
|     if (av_buffersrc_get_nb_failed_requests(ifp->filter))
 | |
|         sub2video_push_ref(ifp, pts2);
 | |
| }
 | |
| 
 | |
| int ifilter_sub2video(InputFilter *ifilter, const AVSubtitle *subtitle)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     int ret;
 | |
| 
 | |
|     if (ifilter->graph->graph) {
 | |
|         if (!subtitle) {
 | |
|             if (ifp->sub2video.end_pts < INT64_MAX)
 | |
|                 sub2video_update(ifp, INT64_MAX, NULL);
 | |
| 
 | |
|             return av_buffersrc_add_frame(ifp->filter, NULL);
 | |
|         }
 | |
| 
 | |
|         sub2video_update(ifp, INT64_MIN, subtitle);
 | |
|     } else if (subtitle) {
 | |
|         AVSubtitle sub;
 | |
| 
 | |
|         if (!ifp->sub2video.queue)
 | |
|             ifp->sub2video.queue = av_fifo_alloc2(8, sizeof(AVSubtitle), AV_FIFO_FLAG_AUTO_GROW);
 | |
|         if (!ifp->sub2video.queue)
 | |
|             return AVERROR(ENOMEM);
 | |
| 
 | |
|         ret = copy_av_subtitle(&sub, subtitle);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
| 
 | |
|         ret = av_fifo_write(ifp->sub2video.queue, &sub, 1);
 | |
|         if (ret < 0) {
 | |
|             avsubtitle_free(&sub);
 | |
|             return ret;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int ifilter_send_eof(InputFilter *ifilter, int64_t pts, AVRational tb)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     int ret;
 | |
| 
 | |
|     ifp->eof = 1;
 | |
| 
 | |
|     if (ifp->filter) {
 | |
|         pts = av_rescale_q_rnd(pts, tb, ifp->time_base,
 | |
|                                AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX);
 | |
| 
 | |
|         ret = av_buffersrc_close(ifp->filter, pts, AV_BUFFERSRC_FLAG_PUSH);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|     } else {
 | |
|         if (ifp->format < 0) {
 | |
|             // the filtergraph was never configured, use the fallback parameters
 | |
|             ifp->format                 = ifp->fallback.format;
 | |
|             ifp->sample_rate            = ifp->fallback.sample_rate;
 | |
|             ifp->width                  = ifp->fallback.width;
 | |
|             ifp->height                 = ifp->fallback.height;
 | |
|             ifp->sample_aspect_ratio    = ifp->fallback.sample_aspect_ratio;
 | |
| 
 | |
|             ret = av_channel_layout_copy(&ifp->ch_layout,
 | |
|                                          &ifp->fallback.ch_layout);
 | |
|             if (ret < 0)
 | |
|                 return ret;
 | |
| 
 | |
|             if (ifilter_has_all_input_formats(ifilter->graph)) {
 | |
|                 ret = configure_filtergraph(ifilter->graph);
 | |
|                 if (ret < 0) {
 | |
|                     av_log(NULL, AV_LOG_ERROR, "Error initializing filters!\n");
 | |
|                     return ret;
 | |
|                 }
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         if (ifp->format < 0) {
 | |
|             av_log(NULL, AV_LOG_ERROR,
 | |
|                    "Cannot determine format of input stream %d:%d after EOF\n",
 | |
|                    ifp->ist->file_index, ifp->ist->index);
 | |
|             return AVERROR_INVALIDDATA;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int ifilter_send_frame(InputFilter *ifilter, AVFrame *frame, int keep_reference)
 | |
| {
 | |
|     InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|     FilterGraph *fg = ifilter->graph;
 | |
|     AVFrameSideData *sd;
 | |
|     int need_reinit, ret;
 | |
| 
 | |
|     /* determine if the parameters for this input changed */
 | |
|     need_reinit = ifp->format != frame->format;
 | |
| 
 | |
|     switch (ifp->type) {
 | |
|     case AVMEDIA_TYPE_AUDIO:
 | |
|         need_reinit |= ifp->sample_rate    != frame->sample_rate ||
 | |
|                        av_channel_layout_compare(&ifp->ch_layout, &frame->ch_layout);
 | |
|         break;
 | |
|     case AVMEDIA_TYPE_VIDEO:
 | |
|         need_reinit |= ifp->width  != frame->width ||
 | |
|                        ifp->height != frame->height;
 | |
|         break;
 | |
|     }
 | |
| 
 | |
|     if (!ifp->ist->reinit_filters && fg->graph)
 | |
|         need_reinit = 0;
 | |
| 
 | |
|     if (!!ifp->hw_frames_ctx != !!frame->hw_frames_ctx ||
 | |
|         (ifp->hw_frames_ctx && ifp->hw_frames_ctx->data != frame->hw_frames_ctx->data))
 | |
|         need_reinit = 1;
 | |
| 
 | |
|     if (sd = av_frame_get_side_data(frame, AV_FRAME_DATA_DISPLAYMATRIX)) {
 | |
|         if (!ifp->displaymatrix_present ||
 | |
|             memcmp(sd->data, ifp->displaymatrix, sizeof(ifp->displaymatrix)))
 | |
|             need_reinit = 1;
 | |
|     } else if (ifp->displaymatrix_present)
 | |
|         need_reinit = 1;
 | |
| 
 | |
|     if (need_reinit) {
 | |
|         ret = ifilter_parameters_from_frame(ifilter, frame);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|     }
 | |
| 
 | |
|     /* (re)init the graph if possible, otherwise buffer the frame and return */
 | |
|     if (need_reinit || !fg->graph) {
 | |
|         if (!ifilter_has_all_input_formats(fg)) {
 | |
|             AVFrame *tmp = av_frame_clone(frame);
 | |
|             if (!tmp)
 | |
|                 return AVERROR(ENOMEM);
 | |
| 
 | |
|             ret = av_fifo_write(ifp->frame_queue, &tmp, 1);
 | |
|             if (ret < 0)
 | |
|                 av_frame_free(&tmp);
 | |
| 
 | |
|             return ret;
 | |
|         }
 | |
| 
 | |
|         ret = reap_filters(0);
 | |
|         if (ret < 0 && ret != AVERROR_EOF) {
 | |
|             av_log(fg, AV_LOG_ERROR, "Error while filtering: %s\n", av_err2str(ret));
 | |
|             return ret;
 | |
|         }
 | |
| 
 | |
|         ret = configure_filtergraph(fg);
 | |
|         if (ret < 0) {
 | |
|             av_log(fg, AV_LOG_ERROR, "Error reinitializing filters!\n");
 | |
|             return ret;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (keep_reference) {
 | |
|         ret = av_frame_ref(ifp->frame, frame);
 | |
|         if (ret < 0)
 | |
|             return ret;
 | |
|     } else
 | |
|         av_frame_move_ref(ifp->frame, frame);
 | |
|     frame = ifp->frame;
 | |
| 
 | |
|     frame->pts       = av_rescale_q(frame->pts,      frame->time_base, ifp->time_base);
 | |
|     frame->duration  = av_rescale_q(frame->duration, frame->time_base, ifp->time_base);
 | |
|     frame->time_base = ifp->time_base;
 | |
| #if LIBAVUTIL_VERSION_MAJOR < 59
 | |
|     AV_NOWARN_DEPRECATED(
 | |
|     frame->pkt_duration = frame->duration;
 | |
|     )
 | |
| #endif
 | |
| 
 | |
|     ret = av_buffersrc_add_frame_flags(ifp->filter, frame,
 | |
|                                        AV_BUFFERSRC_FLAG_PUSH);
 | |
|     if (ret < 0) {
 | |
|         av_frame_unref(frame);
 | |
|         if (ret != AVERROR_EOF)
 | |
|             av_log(fg, AV_LOG_ERROR, "Error while filtering: %s\n", av_err2str(ret));
 | |
|         return ret;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| int fg_transcode_step(FilterGraph *graph, InputStream **best_ist)
 | |
| {
 | |
|     int i, ret;
 | |
|     int nb_requests, nb_requests_max = 0;
 | |
|     InputStream *ist;
 | |
| 
 | |
|     if (!graph->graph) {
 | |
|         for (int i = 0; i < graph->nb_inputs; i++) {
 | |
|             InputFilter *ifilter = graph->inputs[i];
 | |
|             InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
|             if (ifp->format < 0 && !ifp->eof) {
 | |
|                 *best_ist = ifp->ist;
 | |
|                 return 0;
 | |
|             }
 | |
|         }
 | |
| 
 | |
|         // graph not configured, but all inputs are either initialized or EOF
 | |
|         for (int i = 0; i < graph->nb_outputs; i++)
 | |
|             graph->outputs[i]->ost->inputs_done = 1;
 | |
| 
 | |
|         return 0;
 | |
|     }
 | |
| 
 | |
|     *best_ist = NULL;
 | |
|     ret = avfilter_graph_request_oldest(graph->graph);
 | |
|     if (ret >= 0)
 | |
|         return reap_filters(0);
 | |
| 
 | |
|     if (ret == AVERROR_EOF) {
 | |
|         ret = reap_filters(1);
 | |
|         for (i = 0; i < graph->nb_outputs; i++)
 | |
|             close_output_stream(graph->outputs[i]->ost);
 | |
|         return ret;
 | |
|     }
 | |
|     if (ret != AVERROR(EAGAIN))
 | |
|         return ret;
 | |
| 
 | |
|     for (i = 0; i < graph->nb_inputs; i++) {
 | |
|         InputFilter *ifilter = graph->inputs[i];
 | |
|         InputFilterPriv *ifp = ifp_from_ifilter(ifilter);
 | |
| 
 | |
|         ist = ifp->ist;
 | |
|         if (input_files[ist->file_index]->eagain || ifp->eof)
 | |
|             continue;
 | |
|         nb_requests = av_buffersrc_get_nb_failed_requests(ifp->filter);
 | |
|         if (nb_requests > nb_requests_max) {
 | |
|             nb_requests_max = nb_requests;
 | |
|             *best_ist = ist;
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (!*best_ist)
 | |
|         for (i = 0; i < graph->nb_outputs; i++)
 | |
|             graph->outputs[i]->ost->unavailable = 1;
 | |
| 
 | |
|     return 0;
 | |
| }
 | 
