mirror of
				https://github.com/nyanmisaka/ffmpeg-rockchip.git
				synced 2025-10-28 03:13:08 +08:00 
			
		
		
		
	
		
			
				
	
	
		
			1027 lines
		
	
	
		
			34 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1027 lines
		
	
	
		
			34 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (c) 2010 Mark Heath mjpeg0 @ silicontrip dot org
 | |
|  * Copyright (c) 2014 Clément Bœsch
 | |
|  * Copyright (c) 2014 Dave Rice @dericed
 | |
|  *
 | |
|  * This file is part of FFmpeg.
 | |
|  *
 | |
|  * FFmpeg is free software; you can redistribute it and/or
 | |
|  * modify it under the terms of the GNU Lesser General Public
 | |
|  * License as published by the Free Software Foundation; either
 | |
|  * version 2.1 of the License, or (at your option) any later version.
 | |
|  *
 | |
|  * FFmpeg is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 | |
|  * Lesser General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU Lesser General Public
 | |
|  * License along with FFmpeg; if not, write to the Free Software
 | |
|  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
 | |
|  */
 | |
| 
 | |
| #include "libavutil/intreadwrite.h"
 | |
| #include "libavutil/opt.h"
 | |
| #include "libavutil/pixdesc.h"
 | |
| #include "internal.h"
 | |
| 
 | |
| enum FilterMode {
 | |
|     FILTER_NONE = -1,
 | |
|     FILTER_TOUT,
 | |
|     FILTER_VREP,
 | |
|     FILTER_BRNG,
 | |
|     FILT_NUMB
 | |
| };
 | |
| 
 | |
| typedef struct SignalstatsContext {
 | |
|     const AVClass *class;
 | |
|     int chromah;    // height of chroma plane
 | |
|     int chromaw;    // width of chroma plane
 | |
|     int hsub;       // horizontal subsampling
 | |
|     int vsub;       // vertical subsampling
 | |
|     int depth;      // pixel depth
 | |
|     int fs;         // pixel count per frame
 | |
|     int cfs;        // pixel count per frame of chroma planes
 | |
|     int outfilter;  // FilterMode
 | |
|     int filters;
 | |
|     AVFrame *frame_prev;
 | |
|     uint8_t rgba_color[4];
 | |
|     int yuv_color[3];
 | |
|     int nb_jobs;
 | |
|     int *jobs_rets;
 | |
| 
 | |
|     int maxsize;    // history stats array size
 | |
|     int *histy, *histu, *histv, *histsat;
 | |
| 
 | |
|     AVFrame *frame_sat;
 | |
|     AVFrame *frame_hue;
 | |
| } SignalstatsContext;
 | |
| 
 | |
| typedef struct ThreadData {
 | |
|     const AVFrame *in;
 | |
|     AVFrame *out;
 | |
| } ThreadData;
 | |
| 
 | |
| typedef struct ThreadDataHueSatMetrics {
 | |
|     const AVFrame *src;
 | |
|     AVFrame *dst_sat, *dst_hue;
 | |
| } ThreadDataHueSatMetrics;
 | |
| 
 | |
| #define OFFSET(x) offsetof(SignalstatsContext, x)
 | |
| #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
 | |
| 
 | |
| static const AVOption signalstats_options[] = {
 | |
|     {"stat", "set statistics filters", OFFSET(filters), AV_OPT_TYPE_FLAGS, {.i64=0}, 0, INT_MAX, FLAGS, "filters"},
 | |
|         {"tout", "analyze pixels for temporal outliers",                0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_TOUT}, 0, 0, FLAGS, "filters"},
 | |
|         {"vrep", "analyze video lines for vertical line repetition",    0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_VREP}, 0, 0, FLAGS, "filters"},
 | |
|         {"brng", "analyze for pixels outside of broadcast range",       0, AV_OPT_TYPE_CONST, {.i64=1<<FILTER_BRNG}, 0, 0, FLAGS, "filters"},
 | |
|     {"out", "set video filter", OFFSET(outfilter), AV_OPT_TYPE_INT, {.i64=FILTER_NONE}, -1, FILT_NUMB-1, FLAGS, "out"},
 | |
|         {"tout", "highlight pixels that depict temporal outliers",              0, AV_OPT_TYPE_CONST, {.i64=FILTER_TOUT}, 0, 0, FLAGS, "out"},
 | |
|         {"vrep", "highlight video lines that depict vertical line repetition",  0, AV_OPT_TYPE_CONST, {.i64=FILTER_VREP}, 0, 0, FLAGS, "out"},
 | |
|         {"brng", "highlight pixels that are outside of broadcast range",        0, AV_OPT_TYPE_CONST, {.i64=FILTER_BRNG}, 0, 0, FLAGS, "out"},
 | |
|     {"c",     "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS},
 | |
|     {"color", "set highlight color", OFFSET(rgba_color), AV_OPT_TYPE_COLOR, {.str="yellow"}, .flags=FLAGS},
 | |
|     {NULL}
 | |
| };
 | |
| 
 | |
| AVFILTER_DEFINE_CLASS(signalstats);
 | |
| 
 | |
| static av_cold int init(AVFilterContext *ctx)
 | |
| {
 | |
|     uint8_t r, g, b;
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
| 
 | |
|     if (s->outfilter != FILTER_NONE)
 | |
|         s->filters |= 1 << s->outfilter;
 | |
| 
 | |
|     r = s->rgba_color[0];
 | |
|     g = s->rgba_color[1];
 | |
|     b = s->rgba_color[2];
 | |
|     s->yuv_color[0] = (( 66*r + 129*g +  25*b + (1<<7)) >> 8) +  16;
 | |
|     s->yuv_color[1] = ((-38*r + -74*g + 112*b + (1<<7)) >> 8) + 128;
 | |
|     s->yuv_color[2] = ((112*r + -94*g + -18*b + (1<<7)) >> 8) + 128;
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static av_cold void uninit(AVFilterContext *ctx)
 | |
| {
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
|     av_frame_free(&s->frame_prev);
 | |
|     av_frame_free(&s->frame_sat);
 | |
|     av_frame_free(&s->frame_hue);
 | |
|     av_freep(&s->jobs_rets);
 | |
|     av_freep(&s->histy);
 | |
|     av_freep(&s->histu);
 | |
|     av_freep(&s->histv);
 | |
|     av_freep(&s->histsat);
 | |
| }
 | |
| 
 | |
| static int query_formats(AVFilterContext *ctx)
 | |
| {
 | |
|     // TODO: add more
 | |
|     static const enum AVPixelFormat pix_fmts[] = {
 | |
|         AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P,
 | |
|         AV_PIX_FMT_YUV440P,
 | |
|         AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P,
 | |
|         AV_PIX_FMT_YUVJ440P,
 | |
|         AV_PIX_FMT_YUV444P9, AV_PIX_FMT_YUV422P9, AV_PIX_FMT_YUV420P9,
 | |
|         AV_PIX_FMT_YUV444P10, AV_PIX_FMT_YUV422P10, AV_PIX_FMT_YUV420P10,
 | |
|         AV_PIX_FMT_YUV440P10,
 | |
|         AV_PIX_FMT_YUV444P12, AV_PIX_FMT_YUV422P12, AV_PIX_FMT_YUV420P12,
 | |
|         AV_PIX_FMT_YUV440P12,
 | |
|         AV_PIX_FMT_YUV444P14, AV_PIX_FMT_YUV422P14, AV_PIX_FMT_YUV420P14,
 | |
|         AV_PIX_FMT_YUV444P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV420P16,
 | |
|         AV_PIX_FMT_NONE
 | |
|     };
 | |
| 
 | |
|     AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts);
 | |
|     if (!fmts_list)
 | |
|         return AVERROR(ENOMEM);
 | |
|     return ff_set_common_formats(ctx, fmts_list);
 | |
| }
 | |
| 
 | |
| static AVFrame *alloc_frame(enum AVPixelFormat pixfmt, int w, int h)
 | |
| {
 | |
|     AVFrame *frame = av_frame_alloc();
 | |
|     if (!frame)
 | |
|         return NULL;
 | |
| 
 | |
|     frame->format = pixfmt;
 | |
|     frame->width  = w;
 | |
|     frame->height = h;
 | |
| 
 | |
|     if (av_frame_get_buffer(frame, 0) < 0) {
 | |
|         av_frame_free(&frame);
 | |
|         return NULL;
 | |
|     }
 | |
| 
 | |
|     return frame;
 | |
| }
 | |
| 
 | |
| static int config_output(AVFilterLink *outlink)
 | |
| {
 | |
|     AVFilterContext *ctx = outlink->src;
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
|     AVFilterLink *inlink = outlink->src->inputs[0];
 | |
|     const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(outlink->format);
 | |
|     s->hsub = desc->log2_chroma_w;
 | |
|     s->vsub = desc->log2_chroma_h;
 | |
|     s->depth = desc->comp[0].depth;
 | |
|     s->maxsize = 1 << s->depth;
 | |
|     s->histy = av_malloc_array(s->maxsize, sizeof(*s->histy));
 | |
|     s->histu = av_malloc_array(s->maxsize, sizeof(*s->histu));
 | |
|     s->histv = av_malloc_array(s->maxsize, sizeof(*s->histv));
 | |
|     s->histsat = av_malloc_array(s->maxsize, sizeof(*s->histsat));
 | |
| 
 | |
|     if (!s->histy || !s->histu || !s->histv || !s->histsat)
 | |
|         return AVERROR(ENOMEM);
 | |
| 
 | |
|     outlink->w = inlink->w;
 | |
|     outlink->h = inlink->h;
 | |
| 
 | |
|     s->chromaw = AV_CEIL_RSHIFT(inlink->w, s->hsub);
 | |
|     s->chromah = AV_CEIL_RSHIFT(inlink->h, s->vsub);
 | |
| 
 | |
|     s->fs = inlink->w * inlink->h;
 | |
|     s->cfs = s->chromaw * s->chromah;
 | |
| 
 | |
|     s->nb_jobs   = FFMAX(1, FFMIN(inlink->h, ff_filter_get_nb_threads(ctx)));
 | |
|     s->jobs_rets = av_malloc_array(s->nb_jobs, sizeof(*s->jobs_rets));
 | |
|     if (!s->jobs_rets)
 | |
|         return AVERROR(ENOMEM);
 | |
| 
 | |
|     s->frame_sat = alloc_frame(s->depth > 8 ? AV_PIX_FMT_GRAY16 : AV_PIX_FMT_GRAY8,  inlink->w, inlink->h);
 | |
|     s->frame_hue = alloc_frame(AV_PIX_FMT_GRAY16, inlink->w, inlink->h);
 | |
|     if (!s->frame_sat || !s->frame_hue)
 | |
|         return AVERROR(ENOMEM);
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static void burn_frame8(const SignalstatsContext *s, AVFrame *f, int x, int y)
 | |
| {
 | |
|     const int chromax = x >> s->hsub;
 | |
|     const int chromay = y >> s->vsub;
 | |
|     f->data[0][y       * f->linesize[0] +       x] = s->yuv_color[0];
 | |
|     f->data[1][chromay * f->linesize[1] + chromax] = s->yuv_color[1];
 | |
|     f->data[2][chromay * f->linesize[2] + chromax] = s->yuv_color[2];
 | |
| }
 | |
| 
 | |
| static void burn_frame16(const SignalstatsContext *s, AVFrame *f, int x, int y)
 | |
| {
 | |
|     const int chromax = x >> s->hsub;
 | |
|     const int chromay = y >> s->vsub;
 | |
|     const int mult = 1 << (s->depth - 8);
 | |
|     AV_WN16(f->data[0] + y       * f->linesize[0] +       x * 2, s->yuv_color[0] * mult);
 | |
|     AV_WN16(f->data[1] + chromay * f->linesize[1] + chromax * 2, s->yuv_color[1] * mult);
 | |
|     AV_WN16(f->data[2] + chromay * f->linesize[2] + chromax * 2, s->yuv_color[2] * mult);
 | |
| }
 | |
| 
 | |
| static int filter8_brng(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     int x, y, score = 0;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
|         const int yc = y >> s->vsub;
 | |
|         const uint8_t *pluma    = &in->data[0][y  * in->linesize[0]];
 | |
|         const uint8_t *pchromau = &in->data[1][yc * in->linesize[1]];
 | |
|         const uint8_t *pchromav = &in->data[2][yc * in->linesize[2]];
 | |
| 
 | |
|         for (x = 0; x < w; x++) {
 | |
|             const int xc = x >> s->hsub;
 | |
|             const int luma    = pluma[x];
 | |
|             const int chromau = pchromau[xc];
 | |
|             const int chromav = pchromav[xc];
 | |
|             const int filt = luma    < 16 || luma    > 235 ||
 | |
|                 chromau < 16 || chromau > 240 ||
 | |
|                 chromav < 16 || chromav > 240;
 | |
|             score += filt;
 | |
|             if (out && filt)
 | |
|                 burn_frame8(s, out, x, y);
 | |
|         }
 | |
|     }
 | |
|     return score;
 | |
| }
 | |
| 
 | |
| static int filter16_brng(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int mult = 1 << (s->depth - 8);
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     int x, y, score = 0;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
|         const int yc = y >> s->vsub;
 | |
|         const uint16_t *pluma    = (uint16_t *)&in->data[0][y  * in->linesize[0]];
 | |
|         const uint16_t *pchromau = (uint16_t *)&in->data[1][yc * in->linesize[1]];
 | |
|         const uint16_t *pchromav = (uint16_t *)&in->data[2][yc * in->linesize[2]];
 | |
| 
 | |
|         for (x = 0; x < w; x++) {
 | |
|             const int xc = x >> s->hsub;
 | |
|             const int luma    = pluma[x];
 | |
|             const int chromau = pchromau[xc];
 | |
|             const int chromav = pchromav[xc];
 | |
|             const int filt = luma    < 16 * mult || luma    > 235 * mult ||
 | |
|                 chromau < 16 * mult || chromau > 240 * mult ||
 | |
|                 chromav < 16 * mult || chromav > 240 * mult;
 | |
|             score += filt;
 | |
|             if (out && filt)
 | |
|                 burn_frame16(s, out, x, y);
 | |
|         }
 | |
|     }
 | |
|     return score;
 | |
| }
 | |
| 
 | |
| static int filter_tout_outlier(uint8_t x, uint8_t y, uint8_t z)
 | |
| {
 | |
|     return ((abs(x - y) + abs (z - y)) / 2) - abs(z - x) > 4; // make 4 configurable?
 | |
| }
 | |
| 
 | |
| static int filter8_tout(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     const uint8_t *p = in->data[0];
 | |
|     int lw = in->linesize[0];
 | |
|     int x, y, score = 0, filt;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
| 
 | |
|         if (y - 1 < 0 || y + 1 >= h)
 | |
|             continue;
 | |
| 
 | |
|         // detect two pixels above and below (to eliminate interlace artefacts)
 | |
|         // should check that video format is infact interlaced.
 | |
| 
 | |
| #define FILTER(i, j) \
 | |
|         filter_tout_outlier(p[(y-j) * lw + x + i], \
 | |
|                             p[    y * lw + x + i], \
 | |
|                             p[(y+j) * lw + x + i])
 | |
| 
 | |
| #define FILTER3(j) (FILTER(-1, j) && FILTER(0, j) && FILTER(1, j))
 | |
| 
 | |
|         if (y - 2 >= 0 && y + 2 < h) {
 | |
|             for (x = 1; x < w - 1; x++) {
 | |
|                 filt = FILTER3(2) && FILTER3(1);
 | |
|                 score += filt;
 | |
|                 if (filt && out)
 | |
|                     burn_frame8(s, out, x, y);
 | |
|             }
 | |
|         } else {
 | |
|             for (x = 1; x < w - 1; x++) {
 | |
|                 filt = FILTER3(1);
 | |
|                 score += filt;
 | |
|                 if (filt && out)
 | |
|                     burn_frame8(s, out, x, y);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     return score;
 | |
| }
 | |
| 
 | |
| static int filter16_tout(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     const uint16_t *p = (uint16_t *)in->data[0];
 | |
|     int lw = in->linesize[0] / 2;
 | |
|     int x, y, score = 0, filt;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
| 
 | |
|         if (y - 1 < 0 || y + 1 >= h)
 | |
|             continue;
 | |
| 
 | |
|         // detect two pixels above and below (to eliminate interlace artefacts)
 | |
|         // should check that video format is infact interlaced.
 | |
| 
 | |
|         if (y - 2 >= 0 && y + 2 < h) {
 | |
|             for (x = 1; x < w - 1; x++) {
 | |
|                 filt = FILTER3(2) && FILTER3(1);
 | |
|                 score += filt;
 | |
|                 if (filt && out)
 | |
|                     burn_frame16(s, out, x, y);
 | |
|             }
 | |
|         } else {
 | |
|             for (x = 1; x < w - 1; x++) {
 | |
|                 filt = FILTER3(1);
 | |
|                 score += filt;
 | |
|                 if (filt && out)
 | |
|                     burn_frame16(s, out, x, y);
 | |
|             }
 | |
|         }
 | |
|     }
 | |
|     return score;
 | |
| }
 | |
| 
 | |
| #define VREP_START 4
 | |
| 
 | |
| static int filter8_vrep(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     const uint8_t *p = in->data[0];
 | |
|     const int lw = in->linesize[0];
 | |
|     int x, y, score = 0;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
|         const int y2lw = (y - VREP_START) * lw;
 | |
|         const int ylw  =  y               * lw;
 | |
|         int filt, totdiff = 0;
 | |
| 
 | |
|         if (y < VREP_START)
 | |
|             continue;
 | |
| 
 | |
|         for (x = 0; x < w; x++)
 | |
|             totdiff += abs(p[y2lw + x] - p[ylw + x]);
 | |
|         filt = totdiff < w;
 | |
| 
 | |
|         score += filt;
 | |
|         if (filt && out)
 | |
|             for (x = 0; x < w; x++)
 | |
|                 burn_frame8(s, out, x, y);
 | |
|     }
 | |
|     return score * w;
 | |
| }
 | |
| 
 | |
| static int filter16_vrep(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     ThreadData *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *in = td->in;
 | |
|     AVFrame *out = td->out;
 | |
|     const int w = in->width;
 | |
|     const int h = in->height;
 | |
|     const int slice_start = (h *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (h * (jobnr+1)) / nb_jobs;
 | |
|     const uint16_t *p = (uint16_t *)in->data[0];
 | |
|     const int lw = in->linesize[0] / 2;
 | |
|     int x, y, score = 0;
 | |
| 
 | |
|     for (y = slice_start; y < slice_end; y++) {
 | |
|         const int y2lw = (y - VREP_START) * lw;
 | |
|         const int ylw  =  y               * lw;
 | |
|         int64_t totdiff = 0;
 | |
|         int filt;
 | |
| 
 | |
|         if (y < VREP_START)
 | |
|             continue;
 | |
| 
 | |
|         for (x = 0; x < w; x++)
 | |
|             totdiff += abs(p[y2lw + x] - p[ylw + x]);
 | |
|         filt = totdiff < w;
 | |
| 
 | |
|         score += filt;
 | |
|         if (filt && out)
 | |
|             for (x = 0; x < w; x++)
 | |
|                 burn_frame16(s, out, x, y);
 | |
|     }
 | |
|     return score * w;
 | |
| }
 | |
| 
 | |
| static const struct {
 | |
|     const char *name;
 | |
|     int (*process8)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
 | |
|     int (*process16)(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs);
 | |
| } filters_def[] = {
 | |
|     {"TOUT", filter8_tout, filter16_tout},
 | |
|     {"VREP", filter8_vrep, filter16_vrep},
 | |
|     {"BRNG", filter8_brng, filter16_brng},
 | |
|     {NULL}
 | |
| };
 | |
| 
 | |
| static int compute_sat_hue_metrics8(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     int i, j;
 | |
|     ThreadDataHueSatMetrics *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *src = td->src;
 | |
|     AVFrame *dst_sat = td->dst_sat;
 | |
|     AVFrame *dst_hue = td->dst_hue;
 | |
| 
 | |
|     const int slice_start = (s->chromah *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (s->chromah * (jobnr+1)) / nb_jobs;
 | |
| 
 | |
|     const int lsz_u = src->linesize[1];
 | |
|     const int lsz_v = src->linesize[2];
 | |
|     const uint8_t *p_u = src->data[1] + slice_start * lsz_u;
 | |
|     const uint8_t *p_v = src->data[2] + slice_start * lsz_v;
 | |
| 
 | |
|     const int lsz_sat = dst_sat->linesize[0];
 | |
|     const int lsz_hue = dst_hue->linesize[0];
 | |
|     uint8_t *p_sat = dst_sat->data[0] + slice_start * lsz_sat;
 | |
|     uint8_t *p_hue = dst_hue->data[0] + slice_start * lsz_hue;
 | |
| 
 | |
|     for (j = slice_start; j < slice_end; j++) {
 | |
|         for (i = 0; i < s->chromaw; i++) {
 | |
|             const int yuvu = p_u[i];
 | |
|             const int yuvv = p_v[i];
 | |
|             p_sat[i] = hypot(yuvu - 128, yuvv - 128); // int or round?
 | |
|             ((int16_t*)p_hue)[i] = fmod(floor((180 / M_PI) * atan2f(yuvu-128, yuvv-128) + 180), 360.);
 | |
|         }
 | |
|         p_u   += lsz_u;
 | |
|         p_v   += lsz_v;
 | |
|         p_sat += lsz_sat;
 | |
|         p_hue += lsz_hue;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static int compute_sat_hue_metrics16(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
 | |
| {
 | |
|     int i, j;
 | |
|     ThreadDataHueSatMetrics *td = arg;
 | |
|     const SignalstatsContext *s = ctx->priv;
 | |
|     const AVFrame *src = td->src;
 | |
|     AVFrame *dst_sat = td->dst_sat;
 | |
|     AVFrame *dst_hue = td->dst_hue;
 | |
|     const int mid = 1 << (s->depth - 1);
 | |
| 
 | |
|     const int slice_start = (s->chromah *  jobnr   ) / nb_jobs;
 | |
|     const int slice_end   = (s->chromah * (jobnr+1)) / nb_jobs;
 | |
| 
 | |
|     const int lsz_u = src->linesize[1] / 2;
 | |
|     const int lsz_v = src->linesize[2] / 2;
 | |
|     const uint16_t *p_u = (uint16_t*)src->data[1] + slice_start * lsz_u;
 | |
|     const uint16_t *p_v = (uint16_t*)src->data[2] + slice_start * lsz_v;
 | |
| 
 | |
|     const int lsz_sat = dst_sat->linesize[0] / 2;
 | |
|     const int lsz_hue = dst_hue->linesize[0] / 2;
 | |
|     uint16_t *p_sat = (uint16_t*)dst_sat->data[0] + slice_start * lsz_sat;
 | |
|     uint16_t *p_hue = (uint16_t*)dst_hue->data[0] + slice_start * lsz_hue;
 | |
| 
 | |
|     for (j = slice_start; j < slice_end; j++) {
 | |
|         for (i = 0; i < s->chromaw; i++) {
 | |
|             const int yuvu = p_u[i];
 | |
|             const int yuvv = p_v[i];
 | |
|             p_sat[i] = hypot(yuvu - mid, yuvv - mid); // int or round?
 | |
|             ((int16_t*)p_hue)[i] = fmod(floor((180 / M_PI) * atan2f(yuvu-mid, yuvv-mid) + 180), 360.);
 | |
|         }
 | |
|         p_u   += lsz_u;
 | |
|         p_v   += lsz_v;
 | |
|         p_sat += lsz_sat;
 | |
|         p_hue += lsz_hue;
 | |
|     }
 | |
| 
 | |
|     return 0;
 | |
| }
 | |
| 
 | |
| static unsigned compute_bit_depth(uint16_t mask)
 | |
| {
 | |
|     return av_popcount(mask);
 | |
| }
 | |
| 
 | |
| static int filter_frame8(AVFilterLink *link, AVFrame *in)
 | |
| {
 | |
|     AVFilterContext *ctx = link->dst;
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
|     AVFilterLink *outlink = ctx->outputs[0];
 | |
|     AVFrame *out = in;
 | |
|     int i, j;
 | |
|     int  w = 0,  cw = 0, // in
 | |
|         pw = 0, cpw = 0; // prev
 | |
|     int fil;
 | |
|     char metabuf[128];
 | |
|     unsigned int *histy = s->histy,
 | |
|                  *histu = s->histu,
 | |
|                  *histv = s->histv,
 | |
|                  histhue[360] = {0},
 | |
|                  *histsat = s->histsat;
 | |
|     int miny  = -1, minu  = -1, minv  = -1;
 | |
|     int maxy  = -1, maxu  = -1, maxv  = -1;
 | |
|     int lowy  = -1, lowu  = -1, lowv  = -1;
 | |
|     int highy = -1, highu = -1, highv = -1;
 | |
|     int minsat = -1, maxsat = -1, lowsat = -1, highsat = -1;
 | |
|     int lowp, highp, clowp, chighp;
 | |
|     int accy, accu, accv;
 | |
|     int accsat, acchue = 0;
 | |
|     int medhue, maxhue;
 | |
|     int toty = 0, totu = 0, totv = 0, totsat=0;
 | |
|     int tothue = 0;
 | |
|     int dify = 0, difu = 0, difv = 0;
 | |
|     uint16_t masky = 0, masku = 0, maskv = 0;
 | |
| 
 | |
|     int filtot[FILT_NUMB] = {0};
 | |
|     AVFrame *prev;
 | |
| 
 | |
|     AVFrame *sat = s->frame_sat;
 | |
|     AVFrame *hue = s->frame_hue;
 | |
|     const uint8_t *p_sat = sat->data[0];
 | |
|     const uint8_t *p_hue = hue->data[0];
 | |
|     const int lsz_sat = sat->linesize[0];
 | |
|     const int lsz_hue = hue->linesize[0];
 | |
|     ThreadDataHueSatMetrics td_huesat = {
 | |
|         .src     = in,
 | |
|         .dst_sat = sat,
 | |
|         .dst_hue = hue,
 | |
|     };
 | |
| 
 | |
|     if (!s->frame_prev)
 | |
|         s->frame_prev = av_frame_clone(in);
 | |
| 
 | |
|     prev = s->frame_prev;
 | |
| 
 | |
|     if (s->outfilter != FILTER_NONE) {
 | |
|         out = av_frame_clone(in);
 | |
|         av_frame_make_writable(out);
 | |
|     }
 | |
| 
 | |
|     ctx->internal->execute(ctx, compute_sat_hue_metrics8, &td_huesat,
 | |
|                            NULL, FFMIN(s->chromah, ff_filter_get_nb_threads(ctx)));
 | |
| 
 | |
|     // Calculate luma histogram and difference with previous frame or field.
 | |
|     memset(s->histy, 0, s->maxsize * sizeof(*s->histy));
 | |
|     for (j = 0; j < link->h; j++) {
 | |
|         for (i = 0; i < link->w; i++) {
 | |
|             const int yuv = in->data[0][w + i];
 | |
| 
 | |
|             masky |= yuv;
 | |
|             histy[yuv]++;
 | |
|             dify += abs(yuv - prev->data[0][pw + i]);
 | |
|         }
 | |
|         w  += in->linesize[0];
 | |
|         pw += prev->linesize[0];
 | |
|     }
 | |
| 
 | |
|     // Calculate chroma histogram and difference with previous frame or field.
 | |
|     memset(s->histu, 0, s->maxsize * sizeof(*s->histu));
 | |
|     memset(s->histv, 0, s->maxsize * sizeof(*s->histv));
 | |
|     memset(s->histsat, 0, s->maxsize * sizeof(*s->histsat));
 | |
|     for (j = 0; j < s->chromah; j++) {
 | |
|         for (i = 0; i < s->chromaw; i++) {
 | |
|             const int yuvu = in->data[1][cw+i];
 | |
|             const int yuvv = in->data[2][cw+i];
 | |
| 
 | |
|             masku |= yuvu;
 | |
|             maskv |= yuvv;
 | |
|             histu[yuvu]++;
 | |
|             difu += abs(yuvu - prev->data[1][cpw+i]);
 | |
|             histv[yuvv]++;
 | |
|             difv += abs(yuvv - prev->data[2][cpw+i]);
 | |
| 
 | |
|             histsat[p_sat[i]]++;
 | |
|             histhue[((int16_t*)p_hue)[i]]++;
 | |
|         }
 | |
|         cw  += in->linesize[1];
 | |
|         cpw += prev->linesize[1];
 | |
|         p_sat += lsz_sat;
 | |
|         p_hue += lsz_hue;
 | |
|     }
 | |
| 
 | |
|     for (fil = 0; fil < FILT_NUMB; fil ++) {
 | |
|         if (s->filters & 1<<fil) {
 | |
|             ThreadData td = {
 | |
|                 .in = in,
 | |
|                 .out = out != in && s->outfilter == fil ? out : NULL,
 | |
|             };
 | |
|             memset(s->jobs_rets, 0, s->nb_jobs * sizeof(*s->jobs_rets));
 | |
|             ctx->internal->execute(ctx, filters_def[fil].process8,
 | |
|                                    &td, s->jobs_rets, s->nb_jobs);
 | |
|             for (i = 0; i < s->nb_jobs; i++)
 | |
|                 filtot[fil] += s->jobs_rets[i];
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     // find low / high based on histogram percentile
 | |
|     // these only need to be calculated once.
 | |
| 
 | |
|     lowp   = lrint(s->fs  * 10 / 100.);
 | |
|     highp  = lrint(s->fs  * 90 / 100.);
 | |
|     clowp  = lrint(s->cfs * 10 / 100.);
 | |
|     chighp = lrint(s->cfs * 90 / 100.);
 | |
| 
 | |
|     accy = accu = accv = accsat = 0;
 | |
|     for (fil = 0; fil < s->maxsize; fil++) {
 | |
|         if (miny   < 0 && histy[fil])   miny = fil;
 | |
|         if (minu   < 0 && histu[fil])   minu = fil;
 | |
|         if (minv   < 0 && histv[fil])   minv = fil;
 | |
|         if (minsat < 0 && histsat[fil]) minsat = fil;
 | |
| 
 | |
|         if (histy[fil])   maxy   = fil;
 | |
|         if (histu[fil])   maxu   = fil;
 | |
|         if (histv[fil])   maxv   = fil;
 | |
|         if (histsat[fil]) maxsat = fil;
 | |
| 
 | |
|         toty   += histy[fil]   * fil;
 | |
|         totu   += histu[fil]   * fil;
 | |
|         totv   += histv[fil]   * fil;
 | |
|         totsat += histsat[fil] * fil;
 | |
| 
 | |
|         accy   += histy[fil];
 | |
|         accu   += histu[fil];
 | |
|         accv   += histv[fil];
 | |
|         accsat += histsat[fil];
 | |
| 
 | |
|         if (lowy   == -1 && accy   >=  lowp) lowy   = fil;
 | |
|         if (lowu   == -1 && accu   >= clowp) lowu   = fil;
 | |
|         if (lowv   == -1 && accv   >= clowp) lowv   = fil;
 | |
|         if (lowsat == -1 && accsat >= clowp) lowsat = fil;
 | |
| 
 | |
|         if (highy   == -1 && accy   >=  highp) highy   = fil;
 | |
|         if (highu   == -1 && accu   >= chighp) highu   = fil;
 | |
|         if (highv   == -1 && accv   >= chighp) highv   = fil;
 | |
|         if (highsat == -1 && accsat >= chighp) highsat = fil;
 | |
|     }
 | |
| 
 | |
|     maxhue = histhue[0];
 | |
|     medhue = -1;
 | |
|     for (fil = 0; fil < 360; fil++) {
 | |
|         tothue += histhue[fil] * fil;
 | |
|         acchue += histhue[fil];
 | |
| 
 | |
|         if (medhue == -1 && acchue > s->cfs / 2)
 | |
|             medhue = fil;
 | |
|         if (histhue[fil] > maxhue) {
 | |
|             maxhue = histhue[fil];
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     av_frame_free(&s->frame_prev);
 | |
|     s->frame_prev = av_frame_clone(in);
 | |
| 
 | |
| #define SET_META(key, fmt, val) do {                                \
 | |
|     snprintf(metabuf, sizeof(metabuf), fmt, val);                   \
 | |
|     av_dict_set(&out->metadata, "lavfi.signalstats." key, metabuf, 0);   \
 | |
| } while (0)
 | |
| 
 | |
|     SET_META("YMIN",    "%d", miny);
 | |
|     SET_META("YLOW",    "%d", lowy);
 | |
|     SET_META("YAVG",    "%g", 1.0 * toty / s->fs);
 | |
|     SET_META("YHIGH",   "%d", highy);
 | |
|     SET_META("YMAX",    "%d", maxy);
 | |
| 
 | |
|     SET_META("UMIN",    "%d", minu);
 | |
|     SET_META("ULOW",    "%d", lowu);
 | |
|     SET_META("UAVG",    "%g", 1.0 * totu / s->cfs);
 | |
|     SET_META("UHIGH",   "%d", highu);
 | |
|     SET_META("UMAX",    "%d", maxu);
 | |
| 
 | |
|     SET_META("VMIN",    "%d", minv);
 | |
|     SET_META("VLOW",    "%d", lowv);
 | |
|     SET_META("VAVG",    "%g", 1.0 * totv / s->cfs);
 | |
|     SET_META("VHIGH",   "%d", highv);
 | |
|     SET_META("VMAX",    "%d", maxv);
 | |
| 
 | |
|     SET_META("SATMIN",  "%d", minsat);
 | |
|     SET_META("SATLOW",  "%d", lowsat);
 | |
|     SET_META("SATAVG",  "%g", 1.0 * totsat / s->cfs);
 | |
|     SET_META("SATHIGH", "%d", highsat);
 | |
|     SET_META("SATMAX",  "%d", maxsat);
 | |
| 
 | |
|     SET_META("HUEMED",  "%d", medhue);
 | |
|     SET_META("HUEAVG",  "%g", 1.0 * tothue / s->cfs);
 | |
| 
 | |
|     SET_META("YDIF",    "%g", 1.0 * dify / s->fs);
 | |
|     SET_META("UDIF",    "%g", 1.0 * difu / s->cfs);
 | |
|     SET_META("VDIF",    "%g", 1.0 * difv / s->cfs);
 | |
| 
 | |
|     SET_META("YBITDEPTH", "%d", compute_bit_depth(masky));
 | |
|     SET_META("UBITDEPTH", "%d", compute_bit_depth(masku));
 | |
|     SET_META("VBITDEPTH", "%d", compute_bit_depth(maskv));
 | |
| 
 | |
|     for (fil = 0; fil < FILT_NUMB; fil ++) {
 | |
|         if (s->filters & 1<<fil) {
 | |
|             char metaname[128];
 | |
|             snprintf(metabuf,  sizeof(metabuf),  "%g", 1.0 * filtot[fil] / s->fs);
 | |
|             snprintf(metaname, sizeof(metaname), "lavfi.signalstats.%s", filters_def[fil].name);
 | |
|             av_dict_set(&out->metadata, metaname, metabuf, 0);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (in != out)
 | |
|         av_frame_free(&in);
 | |
|     return ff_filter_frame(outlink, out);
 | |
| }
 | |
| 
 | |
| static int filter_frame16(AVFilterLink *link, AVFrame *in)
 | |
| {
 | |
|     AVFilterContext *ctx = link->dst;
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
|     AVFilterLink *outlink = ctx->outputs[0];
 | |
|     AVFrame *out = in;
 | |
|     int i, j;
 | |
|     int  w = 0,  cw = 0, // in
 | |
|         pw = 0, cpw = 0; // prev
 | |
|     int fil;
 | |
|     char metabuf[128];
 | |
|     unsigned int *histy = s->histy,
 | |
|                  *histu = s->histu,
 | |
|                  *histv = s->histv,
 | |
|                  histhue[360] = {0},
 | |
|                  *histsat = s->histsat;
 | |
|     int miny  = -1, minu  = -1, minv  = -1;
 | |
|     int maxy  = -1, maxu  = -1, maxv  = -1;
 | |
|     int lowy  = -1, lowu  = -1, lowv  = -1;
 | |
|     int highy = -1, highu = -1, highv = -1;
 | |
|     int minsat = -1, maxsat = -1, lowsat = -1, highsat = -1;
 | |
|     int lowp, highp, clowp, chighp;
 | |
|     int accy, accu, accv;
 | |
|     int accsat, acchue = 0;
 | |
|     int medhue, maxhue;
 | |
|     int64_t toty = 0, totu = 0, totv = 0, totsat=0;
 | |
|     int64_t tothue = 0;
 | |
|     int64_t dify = 0, difu = 0, difv = 0;
 | |
|     uint16_t masky = 0, masku = 0, maskv = 0;
 | |
| 
 | |
|     int filtot[FILT_NUMB] = {0};
 | |
|     AVFrame *prev;
 | |
| 
 | |
|     AVFrame *sat = s->frame_sat;
 | |
|     AVFrame *hue = s->frame_hue;
 | |
|     const uint16_t *p_sat = (uint16_t *)sat->data[0];
 | |
|     const uint16_t *p_hue = (uint16_t *)hue->data[0];
 | |
|     const int lsz_sat = sat->linesize[0] / 2;
 | |
|     const int lsz_hue = hue->linesize[0] / 2;
 | |
|     ThreadDataHueSatMetrics td_huesat = {
 | |
|         .src     = in,
 | |
|         .dst_sat = sat,
 | |
|         .dst_hue = hue,
 | |
|     };
 | |
| 
 | |
|     if (!s->frame_prev)
 | |
|         s->frame_prev = av_frame_clone(in);
 | |
| 
 | |
|     prev = s->frame_prev;
 | |
| 
 | |
|     if (s->outfilter != FILTER_NONE) {
 | |
|         out = av_frame_clone(in);
 | |
|         av_frame_make_writable(out);
 | |
|     }
 | |
| 
 | |
|     ctx->internal->execute(ctx, compute_sat_hue_metrics16, &td_huesat,
 | |
|                            NULL, FFMIN(s->chromah, ff_filter_get_nb_threads(ctx)));
 | |
| 
 | |
|     // Calculate luma histogram and difference with previous frame or field.
 | |
|     memset(s->histy, 0, s->maxsize * sizeof(*s->histy));
 | |
|     for (j = 0; j < link->h; j++) {
 | |
|         for (i = 0; i < link->w; i++) {
 | |
|             const int yuv = AV_RN16(in->data[0] + w + i * 2);
 | |
| 
 | |
|             masky |= yuv;
 | |
|             histy[yuv]++;
 | |
|             dify += abs(yuv - (int)AV_RN16(prev->data[0] + pw + i * 2));
 | |
|         }
 | |
|         w  += in->linesize[0];
 | |
|         pw += prev->linesize[0];
 | |
|     }
 | |
| 
 | |
|     // Calculate chroma histogram and difference with previous frame or field.
 | |
|     memset(s->histu, 0, s->maxsize * sizeof(*s->histu));
 | |
|     memset(s->histv, 0, s->maxsize * sizeof(*s->histv));
 | |
|     memset(s->histsat, 0, s->maxsize * sizeof(*s->histsat));
 | |
|     for (j = 0; j < s->chromah; j++) {
 | |
|         for (i = 0; i < s->chromaw; i++) {
 | |
|             const int yuvu = AV_RN16(in->data[1] + cw + i * 2);
 | |
|             const int yuvv = AV_RN16(in->data[2] + cw + i * 2);
 | |
| 
 | |
|             masku |= yuvu;
 | |
|             maskv |= yuvv;
 | |
|             histu[yuvu]++;
 | |
|             difu += abs(yuvu - (int)AV_RN16(prev->data[1] + cpw + i * 2));
 | |
|             histv[yuvv]++;
 | |
|             difv += abs(yuvv - (int)AV_RN16(prev->data[2] + cpw + i * 2));
 | |
| 
 | |
|             histsat[p_sat[i]]++;
 | |
|             histhue[((int16_t*)p_hue)[i]]++;
 | |
|         }
 | |
|         cw  += in->linesize[1];
 | |
|         cpw += prev->linesize[1];
 | |
|         p_sat += lsz_sat;
 | |
|         p_hue += lsz_hue;
 | |
|     }
 | |
| 
 | |
|     for (fil = 0; fil < FILT_NUMB; fil ++) {
 | |
|         if (s->filters & 1<<fil) {
 | |
|             ThreadData td = {
 | |
|                 .in = in,
 | |
|                 .out = out != in && s->outfilter == fil ? out : NULL,
 | |
|             };
 | |
|             memset(s->jobs_rets, 0, s->nb_jobs * sizeof(*s->jobs_rets));
 | |
|             ctx->internal->execute(ctx, filters_def[fil].process16,
 | |
|                                    &td, s->jobs_rets, s->nb_jobs);
 | |
|             for (i = 0; i < s->nb_jobs; i++)
 | |
|                 filtot[fil] += s->jobs_rets[i];
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     // find low / high based on histogram percentile
 | |
|     // these only need to be calculated once.
 | |
| 
 | |
|     lowp   = lrint(s->fs  * 10 / 100.);
 | |
|     highp  = lrint(s->fs  * 90 / 100.);
 | |
|     clowp  = lrint(s->cfs * 10 / 100.);
 | |
|     chighp = lrint(s->cfs * 90 / 100.);
 | |
| 
 | |
|     accy = accu = accv = accsat = 0;
 | |
|     for (fil = 0; fil < s->maxsize; fil++) {
 | |
|         if (miny   < 0 && histy[fil])   miny = fil;
 | |
|         if (minu   < 0 && histu[fil])   minu = fil;
 | |
|         if (minv   < 0 && histv[fil])   minv = fil;
 | |
|         if (minsat < 0 && histsat[fil]) minsat = fil;
 | |
| 
 | |
|         if (histy[fil])   maxy   = fil;
 | |
|         if (histu[fil])   maxu   = fil;
 | |
|         if (histv[fil])   maxv   = fil;
 | |
|         if (histsat[fil]) maxsat = fil;
 | |
| 
 | |
|         toty   += histy[fil]   * fil;
 | |
|         totu   += histu[fil]   * fil;
 | |
|         totv   += histv[fil]   * fil;
 | |
|         totsat += histsat[fil] * fil;
 | |
| 
 | |
|         accy   += histy[fil];
 | |
|         accu   += histu[fil];
 | |
|         accv   += histv[fil];
 | |
|         accsat += histsat[fil];
 | |
| 
 | |
|         if (lowy   == -1 && accy   >=  lowp) lowy   = fil;
 | |
|         if (lowu   == -1 && accu   >= clowp) lowu   = fil;
 | |
|         if (lowv   == -1 && accv   >= clowp) lowv   = fil;
 | |
|         if (lowsat == -1 && accsat >= clowp) lowsat = fil;
 | |
| 
 | |
|         if (highy   == -1 && accy   >=  highp) highy   = fil;
 | |
|         if (highu   == -1 && accu   >= chighp) highu   = fil;
 | |
|         if (highv   == -1 && accv   >= chighp) highv   = fil;
 | |
|         if (highsat == -1 && accsat >= chighp) highsat = fil;
 | |
|     }
 | |
| 
 | |
|     maxhue = histhue[0];
 | |
|     medhue = -1;
 | |
|     for (fil = 0; fil < 360; fil++) {
 | |
|         tothue += histhue[fil] * fil;
 | |
|         acchue += histhue[fil];
 | |
| 
 | |
|         if (medhue == -1 && acchue > s->cfs / 2)
 | |
|             medhue = fil;
 | |
|         if (histhue[fil] > maxhue) {
 | |
|             maxhue = histhue[fil];
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     av_frame_free(&s->frame_prev);
 | |
|     s->frame_prev = av_frame_clone(in);
 | |
| 
 | |
|     SET_META("YMIN",    "%d", miny);
 | |
|     SET_META("YLOW",    "%d", lowy);
 | |
|     SET_META("YAVG",    "%g", 1.0 * toty / s->fs);
 | |
|     SET_META("YHIGH",   "%d", highy);
 | |
|     SET_META("YMAX",    "%d", maxy);
 | |
| 
 | |
|     SET_META("UMIN",    "%d", minu);
 | |
|     SET_META("ULOW",    "%d", lowu);
 | |
|     SET_META("UAVG",    "%g", 1.0 * totu / s->cfs);
 | |
|     SET_META("UHIGH",   "%d", highu);
 | |
|     SET_META("UMAX",    "%d", maxu);
 | |
| 
 | |
|     SET_META("VMIN",    "%d", minv);
 | |
|     SET_META("VLOW",    "%d", lowv);
 | |
|     SET_META("VAVG",    "%g", 1.0 * totv / s->cfs);
 | |
|     SET_META("VHIGH",   "%d", highv);
 | |
|     SET_META("VMAX",    "%d", maxv);
 | |
| 
 | |
|     SET_META("SATMIN",  "%d", minsat);
 | |
|     SET_META("SATLOW",  "%d", lowsat);
 | |
|     SET_META("SATAVG",  "%g", 1.0 * totsat / s->cfs);
 | |
|     SET_META("SATHIGH", "%d", highsat);
 | |
|     SET_META("SATMAX",  "%d", maxsat);
 | |
| 
 | |
|     SET_META("HUEMED",  "%d", medhue);
 | |
|     SET_META("HUEAVG",  "%g", 1.0 * tothue / s->cfs);
 | |
| 
 | |
|     SET_META("YDIF",    "%g", 1.0 * dify / s->fs);
 | |
|     SET_META("UDIF",    "%g", 1.0 * difu / s->cfs);
 | |
|     SET_META("VDIF",    "%g", 1.0 * difv / s->cfs);
 | |
| 
 | |
|     SET_META("YBITDEPTH", "%d", compute_bit_depth(masky));
 | |
|     SET_META("UBITDEPTH", "%d", compute_bit_depth(masku));
 | |
|     SET_META("VBITDEPTH", "%d", compute_bit_depth(maskv));
 | |
| 
 | |
|     for (fil = 0; fil < FILT_NUMB; fil ++) {
 | |
|         if (s->filters & 1<<fil) {
 | |
|             char metaname[128];
 | |
|             snprintf(metabuf,  sizeof(metabuf),  "%g", 1.0 * filtot[fil] / s->fs);
 | |
|             snprintf(metaname, sizeof(metaname), "lavfi.signalstats.%s", filters_def[fil].name);
 | |
|             av_dict_set(&out->metadata, metaname, metabuf, 0);
 | |
|         }
 | |
|     }
 | |
| 
 | |
|     if (in != out)
 | |
|         av_frame_free(&in);
 | |
|     return ff_filter_frame(outlink, out);
 | |
| }
 | |
| 
 | |
| static int filter_frame(AVFilterLink *link, AVFrame *in)
 | |
| {
 | |
|     AVFilterContext *ctx = link->dst;
 | |
|     SignalstatsContext *s = ctx->priv;
 | |
| 
 | |
|     if (s->depth > 8)
 | |
|         return filter_frame16(link, in);
 | |
|     else
 | |
|         return filter_frame8(link, in);
 | |
| }
 | |
| 
 | |
| static const AVFilterPad signalstats_inputs[] = {
 | |
|     {
 | |
|         .name           = "default",
 | |
|         .type           = AVMEDIA_TYPE_VIDEO,
 | |
|         .filter_frame   = filter_frame,
 | |
|     },
 | |
|     { NULL }
 | |
| };
 | |
| 
 | |
| static const AVFilterPad signalstats_outputs[] = {
 | |
|     {
 | |
|         .name           = "default",
 | |
|         .config_props   = config_output,
 | |
|         .type           = AVMEDIA_TYPE_VIDEO,
 | |
|     },
 | |
|     { NULL }
 | |
| };
 | |
| 
 | |
| AVFilter ff_vf_signalstats = {
 | |
|     .name          = "signalstats",
 | |
|     .description   = "Generate statistics from video analysis.",
 | |
|     .init          = init,
 | |
|     .uninit        = uninit,
 | |
|     .query_formats = query_formats,
 | |
|     .priv_size     = sizeof(SignalstatsContext),
 | |
|     .inputs        = signalstats_inputs,
 | |
|     .outputs       = signalstats_outputs,
 | |
|     .priv_class    = &signalstats_class,
 | |
|     .flags         = AVFILTER_FLAG_SLICE_THREADS,
 | |
| };
 | 
