yading@10: /* yading@10: * Copyright (c) 2011 Stefano Sabatini yading@10: * yading@10: * This file is part of FFmpeg. yading@10: * yading@10: * FFmpeg is free software; you can redistribute it and/or yading@10: * modify it under the terms of the GNU Lesser General Public yading@10: * License as published by the Free Software Foundation; either yading@10: * version 2.1 of the License, or (at your option) any later version. yading@10: * yading@10: * FFmpeg is distributed in the hope that it will be useful, yading@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of yading@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU yading@10: * Lesser General Public License for more details. yading@10: * yading@10: * You should have received a copy of the GNU Lesser General Public yading@10: * License along with FFmpeg; if not, write to the Free Software yading@10: * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA yading@10: */ yading@10: yading@10: /** yading@10: * @file yading@10: * filter for selecting which frame passes in the filterchain yading@10: */ yading@10: yading@10: #include "libavutil/avstring.h" yading@10: #include "libavutil/eval.h" yading@10: #include "libavutil/fifo.h" yading@10: #include "libavutil/internal.h" yading@10: #include "libavutil/opt.h" yading@10: #include "avfilter.h" yading@10: #include "audio.h" yading@10: #include "formats.h" yading@10: #include "internal.h" yading@10: #include "video.h" yading@10: yading@10: #if CONFIG_AVCODEC yading@10: #include "libavcodec/dsputil.h" yading@10: #endif yading@10: yading@10: static const char *const var_names[] = { yading@10: "TB", ///< timebase yading@10: yading@10: "pts", ///< original pts in the file of the frame yading@10: "start_pts", ///< first PTS in the stream, expressed in TB units yading@10: "prev_pts", ///< previous frame PTS yading@10: "prev_selected_pts", ///< previous selected frame PTS yading@10: yading@10: "t", ///< first PTS in seconds yading@10: "start_t", ///< first PTS in the stream, expressed in seconds yading@10: "prev_t", ///< previous frame time yading@10: "prev_selected_t", ///< previously selected time yading@10: yading@10: "pict_type", ///< the type of picture in the movie yading@10: "I", yading@10: "P", yading@10: "B", yading@10: "S", yading@10: "SI", yading@10: "SP", yading@10: "BI", yading@10: yading@10: "interlace_type", ///< the frame interlace type yading@10: "PROGRESSIVE", yading@10: "TOPFIRST", yading@10: "BOTTOMFIRST", yading@10: yading@10: "consumed_samples_n",///< number of samples consumed by the filter (only audio) yading@10: "samples_n", ///< number of samples in the current frame (only audio) yading@10: "sample_rate", ///< sample rate (only audio) yading@10: yading@10: "n", ///< frame number (starting from zero) yading@10: "selected_n", ///< selected frame number (starting from zero) yading@10: "prev_selected_n", ///< number of the last selected frame yading@10: yading@10: "key", ///< tell if the frame is a key frame yading@10: "pos", ///< original position in the file of the frame yading@10: yading@10: "scene", yading@10: yading@10: NULL yading@10: }; yading@10: yading@10: enum var_name { yading@10: VAR_TB, yading@10: yading@10: VAR_PTS, yading@10: VAR_START_PTS, yading@10: VAR_PREV_PTS, yading@10: VAR_PREV_SELECTED_PTS, yading@10: yading@10: VAR_T, yading@10: VAR_START_T, yading@10: VAR_PREV_T, yading@10: VAR_PREV_SELECTED_T, yading@10: yading@10: VAR_PICT_TYPE, yading@10: VAR_PICT_TYPE_I, yading@10: VAR_PICT_TYPE_P, yading@10: VAR_PICT_TYPE_B, yading@10: VAR_PICT_TYPE_S, yading@10: VAR_PICT_TYPE_SI, yading@10: VAR_PICT_TYPE_SP, yading@10: VAR_PICT_TYPE_BI, yading@10: yading@10: VAR_INTERLACE_TYPE, yading@10: VAR_INTERLACE_TYPE_P, yading@10: VAR_INTERLACE_TYPE_T, yading@10: VAR_INTERLACE_TYPE_B, yading@10: yading@10: VAR_CONSUMED_SAMPLES_N, yading@10: VAR_SAMPLES_N, yading@10: VAR_SAMPLE_RATE, yading@10: yading@10: VAR_N, yading@10: VAR_SELECTED_N, yading@10: VAR_PREV_SELECTED_N, yading@10: yading@10: VAR_KEY, yading@10: VAR_POS, yading@10: yading@10: VAR_SCENE, yading@10: yading@10: VAR_VARS_NB yading@10: }; yading@10: yading@10: typedef struct { yading@10: const AVClass *class; yading@10: char *expr_str; yading@10: AVExpr *expr; yading@10: double var_values[VAR_VARS_NB]; yading@10: int do_scene_detect; ///< 1 if the expression requires scene detection variables, 0 otherwise yading@10: #if CONFIG_AVCODEC yading@10: AVCodecContext *avctx; ///< codec context required for the DSPContext (scene detect only) yading@10: DSPContext c; ///< context providing optimized SAD methods (scene detect only) yading@10: double prev_mafd; ///< previous MAFD (scene detect only) yading@10: #endif yading@10: AVFrame *prev_picref; ///< previous frame (scene detect only) yading@10: double select; yading@10: int select_out; ///< mark the selected output pad index yading@10: int nb_outputs; yading@10: } SelectContext; yading@10: yading@10: #define OFFSET(x) offsetof(SelectContext, x) yading@10: #define DEFINE_OPTIONS(filt_name, FLAGS) \ yading@10: static const AVOption filt_name##_options[] = { \ yading@10: { "expr", "set an expression to use for selecting frames", OFFSET(expr_str), AV_OPT_TYPE_STRING, { .str = "1" }, .flags=FLAGS }, \ yading@10: { "e", "set an expression to use for selecting frames", OFFSET(expr_str), AV_OPT_TYPE_STRING, { .str = "1" }, .flags=FLAGS }, \ yading@10: { "outputs", "set the number of outputs", OFFSET(nb_outputs), AV_OPT_TYPE_INT, {.i64 = 1}, 1, INT_MAX, .flags=FLAGS }, \ yading@10: { "n", "set the number of outputs", OFFSET(nb_outputs), AV_OPT_TYPE_INT, {.i64 = 1}, 1, INT_MAX, .flags=FLAGS }, \ yading@10: { NULL } \ yading@10: } yading@10: yading@10: static int request_frame(AVFilterLink *outlink); yading@10: yading@10: static av_cold int init(AVFilterContext *ctx) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: int i, ret; yading@10: yading@10: if ((ret = av_expr_parse(&select->expr, select->expr_str, yading@10: var_names, NULL, NULL, NULL, NULL, 0, ctx)) < 0) { yading@10: av_log(ctx, AV_LOG_ERROR, "Error while parsing expression '%s'\n", yading@10: select->expr_str); yading@10: return ret; yading@10: } yading@10: select->do_scene_detect = !!strstr(select->expr_str, "scene"); yading@10: yading@10: for (i = 0; i < select->nb_outputs; i++) { yading@10: AVFilterPad pad = { 0 }; yading@10: yading@10: pad.name = av_asprintf("output%d", i); yading@10: if (!pad.name) yading@10: return AVERROR(ENOMEM); yading@10: pad.type = ctx->filter->inputs[0].type; yading@10: pad.request_frame = request_frame; yading@10: ff_insert_outpad(ctx, i, &pad); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: #define INTERLACE_TYPE_P 0 yading@10: #define INTERLACE_TYPE_T 1 yading@10: #define INTERLACE_TYPE_B 2 yading@10: yading@10: static int config_input(AVFilterLink *inlink) yading@10: { yading@10: SelectContext *select = inlink->dst->priv; yading@10: yading@10: select->var_values[VAR_N] = 0.0; yading@10: select->var_values[VAR_SELECTED_N] = 0.0; yading@10: yading@10: select->var_values[VAR_TB] = av_q2d(inlink->time_base); yading@10: yading@10: select->var_values[VAR_PREV_PTS] = NAN; yading@10: select->var_values[VAR_PREV_SELECTED_PTS] = NAN; yading@10: select->var_values[VAR_PREV_SELECTED_T] = NAN; yading@10: select->var_values[VAR_PREV_T] = NAN; yading@10: select->var_values[VAR_START_PTS] = NAN; yading@10: select->var_values[VAR_START_T] = NAN; yading@10: yading@10: select->var_values[VAR_PICT_TYPE_I] = AV_PICTURE_TYPE_I; yading@10: select->var_values[VAR_PICT_TYPE_P] = AV_PICTURE_TYPE_P; yading@10: select->var_values[VAR_PICT_TYPE_B] = AV_PICTURE_TYPE_B; yading@10: select->var_values[VAR_PICT_TYPE_SI] = AV_PICTURE_TYPE_SI; yading@10: select->var_values[VAR_PICT_TYPE_SP] = AV_PICTURE_TYPE_SP; yading@10: yading@10: select->var_values[VAR_INTERLACE_TYPE_P] = INTERLACE_TYPE_P; yading@10: select->var_values[VAR_INTERLACE_TYPE_T] = INTERLACE_TYPE_T; yading@10: select->var_values[VAR_INTERLACE_TYPE_B] = INTERLACE_TYPE_B; yading@10: yading@10: select->var_values[VAR_PICT_TYPE] = NAN; yading@10: select->var_values[VAR_INTERLACE_TYPE] = NAN; yading@10: select->var_values[VAR_SCENE] = NAN; yading@10: select->var_values[VAR_CONSUMED_SAMPLES_N] = NAN; yading@10: select->var_values[VAR_SAMPLES_N] = NAN; yading@10: yading@10: select->var_values[VAR_SAMPLE_RATE] = yading@10: inlink->type == AVMEDIA_TYPE_AUDIO ? inlink->sample_rate : NAN; yading@10: yading@10: #if CONFIG_AVCODEC yading@10: if (select->do_scene_detect) { yading@10: select->avctx = avcodec_alloc_context3(NULL); yading@10: if (!select->avctx) yading@10: return AVERROR(ENOMEM); yading@10: avpriv_dsputil_init(&select->c, select->avctx); yading@10: } yading@10: #endif yading@10: return 0; yading@10: } yading@10: yading@10: #if CONFIG_AVCODEC yading@10: static double get_scene_score(AVFilterContext *ctx, AVFrame *frame) yading@10: { yading@10: double ret = 0; yading@10: SelectContext *select = ctx->priv; yading@10: AVFrame *prev_picref = select->prev_picref; yading@10: yading@10: if (prev_picref && yading@10: frame->height == prev_picref->height && yading@10: frame->width == prev_picref->width && yading@10: frame->linesize[0] == prev_picref->linesize[0]) { yading@10: int x, y, nb_sad = 0; yading@10: int64_t sad = 0; yading@10: double mafd, diff; yading@10: uint8_t *p1 = frame->data[0]; yading@10: uint8_t *p2 = prev_picref->data[0]; yading@10: const int linesize = frame->linesize[0]; yading@10: yading@10: for (y = 0; y < frame->height - 8; y += 8) { yading@10: for (x = 0; x < frame->width*3 - 8; x += 8) { yading@10: sad += select->c.sad[1](select, p1 + x, p2 + x, yading@10: linesize, 8); yading@10: nb_sad += 8 * 8; yading@10: } yading@10: p1 += 8 * linesize; yading@10: p2 += 8 * linesize; yading@10: } yading@10: emms_c(); yading@10: mafd = nb_sad ? sad / nb_sad : 0; yading@10: diff = fabs(mafd - select->prev_mafd); yading@10: ret = av_clipf(FFMIN(mafd, diff) / 100., 0, 1); yading@10: select->prev_mafd = mafd; yading@10: av_frame_free(&prev_picref); yading@10: } yading@10: select->prev_picref = av_frame_clone(frame); yading@10: return ret; yading@10: } yading@10: #endif yading@10: yading@10: #define D2TS(d) (isnan(d) ? AV_NOPTS_VALUE : (int64_t)(d)) yading@10: #define TS2D(ts) ((ts) == AV_NOPTS_VALUE ? NAN : (double)(ts)) yading@10: yading@10: static void select_frame(AVFilterContext *ctx, AVFrame *frame) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: AVFilterLink *inlink = ctx->inputs[0]; yading@10: double res; yading@10: yading@10: if (isnan(select->var_values[VAR_START_PTS])) yading@10: select->var_values[VAR_START_PTS] = TS2D(frame->pts); yading@10: if (isnan(select->var_values[VAR_START_T])) yading@10: select->var_values[VAR_START_T] = TS2D(frame->pts) * av_q2d(inlink->time_base); yading@10: yading@10: select->var_values[VAR_PTS] = TS2D(frame->pts); yading@10: select->var_values[VAR_T ] = TS2D(frame->pts) * av_q2d(inlink->time_base); yading@10: select->var_values[VAR_POS] = av_frame_get_pkt_pos(frame) == -1 ? NAN : av_frame_get_pkt_pos(frame); yading@10: yading@10: switch (inlink->type) { yading@10: case AVMEDIA_TYPE_AUDIO: yading@10: select->var_values[VAR_SAMPLES_N] = frame->nb_samples; yading@10: break; yading@10: yading@10: case AVMEDIA_TYPE_VIDEO: yading@10: select->var_values[VAR_INTERLACE_TYPE] = yading@10: !frame->interlaced_frame ? INTERLACE_TYPE_P : yading@10: frame->top_field_first ? INTERLACE_TYPE_T : INTERLACE_TYPE_B; yading@10: select->var_values[VAR_PICT_TYPE] = frame->pict_type; yading@10: #if CONFIG_AVCODEC yading@10: if (select->do_scene_detect) { yading@10: char buf[32]; yading@10: select->var_values[VAR_SCENE] = get_scene_score(ctx, frame); yading@10: // TODO: document metadata yading@10: snprintf(buf, sizeof(buf), "%f", select->var_values[VAR_SCENE]); yading@10: av_dict_set(avpriv_frame_get_metadatap(frame), "lavfi.scene_score", buf, 0); yading@10: } yading@10: #endif yading@10: break; yading@10: } yading@10: yading@10: select->select = res = av_expr_eval(select->expr, select->var_values, NULL); yading@10: av_log(inlink->dst, AV_LOG_DEBUG, yading@10: "n:%f pts:%f t:%f key:%d", yading@10: select->var_values[VAR_N], yading@10: select->var_values[VAR_PTS], yading@10: select->var_values[VAR_T], yading@10: (int)select->var_values[VAR_KEY]); yading@10: yading@10: switch (inlink->type) { yading@10: case AVMEDIA_TYPE_VIDEO: yading@10: av_log(inlink->dst, AV_LOG_DEBUG, " interlace_type:%c pict_type:%c scene:%f", yading@10: select->var_values[VAR_INTERLACE_TYPE] == INTERLACE_TYPE_P ? 'P' : yading@10: select->var_values[VAR_INTERLACE_TYPE] == INTERLACE_TYPE_T ? 'T' : yading@10: select->var_values[VAR_INTERLACE_TYPE] == INTERLACE_TYPE_B ? 'B' : '?', yading@10: av_get_picture_type_char(select->var_values[VAR_PICT_TYPE]), yading@10: select->var_values[VAR_SCENE]); yading@10: break; yading@10: case AVMEDIA_TYPE_AUDIO: yading@10: av_log(inlink->dst, AV_LOG_DEBUG, " samples_n:%d consumed_samples_n:%d", yading@10: (int)select->var_values[VAR_SAMPLES_N], yading@10: (int)select->var_values[VAR_CONSUMED_SAMPLES_N]); yading@10: break; yading@10: } yading@10: yading@10: if (res == 0) { yading@10: select->select_out = -1; /* drop */ yading@10: } else if (isnan(res) || res < 0) { yading@10: select->select_out = 0; /* first output */ yading@10: } else { yading@10: select->select_out = FFMIN(ceilf(res)-1, select->nb_outputs-1); /* other outputs */ yading@10: } yading@10: yading@10: av_log(inlink->dst, AV_LOG_DEBUG, " -> select:%f select_out:%d\n", res, select->select_out); yading@10: yading@10: if (res) { yading@10: select->var_values[VAR_PREV_SELECTED_N] = select->var_values[VAR_N]; yading@10: select->var_values[VAR_PREV_SELECTED_PTS] = select->var_values[VAR_PTS]; yading@10: select->var_values[VAR_PREV_SELECTED_T] = select->var_values[VAR_T]; yading@10: select->var_values[VAR_SELECTED_N] += 1.0; yading@10: if (inlink->type == AVMEDIA_TYPE_AUDIO) yading@10: select->var_values[VAR_CONSUMED_SAMPLES_N] += frame->nb_samples; yading@10: } yading@10: yading@10: select->var_values[VAR_N] += 1.0; yading@10: select->var_values[VAR_PREV_PTS] = select->var_values[VAR_PTS]; yading@10: select->var_values[VAR_PREV_T] = select->var_values[VAR_T]; yading@10: } yading@10: yading@10: static int filter_frame(AVFilterLink *inlink, AVFrame *frame) yading@10: { yading@10: AVFilterContext *ctx = inlink->dst; yading@10: SelectContext *select = ctx->priv; yading@10: yading@10: select_frame(ctx, frame); yading@10: if (select->select) yading@10: return ff_filter_frame(ctx->outputs[select->select_out], frame); yading@10: yading@10: av_frame_free(&frame); yading@10: return 0; yading@10: } yading@10: yading@10: static int request_frame(AVFilterLink *outlink) yading@10: { yading@10: AVFilterContext *ctx = outlink->src; yading@10: SelectContext *select = ctx->priv; yading@10: AVFilterLink *inlink = outlink->src->inputs[0]; yading@10: int out_no = FF_OUTLINK_IDX(outlink); yading@10: yading@10: do { yading@10: int ret = ff_request_frame(inlink); yading@10: if (ret < 0) yading@10: return ret; yading@10: } while (select->select_out != out_no); yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static av_cold void uninit(AVFilterContext *ctx) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: int i; yading@10: yading@10: av_expr_free(select->expr); yading@10: select->expr = NULL; yading@10: yading@10: for (i = 0; i < ctx->nb_outputs; i++) yading@10: av_freep(&ctx->output_pads[i].name); yading@10: yading@10: #if CONFIG_AVCODEC yading@10: if (select->do_scene_detect) { yading@10: av_frame_free(&select->prev_picref); yading@10: if (select->avctx) { yading@10: avcodec_close(select->avctx); yading@10: av_freep(&select->avctx); yading@10: } yading@10: } yading@10: #endif yading@10: } yading@10: yading@10: static int query_formats(AVFilterContext *ctx) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: yading@10: if (!select->do_scene_detect) { yading@10: return ff_default_query_formats(ctx); yading@10: } else { yading@10: static const enum AVPixelFormat pix_fmts[] = { yading@10: AV_PIX_FMT_RGB24, AV_PIX_FMT_BGR24, yading@10: AV_PIX_FMT_NONE yading@10: }; yading@10: ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); yading@10: } yading@10: return 0; yading@10: } yading@10: yading@10: #if CONFIG_ASELECT_FILTER yading@10: yading@10: DEFINE_OPTIONS(aselect, AV_OPT_FLAG_AUDIO_PARAM|AV_OPT_FLAG_FILTERING_PARAM); yading@10: AVFILTER_DEFINE_CLASS(aselect); yading@10: yading@10: static av_cold int aselect_init(AVFilterContext *ctx) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: int ret; yading@10: yading@10: if ((ret = init(ctx)) < 0) yading@10: return ret; yading@10: yading@10: if (select->do_scene_detect) { yading@10: av_log(ctx, AV_LOG_ERROR, "Scene detection is ignored in aselect filter\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static const AVFilterPad avfilter_af_aselect_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_AUDIO, yading@10: .get_audio_buffer = ff_null_get_audio_buffer, yading@10: .config_props = config_input, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_af_aselect = { yading@10: .name = "aselect", yading@10: .description = NULL_IF_CONFIG_SMALL("Select audio frames to pass in output."), yading@10: .init = aselect_init, yading@10: .uninit = uninit, yading@10: .priv_size = sizeof(SelectContext), yading@10: .inputs = avfilter_af_aselect_inputs, yading@10: .priv_class = &aselect_class, yading@10: .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS, yading@10: }; yading@10: #endif /* CONFIG_ASELECT_FILTER */ yading@10: yading@10: #if CONFIG_SELECT_FILTER yading@10: yading@10: DEFINE_OPTIONS(select, AV_OPT_FLAG_VIDEO_PARAM|AV_OPT_FLAG_FILTERING_PARAM); yading@10: AVFILTER_DEFINE_CLASS(select); yading@10: yading@10: static av_cold int select_init(AVFilterContext *ctx) yading@10: { yading@10: SelectContext *select = ctx->priv; yading@10: int ret; yading@10: yading@10: if ((ret = init(ctx)) < 0) yading@10: return ret; yading@10: yading@10: if (select->do_scene_detect && !CONFIG_AVCODEC) { yading@10: av_log(ctx, AV_LOG_ERROR, "Scene detection is not available without libavcodec.\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static const AVFilterPad avfilter_vf_select_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .get_video_buffer = ff_null_get_video_buffer, yading@10: .config_props = config_input, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_vf_select = { yading@10: .name = "select", yading@10: .description = NULL_IF_CONFIG_SMALL("Select video frames to pass in output."), yading@10: .init = select_init, yading@10: .uninit = uninit, yading@10: .query_formats = query_formats, yading@10: yading@10: .priv_size = sizeof(SelectContext), yading@10: .priv_class = &select_class, yading@10: yading@10: .inputs = avfilter_vf_select_inputs, yading@10: .flags = AVFILTER_FLAG_DYNAMIC_OUTPUTS, yading@10: }; yading@10: #endif /* CONFIG_SELECT_FILTER */