yading@10: /* yading@10: * Copyright (c) 2013 Paul B Mahol yading@10: * yading@10: * This file is part of FFmpeg. yading@10: * yading@10: * FFmpeg is free software; you can redistribute it and/or yading@10: * modify it under the terms of the GNU Lesser General Public yading@10: * License as published by the Free Software Foundation; either yading@10: * version 2.1 of the License, or (at your option) any later version. yading@10: * yading@10: * FFmpeg is distributed in the hope that it will be useful, yading@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of yading@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU yading@10: * Lesser General Public License for more details. yading@10: * yading@10: * You should have received a copy of the GNU Lesser General Public yading@10: * License along with FFmpeg; if not, write to the Free Software yading@10: * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA yading@10: */ yading@10: yading@10: #include "libavutil/imgutils.h" yading@10: #include "libavutil/eval.h" yading@10: #include "libavutil/opt.h" yading@10: #include "libavutil/pixfmt.h" yading@10: #include "avfilter.h" yading@10: #include "bufferqueue.h" yading@10: #include "formats.h" yading@10: #include "internal.h" yading@10: #include "video.h" yading@10: yading@10: #define TOP 0 yading@10: #define BOTTOM 1 yading@10: yading@10: enum BlendMode { yading@10: BLEND_UNSET = -1, yading@10: BLEND_NORMAL, yading@10: BLEND_ADDITION, yading@10: BLEND_AND, yading@10: BLEND_AVERAGE, yading@10: BLEND_BURN, yading@10: BLEND_DARKEN, yading@10: BLEND_DIFFERENCE, yading@10: BLEND_DIVIDE, yading@10: BLEND_DODGE, yading@10: BLEND_EXCLUSION, yading@10: BLEND_HARDLIGHT, yading@10: BLEND_LIGHTEN, yading@10: BLEND_MULTIPLY, yading@10: BLEND_NEGATION, yading@10: BLEND_OR, yading@10: BLEND_OVERLAY, yading@10: BLEND_PHOENIX, yading@10: BLEND_PINLIGHT, yading@10: BLEND_REFLECT, yading@10: BLEND_SCREEN, yading@10: BLEND_SOFTLIGHT, yading@10: BLEND_SUBTRACT, yading@10: BLEND_VIVIDLIGHT, yading@10: BLEND_XOR, yading@10: BLEND_NB yading@10: }; yading@10: yading@10: static const char *const var_names[] = { "X", "Y", "W", "H", "SW", "SH", "T", "N", "A", "B", "TOP", "BOTTOM", NULL }; yading@10: enum { VAR_X, VAR_Y, VAR_W, VAR_H, VAR_SW, VAR_SH, VAR_T, VAR_N, VAR_A, VAR_B, VAR_TOP, VAR_BOTTOM, VAR_VARS_NB }; yading@10: yading@10: typedef struct FilterParams { yading@10: enum BlendMode mode; yading@10: double values[VAR_VARS_NB]; yading@10: double opacity; yading@10: AVExpr *e; yading@10: char *expr_str; yading@10: void (*blend)(const uint8_t *top, int top_linesize, yading@10: const uint8_t *bottom, int bottom_linesize, yading@10: uint8_t *dst, int dst_linesize, yading@10: int width, int height, struct FilterParams *param); yading@10: } FilterParams; yading@10: yading@10: typedef struct { yading@10: const AVClass *class; yading@10: struct FFBufQueue queue_top; yading@10: struct FFBufQueue queue_bottom; yading@10: int hsub, vsub; ///< chroma subsampling values yading@10: int frame_requested; yading@10: int framenum; yading@10: char *all_expr; yading@10: enum BlendMode all_mode; yading@10: double all_opacity; yading@10: yading@10: FilterParams params[4]; yading@10: } BlendContext; yading@10: yading@10: #define OFFSET(x) offsetof(BlendContext, x) yading@10: #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM yading@10: yading@10: static const AVOption blend_options[] = { yading@10: { "c0_mode", "set component #0 blend mode", OFFSET(params[0].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"}, yading@10: { "c1_mode", "set component #1 blend mode", OFFSET(params[1].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"}, yading@10: { "c2_mode", "set component #2 blend mode", OFFSET(params[2].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"}, yading@10: { "c3_mode", "set component #3 blend mode", OFFSET(params[3].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"}, yading@10: { "all_mode", "set blend mode for all components", OFFSET(all_mode), AV_OPT_TYPE_INT, {.i64=-1},-1, BLEND_NB-1, FLAGS, "mode"}, yading@10: { "addition", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_ADDITION}, 0, 0, FLAGS, "mode" }, yading@10: { "and", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AND}, 0, 0, FLAGS, "mode" }, yading@10: { "average", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AVERAGE}, 0, 0, FLAGS, "mode" }, yading@10: { "burn", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_BURN}, 0, 0, FLAGS, "mode" }, yading@10: { "darken", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DARKEN}, 0, 0, FLAGS, "mode" }, yading@10: { "difference", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIFFERENCE}, 0, 0, FLAGS, "mode" }, yading@10: { "divide", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIVIDE}, 0, 0, FLAGS, "mode" }, yading@10: { "dodge", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DODGE}, 0, 0, FLAGS, "mode" }, yading@10: { "exclusion", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_EXCLUSION}, 0, 0, FLAGS, "mode" }, yading@10: { "hardlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_HARDLIGHT}, 0, 0, FLAGS, "mode" }, yading@10: { "lighten", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_LIGHTEN}, 0, 0, FLAGS, "mode" }, yading@10: { "multiply", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_MULTIPLY}, 0, 0, FLAGS, "mode" }, yading@10: { "negation", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NEGATION}, 0, 0, FLAGS, "mode" }, yading@10: { "normal", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NORMAL}, 0, 0, FLAGS, "mode" }, yading@10: { "or", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OR}, 0, 0, FLAGS, "mode" }, yading@10: { "overlay", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OVERLAY}, 0, 0, FLAGS, "mode" }, yading@10: { "phoenix", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PHOENIX}, 0, 0, FLAGS, "mode" }, yading@10: { "pinlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PINLIGHT}, 0, 0, FLAGS, "mode" }, yading@10: { "reflect", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_REFLECT}, 0, 0, FLAGS, "mode" }, yading@10: { "screen", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SCREEN}, 0, 0, FLAGS, "mode" }, yading@10: { "softlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SOFTLIGHT}, 0, 0, FLAGS, "mode" }, yading@10: { "subtract", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SUBTRACT}, 0, 0, FLAGS, "mode" }, yading@10: { "vividlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_VIVIDLIGHT}, 0, 0, FLAGS, "mode" }, yading@10: { "xor", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_XOR}, 0, 0, FLAGS, "mode" }, yading@10: { "c0_expr", "set color component #0 expression", OFFSET(params[0].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS }, yading@10: { "c1_expr", "set color component #1 expression", OFFSET(params[1].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS }, yading@10: { "c2_expr", "set color component #2 expression", OFFSET(params[2].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS }, yading@10: { "c3_expr", "set color component #3 expression", OFFSET(params[3].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS }, yading@10: { "all_expr", "set expression for all color components", OFFSET(all_expr), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS }, yading@10: { "c0_opacity", "set color component #0 opacity", OFFSET(params[0].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS }, yading@10: { "c1_opacity", "set color component #1 opacity", OFFSET(params[1].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS }, yading@10: { "c2_opacity", "set color component #2 opacity", OFFSET(params[2].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS }, yading@10: { "c3_opacity", "set color component #3 opacity", OFFSET(params[3].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS }, yading@10: { "all_opacity", "set opacity for all color components", OFFSET(all_opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS}, yading@10: { NULL }, yading@10: }; yading@10: yading@10: AVFILTER_DEFINE_CLASS(blend); yading@10: yading@10: static void blend_normal(const uint8_t *top, int top_linesize, yading@10: const uint8_t *bottom, int bottom_linesize, yading@10: uint8_t *dst, int dst_linesize, yading@10: int width, int height, FilterParams *param) yading@10: { yading@10: av_image_copy_plane(dst, dst_linesize, top, top_linesize, width, height); yading@10: } yading@10: yading@10: #define DEFINE_BLEND(name, expr) \ yading@10: static void blend_## name(const uint8_t *top, int top_linesize, \ yading@10: const uint8_t *bottom, int bottom_linesize, \ yading@10: uint8_t *dst, int dst_linesize, \ yading@10: int width, int height, FilterParams *param) \ yading@10: { \ yading@10: double opacity = param->opacity; \ yading@10: int i, j; \ yading@10: \ yading@10: for (i = 0; i < height; i++) { \ yading@10: for (j = 0; j < width; j++) { \ yading@10: dst[j] = top[j] + ((expr) - top[j]) * opacity; \ yading@10: } \ yading@10: dst += dst_linesize; \ yading@10: top += top_linesize; \ yading@10: bottom += bottom_linesize; \ yading@10: } \ yading@10: } yading@10: yading@10: #define A top[j] yading@10: #define B bottom[j] yading@10: yading@10: #define MULTIPLY(x, a, b) (x * ((a * b) / 255)) yading@10: #define SCREEN(x, a, b) (255 - x * ((255 - a) * (255 - b) / 255)) yading@10: #define BURN(a, b) ((a == 0) ? a : FFMAX(0, 255 - ((255 - b) << 8) / a)) yading@10: #define DODGE(a, b) ((a == 255) ? a : FFMIN(255, ((b << 8) / (255 - a)))) yading@10: yading@10: DEFINE_BLEND(addition, FFMIN(255, A + B)) yading@10: DEFINE_BLEND(average, (A + B) / 2) yading@10: DEFINE_BLEND(subtract, FFMAX(0, A - B)) yading@10: DEFINE_BLEND(multiply, MULTIPLY(1, A, B)) yading@10: DEFINE_BLEND(negation, 255 - FFABS(255 - A - B)) yading@10: DEFINE_BLEND(difference, FFABS(A - B)) yading@10: DEFINE_BLEND(screen, SCREEN(1, A, B)) yading@10: DEFINE_BLEND(overlay, (A < 128) ? MULTIPLY(2, A, B) : SCREEN(2, A, B)) yading@10: DEFINE_BLEND(hardlight, (B < 128) ? MULTIPLY(2, B, A) : SCREEN(2, B, A)) yading@10: DEFINE_BLEND(darken, FFMIN(A, B)) yading@10: DEFINE_BLEND(lighten, FFMAX(A, B)) yading@10: DEFINE_BLEND(divide, ((float)A / ((float)B) * 255)) yading@10: DEFINE_BLEND(dodge, DODGE(A, B)) yading@10: DEFINE_BLEND(burn, BURN(A, B)) yading@10: DEFINE_BLEND(softlight, (A > 127) ? B + (255 - B) * (A - 127.5) / 127.5 * (0.5 - FFABS(B - 127.5) / 255): B - B * ((127.5 - A) / 127.5) * (0.5 - FFABS(B - 127.5)/255)) yading@10: DEFINE_BLEND(exclusion, A + B - 2 * A * B / 255) yading@10: DEFINE_BLEND(pinlight, (B < 128) ? FFMIN(A, 2 * B) : FFMAX(A, 2 * (B - 128))) yading@10: DEFINE_BLEND(phoenix, FFMIN(A, B) - FFMAX(A, B) + 255) yading@10: DEFINE_BLEND(reflect, (B == 255) ? B : FFMIN(255, (A * A / (255 - B)))) yading@10: DEFINE_BLEND(and, A & B) yading@10: DEFINE_BLEND(or, A | B) yading@10: DEFINE_BLEND(xor, A ^ B) yading@10: DEFINE_BLEND(vividlight, (B < 128) ? BURN(A, 2 * B) : DODGE(A, 2 * (B - 128))) yading@10: yading@10: static void blend_expr(const uint8_t *top, int top_linesize, yading@10: const uint8_t *bottom, int bottom_linesize, yading@10: uint8_t *dst, int dst_linesize, yading@10: int width, int height, yading@10: FilterParams *param) yading@10: { yading@10: AVExpr *e = param->e; yading@10: double *values = param->values; yading@10: int y, x; yading@10: yading@10: for (y = 0; y < height; y++) { yading@10: values[VAR_Y] = y; yading@10: for (x = 0; x < width; x++) { yading@10: values[VAR_X] = x; yading@10: values[VAR_TOP] = values[VAR_A] = top[x]; yading@10: values[VAR_BOTTOM] = values[VAR_B] = bottom[x]; yading@10: dst[x] = av_expr_eval(e, values, NULL); yading@10: } yading@10: dst += dst_linesize; yading@10: top += top_linesize; yading@10: bottom += bottom_linesize; yading@10: } yading@10: } yading@10: yading@10: static av_cold int init(AVFilterContext *ctx) yading@10: { yading@10: BlendContext *b = ctx->priv; yading@10: int ret, plane; yading@10: yading@10: for (plane = 0; plane < FF_ARRAY_ELEMS(b->params); plane++) { yading@10: FilterParams *param = &b->params[plane]; yading@10: yading@10: if (b->all_mode >= 0) yading@10: param->mode = b->all_mode; yading@10: if (b->all_opacity < 1) yading@10: param->opacity = b->all_opacity; yading@10: yading@10: switch (param->mode) { yading@10: case BLEND_ADDITION: param->blend = blend_addition; break; yading@10: case BLEND_AND: param->blend = blend_and; break; yading@10: case BLEND_AVERAGE: param->blend = blend_average; break; yading@10: case BLEND_BURN: param->blend = blend_burn; break; yading@10: case BLEND_DARKEN: param->blend = blend_darken; break; yading@10: case BLEND_DIFFERENCE: param->blend = blend_difference; break; yading@10: case BLEND_DIVIDE: param->blend = blend_divide; break; yading@10: case BLEND_DODGE: param->blend = blend_dodge; break; yading@10: case BLEND_EXCLUSION: param->blend = blend_exclusion; break; yading@10: case BLEND_HARDLIGHT: param->blend = blend_hardlight; break; yading@10: case BLEND_LIGHTEN: param->blend = blend_lighten; break; yading@10: case BLEND_MULTIPLY: param->blend = blend_multiply; break; yading@10: case BLEND_NEGATION: param->blend = blend_negation; break; yading@10: case BLEND_NORMAL: param->blend = blend_normal; break; yading@10: case BLEND_OR: param->blend = blend_or; break; yading@10: case BLEND_OVERLAY: param->blend = blend_overlay; break; yading@10: case BLEND_PHOENIX: param->blend = blend_phoenix; break; yading@10: case BLEND_PINLIGHT: param->blend = blend_pinlight; break; yading@10: case BLEND_REFLECT: param->blend = blend_reflect; break; yading@10: case BLEND_SCREEN: param->blend = blend_screen; break; yading@10: case BLEND_SOFTLIGHT: param->blend = blend_softlight; break; yading@10: case BLEND_SUBTRACT: param->blend = blend_subtract; break; yading@10: case BLEND_VIVIDLIGHT: param->blend = blend_vividlight; break; yading@10: case BLEND_XOR: param->blend = blend_xor; break; yading@10: } yading@10: yading@10: if (b->all_expr && !param->expr_str) { yading@10: param->expr_str = av_strdup(b->all_expr); yading@10: if (!param->expr_str) yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: if (param->expr_str) { yading@10: ret = av_expr_parse(¶m->e, param->expr_str, var_names, yading@10: NULL, NULL, NULL, NULL, 0, ctx); yading@10: if (ret < 0) yading@10: return ret; yading@10: param->blend = blend_expr; yading@10: } yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static int query_formats(AVFilterContext *ctx) yading@10: { yading@10: static const enum AVPixelFormat pix_fmts[] = { yading@10: AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA420P, yading@10: AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P,AV_PIX_FMT_YUVJ420P, yading@10: AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV410P, yading@10: AV_PIX_FMT_GBRP, AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE yading@10: }; yading@10: yading@10: ff_set_common_formats(ctx, ff_make_format_list(pix_fmts)); yading@10: return 0; yading@10: } yading@10: yading@10: static int config_output(AVFilterLink *outlink) yading@10: { yading@10: AVFilterContext *ctx = outlink->src; yading@10: AVFilterLink *toplink = ctx->inputs[TOP]; yading@10: AVFilterLink *bottomlink = ctx->inputs[BOTTOM]; yading@10: yading@10: if (toplink->format != bottomlink->format) { yading@10: av_log(ctx, AV_LOG_ERROR, "inputs must be of same pixel format\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: if (toplink->w != bottomlink->w || yading@10: toplink->h != bottomlink->h || yading@10: toplink->sample_aspect_ratio.num != bottomlink->sample_aspect_ratio.num || yading@10: toplink->sample_aspect_ratio.den != bottomlink->sample_aspect_ratio.den) { yading@10: av_log(ctx, AV_LOG_ERROR, "First input link %s parameters " yading@10: "(size %dx%d, SAR %d:%d) do not match the corresponding " yading@10: "second input link %s parameters (%dx%d, SAR %d:%d)\n", yading@10: ctx->input_pads[TOP].name, toplink->w, toplink->h, yading@10: toplink->sample_aspect_ratio.num, yading@10: toplink->sample_aspect_ratio.den, yading@10: ctx->input_pads[BOTTOM].name, bottomlink->w, bottomlink->h, yading@10: bottomlink->sample_aspect_ratio.num, yading@10: bottomlink->sample_aspect_ratio.den); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: outlink->w = toplink->w; yading@10: outlink->h = bottomlink->h; yading@10: outlink->time_base = toplink->time_base; yading@10: outlink->sample_aspect_ratio = toplink->sample_aspect_ratio; yading@10: outlink->frame_rate = toplink->frame_rate; yading@10: return 0; yading@10: } yading@10: yading@10: static int config_input_top(AVFilterLink *inlink) yading@10: { yading@10: BlendContext *b = inlink->dst->priv; yading@10: const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(inlink->format); yading@10: yading@10: b->hsub = pix_desc->log2_chroma_w; yading@10: b->vsub = pix_desc->log2_chroma_h; yading@10: return 0; yading@10: } yading@10: yading@10: static av_cold void uninit(AVFilterContext *ctx) yading@10: { yading@10: BlendContext *b = ctx->priv; yading@10: int i; yading@10: yading@10: ff_bufqueue_discard_all(&b->queue_top); yading@10: ff_bufqueue_discard_all(&b->queue_bottom); yading@10: yading@10: for (i = 0; i < FF_ARRAY_ELEMS(b->params); i++) yading@10: av_expr_free(b->params[i].e); yading@10: } yading@10: yading@10: static int request_frame(AVFilterLink *outlink) yading@10: { yading@10: AVFilterContext *ctx = outlink->src; yading@10: BlendContext *b = ctx->priv; yading@10: int in, ret; yading@10: yading@10: b->frame_requested = 1; yading@10: while (b->frame_requested) { yading@10: in = ff_bufqueue_peek(&b->queue_top, 0) ? BOTTOM : TOP; yading@10: ret = ff_request_frame(ctx->inputs[in]); yading@10: if (ret < 0) yading@10: return ret; yading@10: } yading@10: return 0; yading@10: } yading@10: yading@10: static void blend_frame(AVFilterContext *ctx, yading@10: AVFrame *top_buf, yading@10: AVFrame *bottom_buf, yading@10: AVFrame *dst_buf) yading@10: { yading@10: BlendContext *b = ctx->priv; yading@10: AVFilterLink *inlink = ctx->inputs[0]; yading@10: FilterParams *param; yading@10: int plane; yading@10: yading@10: for (plane = 0; dst_buf->data[plane]; plane++) { yading@10: int hsub = plane == 1 || plane == 2 ? b->hsub : 0; yading@10: int vsub = plane == 1 || plane == 2 ? b->vsub : 0; yading@10: int outw = dst_buf->width >> hsub; yading@10: int outh = dst_buf->height >> vsub; yading@10: uint8_t *dst = dst_buf->data[plane]; yading@10: uint8_t *top = top_buf->data[plane]; yading@10: uint8_t *bottom = bottom_buf->data[plane]; yading@10: yading@10: param = &b->params[plane]; yading@10: param->values[VAR_N] = b->framenum++; yading@10: param->values[VAR_T] = dst_buf->pts == AV_NOPTS_VALUE ? NAN : dst_buf->pts * av_q2d(inlink->time_base); yading@10: param->values[VAR_W] = outw; yading@10: param->values[VAR_H] = outh; yading@10: param->values[VAR_SW] = outw / dst_buf->width; yading@10: param->values[VAR_SH] = outh / dst_buf->height; yading@10: param->blend(top, top_buf->linesize[plane], yading@10: bottom, bottom_buf->linesize[plane], yading@10: dst, dst_buf->linesize[plane], outw, outh, param); yading@10: } yading@10: } yading@10: yading@10: static int filter_frame(AVFilterLink *inlink, AVFrame *buf) yading@10: { yading@10: AVFilterContext *ctx = inlink->dst; yading@10: AVFilterLink *outlink = ctx->outputs[0]; yading@10: BlendContext *b = ctx->priv; yading@10: yading@10: int ret = 0; yading@10: int is_bottom = (inlink == ctx->inputs[BOTTOM]); yading@10: struct FFBufQueue *queue = yading@10: (is_bottom ? &b->queue_bottom : &b->queue_top); yading@10: ff_bufqueue_add(ctx, queue, buf); yading@10: yading@10: while (1) { yading@10: AVFrame *top_buf, *bottom_buf, *out_buf; yading@10: yading@10: if (!ff_bufqueue_peek(&b->queue_top, 0) || yading@10: !ff_bufqueue_peek(&b->queue_bottom, 0)) break; yading@10: yading@10: top_buf = ff_bufqueue_get(&b->queue_top); yading@10: bottom_buf = ff_bufqueue_get(&b->queue_bottom); yading@10: yading@10: out_buf = ff_get_video_buffer(outlink, outlink->w, outlink->h); yading@10: if (!out_buf) { yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: av_frame_copy_props(out_buf, top_buf); yading@10: yading@10: b->frame_requested = 0; yading@10: blend_frame(ctx, top_buf, bottom_buf, out_buf); yading@10: ret = ff_filter_frame(ctx->outputs[0], out_buf); yading@10: av_frame_free(&top_buf); yading@10: av_frame_free(&bottom_buf); yading@10: } yading@10: return ret; yading@10: } yading@10: yading@10: static const AVFilterPad blend_inputs[] = { yading@10: { yading@10: .name = "top", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .config_props = config_input_top, yading@10: .filter_frame = filter_frame, yading@10: },{ yading@10: .name = "bottom", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: static const AVFilterPad blend_outputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .config_props = config_output, yading@10: .request_frame = request_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_vf_blend = { yading@10: .name = "blend", yading@10: .description = NULL_IF_CONFIG_SMALL("Blend two video frames into each other."), yading@10: .init = init, yading@10: .uninit = uninit, yading@10: .priv_size = sizeof(BlendContext), yading@10: .query_formats = query_formats, yading@10: .inputs = blend_inputs, yading@10: .outputs = blend_outputs, yading@10: .priv_class = &blend_class, yading@10: };