yading@10: /* yading@10: * Copyright (c) 2003 Michael Zucchi yading@10: * Copyright (c) 2010 Baptiste Coudurier yading@10: * Copyright (c) 2011 Stefano Sabatini yading@10: * Copyright (c) 2013 Vittorio Giovara yading@10: * yading@10: * This file is part of FFmpeg. yading@10: * yading@10: * FFmpeg is free software; you can redistribute it and/or modify yading@10: * it under the terms of the GNU General Public License as published by yading@10: * the Free Software Foundation; either version 2 of the License, or yading@10: * (at your option) any later version. yading@10: * yading@10: * FFmpeg is distributed in the hope that it will be useful, yading@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of yading@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the yading@10: * GNU General Public License for more details. yading@10: * yading@10: * You should have received a copy of the GNU General Public License along yading@10: * with FFmpeg; if not, write to the Free Software Foundation, Inc., yading@10: * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. yading@10: */ yading@10: yading@10: /** yading@10: * @file yading@10: * progressive to interlaced content filter, inspired by heavy debugging of tinterlace filter yading@10: */ yading@10: yading@10: #include "libavutil/common.h" yading@10: #include "libavutil/opt.h" yading@10: #include "libavutil/imgutils.h" yading@10: #include "libavutil/avassert.h" yading@10: yading@10: #include "formats.h" yading@10: #include "avfilter.h" yading@10: #include "internal.h" yading@10: #include "video.h" yading@10: yading@10: enum ScanMode { yading@10: MODE_TFF = 0, yading@10: MODE_BFF = 1, yading@10: }; yading@10: yading@10: enum FieldType { yading@10: FIELD_UPPER = 0, yading@10: FIELD_LOWER = 1, yading@10: }; yading@10: yading@10: typedef struct { yading@10: const AVClass *class; yading@10: enum ScanMode scan; // top or bottom field first scanning yading@10: int lowpass; // enable or disable low pass filterning yading@10: AVFrame *cur, *next; // the two frames from which the new one is obtained yading@10: int got_output; // signal an output frame is reday to request_frame() yading@10: } InterlaceContext; yading@10: yading@10: #define OFFSET(x) offsetof(InterlaceContext, x) yading@10: #define V AV_OPT_FLAG_VIDEO_PARAM yading@10: static const AVOption options[] = { yading@10: { "scan", "scanning mode", OFFSET(scan), yading@10: AV_OPT_TYPE_INT, {.i64 = MODE_TFF }, 0, 1, .flags = V, .unit = "scan" }, yading@10: { "tff", "top field first", 0, yading@10: AV_OPT_TYPE_CONST, {.i64 = MODE_TFF }, INT_MIN, INT_MAX, .flags = V, .unit = "scan" }, yading@10: { "bff", "bottom field first", 0, yading@10: AV_OPT_TYPE_CONST, {.i64 = MODE_BFF }, INT_MIN, INT_MAX, .flags = V, .unit = "scan" }, yading@10: { "lowpass", "enable vertical low-pass filter", OFFSET(lowpass), yading@10: AV_OPT_TYPE_INT, {.i64 = 1 }, 0, 1, .flags = V }, yading@10: { NULL } yading@10: }; yading@10: yading@10: static const AVClass class = { yading@10: .class_name = "interlace filter", yading@10: .item_name = av_default_item_name, yading@10: .option = options, yading@10: .version = LIBAVUTIL_VERSION_INT, yading@10: }; yading@10: yading@10: yading@10: static const enum AVPixelFormat formats_supported[] = { yading@10: AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, yading@10: AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV410P, AV_PIX_FMT_YUVA420P, yading@10: AV_PIX_FMT_GRAY8, AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, yading@10: AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_NONE yading@10: }; yading@10: yading@10: static int query_formats(AVFilterContext *ctx) yading@10: { yading@10: ff_set_common_formats(ctx, ff_make_format_list(formats_supported)); yading@10: return 0; yading@10: } yading@10: yading@10: static av_cold void uninit(AVFilterContext *ctx) yading@10: { yading@10: InterlaceContext *s = ctx->priv; yading@10: yading@10: av_frame_free(&s->cur); yading@10: av_frame_free(&s->next); yading@10: } yading@10: yading@10: static int config_out_props(AVFilterLink *outlink) yading@10: { yading@10: AVFilterContext *ctx = outlink->src; yading@10: AVFilterLink *inlink = outlink->src->inputs[0]; yading@10: InterlaceContext *s = ctx->priv; yading@10: yading@10: if (inlink->h < 2) { yading@10: av_log(ctx, AV_LOG_ERROR, "input video height is too small\n"); yading@10: return AVERROR_INVALIDDATA; yading@10: } yading@10: // same input size yading@10: outlink->w = inlink->w; yading@10: outlink->h = inlink->h; yading@10: outlink->time_base = inlink->time_base; yading@10: outlink->frame_rate = inlink->frame_rate; yading@10: // half framerate yading@10: outlink->time_base.num *= 2; yading@10: outlink->frame_rate.den *= 2; yading@10: yading@10: av_log(ctx, AV_LOG_VERBOSE, "%s interlacing %s lowpass filter\n", yading@10: s->scan == MODE_TFF ? "tff" : "bff", (s->lowpass) ? "with" : "without"); yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static void copy_picture_field(AVFrame *src_frame, AVFrame *dst_frame, yading@10: AVFilterLink *inlink, enum FieldType field_type, yading@10: int lowpass) yading@10: { yading@10: const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(inlink->format); yading@10: int vsub = desc->log2_chroma_h; yading@10: int plane, i, j; yading@10: yading@10: for (plane = 0; plane < desc->nb_components; plane++) { yading@10: int lines = (plane == 1 || plane == 2) ? inlink->h >> vsub : inlink->h; yading@10: int linesize = av_image_get_linesize(inlink->format, inlink->w, plane); yading@10: uint8_t *dstp = dst_frame->data[plane]; yading@10: const uint8_t *srcp = src_frame->data[plane]; yading@10: yading@10: av_assert0(linesize >= 0); yading@10: yading@10: lines /= 2; yading@10: if (field_type == FIELD_LOWER) yading@10: srcp += src_frame->linesize[plane]; yading@10: if (field_type == FIELD_LOWER) yading@10: dstp += dst_frame->linesize[plane]; yading@10: if (lowpass) { yading@10: int srcp_linesize = src_frame->linesize[plane] * 2; yading@10: int dstp_linesize = dst_frame->linesize[plane] * 2; yading@10: for (j = lines; j > 0; j--) { yading@10: const uint8_t *srcp_above = srcp - src_frame->linesize[plane]; yading@10: const uint8_t *srcp_below = srcp + src_frame->linesize[plane]; yading@10: if (j == lines) yading@10: srcp_above = srcp; // there is no line above yading@10: if (j == 1) yading@10: srcp_below = srcp; // there is no line below yading@10: for (i = 0; i < linesize; i++) { yading@10: // this calculation is an integer representation of yading@10: // '0.5 * current + 0.25 * above + 0.25 + below' yading@10: // '1 +' is for rounding. yading@10: dstp[i] = (1 + srcp[i] + srcp[i] + srcp_above[i] + srcp_below[i]) >> 2; yading@10: } yading@10: dstp += dstp_linesize; yading@10: srcp += srcp_linesize; yading@10: } yading@10: } else { yading@10: av_image_copy_plane(dstp, dst_frame->linesize[plane] * 2, yading@10: srcp, src_frame->linesize[plane] * 2, yading@10: linesize, lines); yading@10: } yading@10: } yading@10: } yading@10: yading@10: static int filter_frame(AVFilterLink *inlink, AVFrame *buf) yading@10: { yading@10: AVFilterContext *ctx = inlink->dst; yading@10: AVFilterLink *outlink = ctx->outputs[0]; yading@10: InterlaceContext *s = ctx->priv; yading@10: AVFrame *out; yading@10: int tff, ret; yading@10: yading@10: av_frame_free(&s->cur); yading@10: s->cur = s->next; yading@10: s->next = buf; yading@10: yading@10: /* we need at least two frames */ yading@10: if (!s->cur || !s->next) yading@10: return 0; yading@10: yading@10: tff = (s->scan == MODE_TFF); yading@10: out = ff_get_video_buffer(outlink, outlink->w, outlink->h); yading@10: if (!out) yading@10: return AVERROR(ENOMEM); yading@10: yading@10: av_frame_copy_props(out, s->cur); yading@10: out->interlaced_frame = 1; yading@10: out->top_field_first = tff; yading@10: out->pts /= 2; // adjust pts to new framerate yading@10: yading@10: /* copy upper/lower field from cur */ yading@10: copy_picture_field(s->cur, out, inlink, tff ? FIELD_UPPER : FIELD_LOWER, s->lowpass); yading@10: av_frame_free(&s->cur); yading@10: yading@10: /* copy lower/upper field from next */ yading@10: copy_picture_field(s->next, out, inlink, tff ? FIELD_LOWER : FIELD_UPPER, s->lowpass); yading@10: av_frame_free(&s->next); yading@10: yading@10: ret = ff_filter_frame(outlink, out); yading@10: s->got_output = 1; yading@10: yading@10: return ret; yading@10: } yading@10: yading@10: static int request_frame(AVFilterLink *outlink) yading@10: { yading@10: AVFilterContext *ctx = outlink->src; yading@10: InterlaceContext *s = ctx->priv; yading@10: int ret = 0; yading@10: yading@10: s->got_output = 0; yading@10: while (ret >= 0 && !s->got_output) yading@10: ret = ff_request_frame(ctx->inputs[0]); yading@10: yading@10: return ret; yading@10: } yading@10: yading@10: static const AVFilterPad inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: static const AVFilterPad outputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .config_props = config_out_props, yading@10: .request_frame = request_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_vf_interlace = { yading@10: .name = "interlace", yading@10: .description = NULL_IF_CONFIG_SMALL("Convert progressive video into interlaced."), yading@10: .uninit = uninit, yading@10: yading@10: .priv_class = &class, yading@10: .priv_size = sizeof(InterlaceContext), yading@10: .query_formats = query_formats, yading@10: yading@10: .inputs = inputs, yading@10: .outputs = outputs, yading@10: }; yading@10: