yading@10: /* yading@10: * Copyright (c) 2011 Stefano Sabatini yading@10: * yading@10: * This file is part of FFmpeg. yading@10: * yading@10: * FFmpeg is free software; you can redistribute it and/or yading@10: * modify it under the terms of the GNU Lesser General Public yading@10: * License as published by the Free Software Foundation; either yading@10: * version 2.1 of the License, or (at your option) any later version. yading@10: * yading@10: * FFmpeg is distributed in the hope that it will be useful, yading@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of yading@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU yading@10: * Lesser General Public License for more details. yading@10: * yading@10: * You should have received a copy of the GNU Lesser General Public yading@10: * License along with FFmpeg; if not, write to the Free Software yading@10: * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA yading@10: */ yading@10: yading@10: /** yading@10: * @file yading@10: * buffer sink yading@10: */ yading@10: yading@10: #include "libavutil/audio_fifo.h" yading@10: #include "libavutil/avassert.h" yading@10: #include "libavutil/channel_layout.h" yading@10: #include "libavutil/common.h" yading@10: #include "libavutil/mathematics.h" yading@10: #include "libavutil/opt.h" yading@10: yading@10: #include "audio.h" yading@10: #include "avfilter.h" yading@10: #include "buffersink.h" yading@10: #include "internal.h" yading@10: yading@10: typedef struct { yading@10: const AVClass *class; yading@10: AVFifoBuffer *fifo; ///< FIFO buffer of video frame references yading@10: unsigned warning_limit; yading@10: yading@10: /* only used for video */ yading@10: enum AVPixelFormat *pixel_fmts; ///< list of accepted pixel formats, must be terminated with -1 yading@10: int pixel_fmts_size; yading@10: yading@10: /* only used for audio */ yading@10: enum AVSampleFormat *sample_fmts; ///< list of accepted sample formats, terminated by AV_SAMPLE_FMT_NONE yading@10: int sample_fmts_size; yading@10: int64_t *channel_layouts; ///< list of accepted channel layouts, terminated by -1 yading@10: int channel_layouts_size; yading@10: int *channel_counts; ///< list of accepted channel counts, terminated by -1 yading@10: int channel_counts_size; yading@10: int all_channel_counts; yading@10: int *sample_rates; ///< list of accepted sample rates, terminated by -1 yading@10: int sample_rates_size; yading@10: yading@10: /* only used for compat API */ yading@10: AVAudioFifo *audio_fifo; ///< FIFO for audio samples yading@10: int64_t next_pts; ///< interpolating audio pts yading@10: } BufferSinkContext; yading@10: yading@10: #define NB_ITEMS(list) (list ## _size / sizeof(*list)) yading@10: yading@10: static av_cold void uninit(AVFilterContext *ctx) yading@10: { yading@10: BufferSinkContext *sink = ctx->priv; yading@10: AVFrame *frame; yading@10: yading@10: if (sink->audio_fifo) yading@10: av_audio_fifo_free(sink->audio_fifo); yading@10: yading@10: if (sink->fifo) { yading@10: while (av_fifo_size(sink->fifo) >= sizeof(AVFilterBufferRef *)) { yading@10: av_fifo_generic_read(sink->fifo, &frame, sizeof(frame), NULL); yading@10: av_frame_free(&frame); yading@10: } yading@10: av_fifo_free(sink->fifo); yading@10: sink->fifo = NULL; yading@10: } yading@10: } yading@10: yading@10: static int add_buffer_ref(AVFilterContext *ctx, AVFrame *ref) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: yading@10: if (av_fifo_space(buf->fifo) < sizeof(AVFilterBufferRef *)) { yading@10: /* realloc fifo size */ yading@10: if (av_fifo_realloc2(buf->fifo, av_fifo_size(buf->fifo) * 2) < 0) { yading@10: av_log(ctx, AV_LOG_ERROR, yading@10: "Cannot buffer more frames. Consume some available frames " yading@10: "before adding new ones.\n"); yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: } yading@10: yading@10: /* cache frame */ yading@10: av_fifo_generic_write(buf->fifo, &ref, sizeof(AVFilterBufferRef *), NULL); yading@10: return 0; yading@10: } yading@10: yading@10: static int filter_frame(AVFilterLink *link, AVFrame *frame) yading@10: { yading@10: AVFilterContext *ctx = link->dst; yading@10: BufferSinkContext *buf = link->dst->priv; yading@10: int ret; yading@10: yading@10: if ((ret = add_buffer_ref(ctx, frame)) < 0) yading@10: return ret; yading@10: if (buf->warning_limit && yading@10: av_fifo_size(buf->fifo) / sizeof(AVFilterBufferRef *) >= buf->warning_limit) { yading@10: av_log(ctx, AV_LOG_WARNING, yading@10: "%d buffers queued in %s, something may be wrong.\n", yading@10: buf->warning_limit, yading@10: (char *)av_x_if_null(ctx->name, ctx->filter->name)); yading@10: buf->warning_limit *= 10; yading@10: } yading@10: return 0; yading@10: } yading@10: yading@10: int av_buffersink_get_frame(AVFilterContext *ctx, AVFrame *frame) yading@10: { yading@10: return av_buffersink_get_frame_flags(ctx, frame, 0); yading@10: } yading@10: yading@10: int attribute_align_arg av_buffersink_get_frame_flags(AVFilterContext *ctx, AVFrame *frame, int flags) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVFilterLink *inlink = ctx->inputs[0]; yading@10: int ret; yading@10: AVFrame *cur_frame; yading@10: yading@10: /* no picref available, fetch it from the filterchain */ yading@10: if (!av_fifo_size(buf->fifo)) { yading@10: if (flags & AV_BUFFERSINK_FLAG_NO_REQUEST) yading@10: return AVERROR(EAGAIN); yading@10: if ((ret = ff_request_frame(inlink)) < 0) yading@10: return ret; yading@10: } yading@10: yading@10: if (!av_fifo_size(buf->fifo)) yading@10: return AVERROR(EINVAL); yading@10: yading@10: if (flags & AV_BUFFERSINK_FLAG_PEEK) { yading@10: cur_frame = *((AVFrame **)av_fifo_peek2(buf->fifo, 0)); yading@10: if ((ret = av_frame_ref(frame, cur_frame)) < 0) yading@10: return ret; yading@10: } else { yading@10: av_fifo_generic_read(buf->fifo, &cur_frame, sizeof(cur_frame), NULL); yading@10: av_frame_move_ref(frame, cur_frame); yading@10: av_frame_free(&cur_frame); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static int read_from_fifo(AVFilterContext *ctx, AVFrame *frame, yading@10: int nb_samples) yading@10: { yading@10: BufferSinkContext *s = ctx->priv; yading@10: AVFilterLink *link = ctx->inputs[0]; yading@10: AVFrame *tmp; yading@10: yading@10: if (!(tmp = ff_get_audio_buffer(link, nb_samples))) yading@10: return AVERROR(ENOMEM); yading@10: av_audio_fifo_read(s->audio_fifo, (void**)tmp->extended_data, nb_samples); yading@10: yading@10: tmp->pts = s->next_pts; yading@10: s->next_pts += av_rescale_q(nb_samples, (AVRational){1, link->sample_rate}, yading@10: link->time_base); yading@10: yading@10: av_frame_move_ref(frame, tmp); yading@10: av_frame_free(&tmp); yading@10: yading@10: return 0; yading@10: yading@10: } yading@10: yading@10: int attribute_align_arg av_buffersink_get_samples(AVFilterContext *ctx, AVFrame *frame, int nb_samples) yading@10: { yading@10: BufferSinkContext *s = ctx->priv; yading@10: AVFilterLink *link = ctx->inputs[0]; yading@10: AVFrame *cur_frame; yading@10: int ret = 0; yading@10: yading@10: if (!s->audio_fifo) { yading@10: int nb_channels = link->channels; yading@10: if (!(s->audio_fifo = av_audio_fifo_alloc(link->format, nb_channels, nb_samples))) yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: yading@10: while (ret >= 0) { yading@10: if (av_audio_fifo_size(s->audio_fifo) >= nb_samples) yading@10: return read_from_fifo(ctx, frame, nb_samples); yading@10: yading@10: if (!(cur_frame = av_frame_alloc())) yading@10: return AVERROR(ENOMEM); yading@10: ret = av_buffersink_get_frame_flags(ctx, cur_frame, 0); yading@10: if (ret == AVERROR_EOF && av_audio_fifo_size(s->audio_fifo)) { yading@10: av_frame_free(&cur_frame); yading@10: return read_from_fifo(ctx, frame, av_audio_fifo_size(s->audio_fifo)); yading@10: } else if (ret < 0) { yading@10: av_frame_free(&cur_frame); yading@10: return ret; yading@10: } yading@10: yading@10: if (cur_frame->pts != AV_NOPTS_VALUE) { yading@10: s->next_pts = cur_frame->pts - yading@10: av_rescale_q(av_audio_fifo_size(s->audio_fifo), yading@10: (AVRational){ 1, link->sample_rate }, yading@10: link->time_base); yading@10: } yading@10: yading@10: ret = av_audio_fifo_write(s->audio_fifo, (void**)cur_frame->extended_data, yading@10: cur_frame->nb_samples); yading@10: av_frame_free(&cur_frame); yading@10: } yading@10: yading@10: return ret; yading@10: yading@10: } yading@10: yading@10: AVBufferSinkParams *av_buffersink_params_alloc(void) yading@10: { yading@10: static const int pixel_fmts[] = { AV_PIX_FMT_NONE }; yading@10: AVBufferSinkParams *params = av_malloc(sizeof(AVBufferSinkParams)); yading@10: if (!params) yading@10: return NULL; yading@10: yading@10: params->pixel_fmts = pixel_fmts; yading@10: return params; yading@10: } yading@10: yading@10: AVABufferSinkParams *av_abuffersink_params_alloc(void) yading@10: { yading@10: AVABufferSinkParams *params = av_mallocz(sizeof(AVABufferSinkParams)); yading@10: yading@10: if (!params) yading@10: return NULL; yading@10: return params; yading@10: } yading@10: yading@10: #define FIFO_INIT_SIZE 8 yading@10: yading@10: static av_cold int common_init(AVFilterContext *ctx) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: yading@10: buf->fifo = av_fifo_alloc(FIFO_INIT_SIZE*sizeof(AVFilterBufferRef *)); yading@10: if (!buf->fifo) { yading@10: av_log(ctx, AV_LOG_ERROR, "Failed to allocate fifo\n"); yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: buf->warning_limit = 100; yading@10: return 0; yading@10: } yading@10: yading@10: void av_buffersink_set_frame_size(AVFilterContext *ctx, unsigned frame_size) yading@10: { yading@10: AVFilterLink *inlink = ctx->inputs[0]; yading@10: yading@10: inlink->min_samples = inlink->max_samples = yading@10: inlink->partial_buf_size = frame_size; yading@10: } yading@10: yading@10: #if FF_API_AVFILTERBUFFER yading@10: static void compat_free_buffer(AVFilterBuffer *buf) yading@10: { yading@10: AVFrame *frame = buf->priv; yading@10: av_frame_free(&frame); yading@10: av_free(buf); yading@10: } yading@10: yading@10: static int attribute_align_arg compat_read(AVFilterContext *ctx, AVFilterBufferRef **pbuf, int nb_samples, int flags) yading@10: { yading@10: AVFilterBufferRef *buf; yading@10: AVFrame *frame; yading@10: int ret; yading@10: yading@10: if (!pbuf) yading@10: return ff_poll_frame(ctx->inputs[0]); yading@10: yading@10: frame = av_frame_alloc(); yading@10: if (!frame) yading@10: return AVERROR(ENOMEM); yading@10: yading@10: if (!nb_samples) yading@10: ret = av_buffersink_get_frame_flags(ctx, frame, flags); yading@10: else yading@10: ret = av_buffersink_get_samples(ctx, frame, nb_samples); yading@10: yading@10: if (ret < 0) yading@10: goto fail; yading@10: yading@10: AV_NOWARN_DEPRECATED( yading@10: if (ctx->inputs[0]->type == AVMEDIA_TYPE_VIDEO) { yading@10: buf = avfilter_get_video_buffer_ref_from_arrays(frame->data, frame->linesize, yading@10: AV_PERM_READ, yading@10: frame->width, frame->height, yading@10: frame->format); yading@10: } else { yading@10: buf = avfilter_get_audio_buffer_ref_from_arrays(frame->extended_data, yading@10: frame->linesize[0], AV_PERM_READ, yading@10: frame->nb_samples, yading@10: frame->format, yading@10: frame->channel_layout); yading@10: } yading@10: if (!buf) { yading@10: ret = AVERROR(ENOMEM); yading@10: goto fail; yading@10: } yading@10: yading@10: avfilter_copy_frame_props(buf, frame); yading@10: ) yading@10: yading@10: buf->buf->priv = frame; yading@10: buf->buf->free = compat_free_buffer; yading@10: yading@10: *pbuf = buf; yading@10: yading@10: return 0; yading@10: fail: yading@10: av_frame_free(&frame); yading@10: return ret; yading@10: } yading@10: yading@10: int av_buffersink_read(AVFilterContext *ctx, AVFilterBufferRef **buf) yading@10: { yading@10: return compat_read(ctx, buf, 0, 0); yading@10: } yading@10: yading@10: int av_buffersink_read_samples(AVFilterContext *ctx, AVFilterBufferRef **buf, yading@10: int nb_samples) yading@10: { yading@10: return compat_read(ctx, buf, nb_samples, 0); yading@10: } yading@10: yading@10: int av_buffersink_get_buffer_ref(AVFilterContext *ctx, yading@10: AVFilterBufferRef **bufref, int flags) yading@10: { yading@10: *bufref = NULL; yading@10: yading@10: av_assert0( !strcmp(ctx->filter->name, "buffersink") yading@10: || !strcmp(ctx->filter->name, "abuffersink") yading@10: || !strcmp(ctx->filter->name, "ffbuffersink") yading@10: || !strcmp(ctx->filter->name, "ffabuffersink")); yading@10: yading@10: return compat_read(ctx, bufref, 0, flags); yading@10: } yading@10: #endif yading@10: yading@10: AVRational av_buffersink_get_frame_rate(AVFilterContext *ctx) yading@10: { yading@10: av_assert0( !strcmp(ctx->filter->name, "buffersink") yading@10: || !strcmp(ctx->filter->name, "ffbuffersink")); yading@10: yading@10: return ctx->inputs[0]->frame_rate; yading@10: } yading@10: yading@10: int attribute_align_arg av_buffersink_poll_frame(AVFilterContext *ctx) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVFilterLink *inlink = ctx->inputs[0]; yading@10: yading@10: av_assert0( !strcmp(ctx->filter->name, "buffersink") yading@10: || !strcmp(ctx->filter->name, "abuffersink") yading@10: || !strcmp(ctx->filter->name, "ffbuffersink") yading@10: || !strcmp(ctx->filter->name, "ffabuffersink")); yading@10: yading@10: return av_fifo_size(buf->fifo)/sizeof(AVFilterBufferRef *) + ff_poll_frame(inlink); yading@10: } yading@10: yading@10: static av_cold int vsink_init(AVFilterContext *ctx, void *opaque) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVBufferSinkParams *params = opaque; yading@10: int ret; yading@10: yading@10: if (params) { yading@10: if ((ret = av_opt_set_int_list(buf, "pix_fmts", params->pixel_fmts, AV_PIX_FMT_NONE, 0)) < 0) yading@10: return ret; yading@10: } yading@10: yading@10: return common_init(ctx); yading@10: } yading@10: yading@10: #define CHECK_LIST_SIZE(field) \ yading@10: if (buf->field ## _size % sizeof(*buf->field)) { \ yading@10: av_log(ctx, AV_LOG_ERROR, "Invalid size for " #field ": %d, " \ yading@10: "should be multiple of %d\n", \ yading@10: buf->field ## _size, (int)sizeof(*buf->field)); \ yading@10: return AVERROR(EINVAL); \ yading@10: } yading@10: static int vsink_query_formats(AVFilterContext *ctx) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVFilterFormats *formats = NULL; yading@10: unsigned i; yading@10: int ret; yading@10: yading@10: CHECK_LIST_SIZE(pixel_fmts) yading@10: if (buf->pixel_fmts_size) { yading@10: for (i = 0; i < NB_ITEMS(buf->pixel_fmts); i++) yading@10: if ((ret = ff_add_format(&formats, buf->pixel_fmts[i])) < 0) yading@10: return ret; yading@10: ff_set_common_formats(ctx, formats); yading@10: } else { yading@10: ff_default_query_formats(ctx); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static av_cold int asink_init(AVFilterContext *ctx, void *opaque) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVABufferSinkParams *params = opaque; yading@10: int ret; yading@10: yading@10: if (params) { yading@10: if ((ret = av_opt_set_int_list(buf, "sample_fmts", params->sample_fmts, AV_SAMPLE_FMT_NONE, 0)) < 0 || yading@10: (ret = av_opt_set_int_list(buf, "sample_rates", params->sample_rates, -1, 0)) < 0 || yading@10: (ret = av_opt_set_int_list(buf, "channel_layouts", params->channel_layouts, -1, 0)) < 0 || yading@10: (ret = av_opt_set_int_list(buf, "channel_counts", params->channel_counts, -1, 0)) < 0 || yading@10: (ret = av_opt_set_int(buf, "all_channel_counts", params->all_channel_counts, 0)) < 0) yading@10: return ret; yading@10: } yading@10: return common_init(ctx); yading@10: } yading@10: yading@10: static int asink_query_formats(AVFilterContext *ctx) yading@10: { yading@10: BufferSinkContext *buf = ctx->priv; yading@10: AVFilterFormats *formats = NULL; yading@10: AVFilterChannelLayouts *layouts = NULL; yading@10: unsigned i; yading@10: int ret; yading@10: yading@10: CHECK_LIST_SIZE(sample_fmts) yading@10: CHECK_LIST_SIZE(sample_rates) yading@10: CHECK_LIST_SIZE(channel_layouts) yading@10: CHECK_LIST_SIZE(channel_counts) yading@10: yading@10: if (buf->sample_fmts_size) { yading@10: for (i = 0; i < NB_ITEMS(buf->sample_fmts); i++) yading@10: if ((ret = ff_add_format(&formats, buf->sample_fmts[i])) < 0) yading@10: return ret; yading@10: ff_set_common_formats(ctx, formats); yading@10: } yading@10: yading@10: if (buf->channel_layouts_size || buf->channel_counts_size || yading@10: buf->all_channel_counts) { yading@10: for (i = 0; i < NB_ITEMS(buf->channel_layouts); i++) yading@10: if ((ret = ff_add_channel_layout(&layouts, buf->channel_layouts[i])) < 0) yading@10: return ret; yading@10: for (i = 0; i < NB_ITEMS(buf->channel_counts); i++) yading@10: if ((ret = ff_add_channel_layout(&layouts, FF_COUNT2LAYOUT(buf->channel_counts[i]))) < 0) yading@10: return ret; yading@10: if (buf->all_channel_counts) { yading@10: if (layouts) yading@10: av_log(ctx, AV_LOG_WARNING, yading@10: "Conflicting all_channel_counts and list in options\n"); yading@10: else if (!(layouts = ff_all_channel_counts())) yading@10: return AVERROR(ENOMEM); yading@10: } yading@10: ff_set_common_channel_layouts(ctx, layouts); yading@10: } yading@10: yading@10: if (buf->sample_rates_size) { yading@10: formats = NULL; yading@10: for (i = 0; i < NB_ITEMS(buf->sample_rates); i++) yading@10: if ((ret = ff_add_format(&formats, buf->sample_rates[i])) < 0) yading@10: return ret; yading@10: ff_set_common_samplerates(ctx, formats); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: #define OFFSET(x) offsetof(BufferSinkContext, x) yading@10: #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM yading@10: static const AVOption buffersink_options[] = { yading@10: { "pix_fmts", "set the supported pixel formats", OFFSET(pixel_fmts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, yading@10: { NULL }, yading@10: }; yading@10: #undef FLAGS yading@10: #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM yading@10: static const AVOption abuffersink_options[] = { yading@10: { "sample_fmts", "set the supported sample formats", OFFSET(sample_fmts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, yading@10: { "sample_rates", "set the supported sample rates", OFFSET(sample_rates), AV_OPT_TYPE_BINARY, .flags = FLAGS }, yading@10: { "channel_layouts", "set the supported channel layouts", OFFSET(channel_layouts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, yading@10: { "channel_counts", "set the supported channel counts", OFFSET(channel_counts), AV_OPT_TYPE_BINARY, .flags = FLAGS }, yading@10: { "all_channel_counts", "accept all channel counts", OFFSET(all_channel_counts), AV_OPT_TYPE_INT, {.i64 = 0}, 0, 1, FLAGS }, yading@10: { NULL }, yading@10: }; yading@10: #undef FLAGS yading@10: yading@10: AVFILTER_DEFINE_CLASS(buffersink); yading@10: AVFILTER_DEFINE_CLASS(abuffersink); yading@10: yading@10: #if FF_API_AVFILTERBUFFER yading@10: yading@10: #define ffbuffersink_options buffersink_options yading@10: #define ffabuffersink_options abuffersink_options yading@10: AVFILTER_DEFINE_CLASS(ffbuffersink); yading@10: AVFILTER_DEFINE_CLASS(ffabuffersink); yading@10: yading@10: static const AVFilterPad ffbuffersink_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL }, yading@10: }; yading@10: yading@10: AVFilter avfilter_vsink_ffbuffersink = { yading@10: .name = "ffbuffersink", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them available to the end of the filter graph."), yading@10: .priv_size = sizeof(BufferSinkContext), yading@10: .priv_class = &ffbuffersink_class, yading@10: .init_opaque = vsink_init, yading@10: .uninit = uninit, yading@10: yading@10: .query_formats = vsink_query_formats, yading@10: .inputs = ffbuffersink_inputs, yading@10: .outputs = NULL, yading@10: }; yading@10: yading@10: static const AVFilterPad ffabuffersink_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_AUDIO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL }, yading@10: }; yading@10: yading@10: AVFilter avfilter_asink_ffabuffersink = { yading@10: .name = "ffabuffersink", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them available to the end of the filter graph."), yading@10: .init_opaque = asink_init, yading@10: .uninit = uninit, yading@10: .priv_size = sizeof(BufferSinkContext), yading@10: .priv_class = &ffabuffersink_class, yading@10: .query_formats = asink_query_formats, yading@10: .inputs = ffabuffersink_inputs, yading@10: .outputs = NULL, yading@10: }; yading@10: #endif /* FF_API_AVFILTERBUFFER */ yading@10: yading@10: static const AVFilterPad avfilter_vsink_buffer_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_vsink_buffer = { yading@10: .name = "buffersink", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them available to the end of the filter graph."), yading@10: .priv_size = sizeof(BufferSinkContext), yading@10: .priv_class = &buffersink_class, yading@10: .init_opaque = vsink_init, yading@10: .uninit = uninit, yading@10: yading@10: .query_formats = vsink_query_formats, yading@10: .inputs = avfilter_vsink_buffer_inputs, yading@10: .outputs = NULL, yading@10: }; yading@10: yading@10: static const AVFilterPad avfilter_asink_abuffer_inputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_AUDIO, yading@10: .filter_frame = filter_frame, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_asink_abuffer = { yading@10: .name = "abuffersink", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them available to the end of the filter graph."), yading@10: .priv_class = &abuffersink_class, yading@10: .priv_size = sizeof(BufferSinkContext), yading@10: .init_opaque = asink_init, yading@10: .uninit = uninit, yading@10: yading@10: .query_formats = asink_query_formats, yading@10: .inputs = avfilter_asink_abuffer_inputs, yading@10: .outputs = NULL, yading@10: };