yading@10: /* yading@10: * Copyright (c) 2008 Vitor Sessak yading@10: * yading@10: * This file is part of FFmpeg. yading@10: * yading@10: * FFmpeg is free software; you can redistribute it and/or yading@10: * modify it under the terms of the GNU Lesser General Public yading@10: * License as published by the Free Software Foundation; either yading@10: * version 2.1 of the License, or (at your option) any later version. yading@10: * yading@10: * FFmpeg is distributed in the hope that it will be useful, yading@10: * but WITHOUT ANY WARRANTY; without even the implied warranty of yading@10: * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU yading@10: * Lesser General Public License for more details. yading@10: * yading@10: * You should have received a copy of the GNU Lesser General Public yading@10: * License along with FFmpeg; if not, write to the Free Software yading@10: * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA yading@10: */ yading@10: yading@10: /** yading@10: * @file yading@10: * memory buffer source filter yading@10: */ yading@10: yading@10: #include yading@10: yading@10: #include "libavutil/channel_layout.h" yading@10: #include "libavutil/common.h" yading@10: #include "libavutil/fifo.h" yading@10: #include "libavutil/frame.h" yading@10: #include "libavutil/imgutils.h" yading@10: #include "libavutil/opt.h" yading@10: #include "libavutil/samplefmt.h" yading@10: #include "audio.h" yading@10: #include "avfilter.h" yading@10: #include "buffersrc.h" yading@10: #include "formats.h" yading@10: #include "internal.h" yading@10: #include "video.h" yading@10: #include "avcodec.h" yading@10: yading@10: typedef struct { yading@10: const AVClass *class; yading@10: AVFifoBuffer *fifo; yading@10: AVRational time_base; ///< time_base to set in the output link yading@10: AVRational frame_rate; ///< frame_rate to set in the output link yading@10: unsigned nb_failed_requests; yading@10: unsigned warning_limit; yading@10: yading@10: /* video only */ yading@10: int w, h; yading@10: enum AVPixelFormat pix_fmt; yading@10: char *pix_fmt_str; yading@10: AVRational pixel_aspect; yading@10: char *sws_param; yading@10: yading@10: /* audio only */ yading@10: int sample_rate; yading@10: enum AVSampleFormat sample_fmt; yading@10: char *sample_fmt_str; yading@10: int channels; yading@10: uint64_t channel_layout; yading@10: char *channel_layout_str; yading@10: yading@10: int eof; yading@10: } BufferSourceContext; yading@10: yading@10: #define CHECK_VIDEO_PARAM_CHANGE(s, c, width, height, format)\ yading@10: if (c->w != width || c->h != height || c->pix_fmt != format) {\ yading@10: av_log(s, AV_LOG_INFO, "Changing frame properties on the fly is not supported by all filters.\n");\ yading@10: } yading@10: yading@10: #define CHECK_AUDIO_PARAM_CHANGE(s, c, srate, ch_layout, ch_count, format)\ yading@10: if (c->sample_fmt != format || c->sample_rate != srate ||\ yading@10: c->channel_layout != ch_layout || c->channels != ch_count) {\ yading@10: av_log(s, AV_LOG_ERROR, "Changing frame properties on the fly is not supported.\n");\ yading@10: return AVERROR(EINVAL);\ yading@10: } yading@10: yading@10: int av_buffersrc_write_frame(AVFilterContext *ctx, const AVFrame *frame) yading@10: { yading@10: return av_buffersrc_add_frame_flags(ctx, (AVFrame *)frame, yading@10: AV_BUFFERSRC_FLAG_KEEP_REF); yading@10: } yading@10: yading@10: int av_buffersrc_add_frame(AVFilterContext *ctx, AVFrame *frame) yading@10: { yading@10: return av_buffersrc_add_frame_flags(ctx, frame, 0); yading@10: } yading@10: yading@10: static int av_buffersrc_add_frame_internal(AVFilterContext *ctx, yading@10: AVFrame *frame, int flags); yading@10: yading@10: int av_buffersrc_add_frame_flags(AVFilterContext *ctx, AVFrame *frame, int flags) yading@10: { yading@10: AVFrame *copy = NULL; yading@10: int ret = 0; yading@10: yading@10: if (frame && frame->channel_layout && yading@10: av_get_channel_layout_nb_channels(frame->channel_layout) != av_frame_get_channels(frame)) { yading@10: av_log(0, AV_LOG_ERROR, "Layout indicates a different number of channels than actually present\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: if (!(flags & AV_BUFFERSRC_FLAG_KEEP_REF) || !frame) yading@10: return av_buffersrc_add_frame_internal(ctx, frame, flags); yading@10: yading@10: if (!(copy = av_frame_alloc())) yading@10: return AVERROR(ENOMEM); yading@10: ret = av_frame_ref(copy, frame); yading@10: if (ret >= 0) yading@10: ret = av_buffersrc_add_frame_internal(ctx, copy, flags); yading@10: yading@10: av_frame_free(©); yading@10: return ret; yading@10: } yading@10: yading@10: static int attribute_align_arg av_buffersrc_add_frame_internal(AVFilterContext *ctx, yading@10: AVFrame *frame, int flags) yading@10: { yading@10: BufferSourceContext *s = ctx->priv; yading@10: AVFrame *copy; yading@10: int ret; yading@10: yading@10: s->nb_failed_requests = 0; yading@10: yading@10: if (!frame) { yading@10: s->eof = 1; yading@10: return 0; yading@10: } else if (s->eof) yading@10: return AVERROR(EINVAL); yading@10: yading@10: if (!(flags & AV_BUFFERSRC_FLAG_NO_CHECK_FORMAT)) { yading@10: yading@10: switch (ctx->outputs[0]->type) { yading@10: case AVMEDIA_TYPE_VIDEO: yading@10: CHECK_VIDEO_PARAM_CHANGE(ctx, s, frame->width, frame->height, yading@10: frame->format); yading@10: break; yading@10: case AVMEDIA_TYPE_AUDIO: yading@10: /* For layouts unknown on input but known on link after negotiation. */ yading@10: if (!frame->channel_layout) yading@10: frame->channel_layout = s->channel_layout; yading@10: CHECK_AUDIO_PARAM_CHANGE(ctx, s, frame->sample_rate, frame->channel_layout, yading@10: av_frame_get_channels(frame), frame->format); yading@10: break; yading@10: default: yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: } yading@10: yading@10: if (!av_fifo_space(s->fifo) && yading@10: (ret = av_fifo_realloc2(s->fifo, av_fifo_size(s->fifo) + yading@10: sizeof(copy))) < 0) yading@10: return ret; yading@10: yading@10: if (!(copy = av_frame_alloc())) yading@10: return AVERROR(ENOMEM); yading@10: av_frame_move_ref(copy, frame); yading@10: yading@10: if ((ret = av_fifo_generic_write(s->fifo, ©, sizeof(copy), NULL)) < 0) { yading@10: av_frame_move_ref(frame, copy); yading@10: av_frame_free(©); yading@10: return ret; yading@10: } yading@10: yading@10: if ((flags & AV_BUFFERSRC_FLAG_PUSH)) yading@10: if ((ret = ctx->output_pads[0].request_frame(ctx->outputs[0])) < 0) yading@10: return ret; yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: #if FF_API_AVFILTERBUFFER yading@10: static void compat_free_buffer(void *opaque, uint8_t *data) yading@10: { yading@10: AVFilterBufferRef *buf = opaque; yading@10: AV_NOWARN_DEPRECATED( yading@10: avfilter_unref_buffer(buf); yading@10: ) yading@10: } yading@10: yading@10: static void compat_unref_buffer(void *opaque, uint8_t *data) yading@10: { yading@10: AVBufferRef *buf = opaque; yading@10: AV_NOWARN_DEPRECATED( yading@10: av_buffer_unref(&buf); yading@10: ) yading@10: } yading@10: yading@10: int av_buffersrc_add_ref(AVFilterContext *ctx, AVFilterBufferRef *buf, yading@10: int flags) yading@10: { yading@10: BufferSourceContext *s = ctx->priv; yading@10: AVFrame *frame = NULL; yading@10: AVBufferRef *dummy_buf = NULL; yading@10: int ret = 0, planes, i; yading@10: yading@10: if (!buf) { yading@10: s->eof = 1; yading@10: return 0; yading@10: } else if (s->eof) yading@10: return AVERROR(EINVAL); yading@10: yading@10: frame = av_frame_alloc(); yading@10: if (!frame) yading@10: return AVERROR(ENOMEM); yading@10: yading@10: dummy_buf = av_buffer_create(NULL, 0, compat_free_buffer, buf, yading@10: (buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY); yading@10: if (!dummy_buf) { yading@10: ret = AVERROR(ENOMEM); yading@10: goto fail; yading@10: } yading@10: yading@10: AV_NOWARN_DEPRECATED( yading@10: if ((ret = avfilter_copy_buf_props(frame, buf)) < 0) yading@10: goto fail; yading@10: ) yading@10: yading@10: #define WRAP_PLANE(ref_out, data, data_size) \ yading@10: do { \ yading@10: AVBufferRef *dummy_ref = av_buffer_ref(dummy_buf); \ yading@10: if (!dummy_ref) { \ yading@10: ret = AVERROR(ENOMEM); \ yading@10: goto fail; \ yading@10: } \ yading@10: ref_out = av_buffer_create(data, data_size, compat_unref_buffer, \ yading@10: dummy_ref, (buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY); \ yading@10: if (!ref_out) { \ yading@10: av_frame_unref(frame); \ yading@10: ret = AVERROR(ENOMEM); \ yading@10: goto fail; \ yading@10: } \ yading@10: } while (0) yading@10: yading@10: if (ctx->outputs[0]->type == AVMEDIA_TYPE_VIDEO) { yading@10: const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format); yading@10: yading@10: planes = av_pix_fmt_count_planes(frame->format); yading@10: if (!desc || planes <= 0) { yading@10: ret = AVERROR(EINVAL); yading@10: goto fail; yading@10: } yading@10: yading@10: for (i = 0; i < planes; i++) { yading@10: int v_shift = (i == 1 || i == 2) ? desc->log2_chroma_h : 0; yading@10: int plane_size = (frame->height >> v_shift) * frame->linesize[i]; yading@10: yading@10: WRAP_PLANE(frame->buf[i], frame->data[i], plane_size); yading@10: } yading@10: } else { yading@10: int planar = av_sample_fmt_is_planar(frame->format); yading@10: int channels = av_get_channel_layout_nb_channels(frame->channel_layout); yading@10: yading@10: planes = planar ? channels : 1; yading@10: yading@10: if (planes > FF_ARRAY_ELEMS(frame->buf)) { yading@10: frame->nb_extended_buf = planes - FF_ARRAY_ELEMS(frame->buf); yading@10: frame->extended_buf = av_mallocz(sizeof(*frame->extended_buf) * yading@10: frame->nb_extended_buf); yading@10: if (!frame->extended_buf) { yading@10: ret = AVERROR(ENOMEM); yading@10: goto fail; yading@10: } yading@10: } yading@10: yading@10: for (i = 0; i < FFMIN(planes, FF_ARRAY_ELEMS(frame->buf)); i++) yading@10: WRAP_PLANE(frame->buf[i], frame->extended_data[i], frame->linesize[0]); yading@10: yading@10: for (i = 0; i < planes - FF_ARRAY_ELEMS(frame->buf); i++) yading@10: WRAP_PLANE(frame->extended_buf[i], yading@10: frame->extended_data[i + FF_ARRAY_ELEMS(frame->buf)], yading@10: frame->linesize[0]); yading@10: } yading@10: yading@10: ret = av_buffersrc_add_frame_flags(ctx, frame, flags); yading@10: yading@10: fail: yading@10: av_buffer_unref(&dummy_buf); yading@10: av_frame_free(&frame); yading@10: yading@10: return ret; yading@10: } yading@10: yading@10: int av_buffersrc_buffer(AVFilterContext *ctx, AVFilterBufferRef *buf) yading@10: { yading@10: return av_buffersrc_add_ref(ctx, buf, 0); yading@10: } yading@10: #endif yading@10: yading@10: static av_cold int init_video(AVFilterContext *ctx) yading@10: { yading@10: BufferSourceContext *c = ctx->priv; yading@10: yading@10: if (c->pix_fmt == AV_PIX_FMT_NONE || !c->w || !c->h || av_q2d(c->time_base) <= 0) { yading@10: av_log(ctx, AV_LOG_ERROR, "Invalid parameters provided.\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: if (!(c->fifo = av_fifo_alloc(sizeof(AVFrame*)))) yading@10: return AVERROR(ENOMEM); yading@10: yading@10: av_log(ctx, AV_LOG_VERBOSE, "w:%d h:%d pixfmt:%s tb:%d/%d fr:%d/%d sar:%d/%d sws_param:%s\n", yading@10: c->w, c->h, av_get_pix_fmt_name(c->pix_fmt), yading@10: c->time_base.num, c->time_base.den, c->frame_rate.num, c->frame_rate.den, yading@10: c->pixel_aspect.num, c->pixel_aspect.den, (char *)av_x_if_null(c->sws_param, "")); yading@10: c->warning_limit = 100; yading@10: return 0; yading@10: } yading@10: yading@10: unsigned av_buffersrc_get_nb_failed_requests(AVFilterContext *buffer_src) yading@10: { yading@10: return ((BufferSourceContext *)buffer_src->priv)->nb_failed_requests; yading@10: } yading@10: yading@10: #define OFFSET(x) offsetof(BufferSourceContext, x) yading@10: #define A AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_AUDIO_PARAM yading@10: #define V AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM yading@10: yading@10: static const AVOption buffer_options[] = { yading@10: { "width", NULL, OFFSET(w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: { "video_size", NULL, OFFSET(w), AV_OPT_TYPE_IMAGE_SIZE, .flags = V }, yading@10: { "height", NULL, OFFSET(h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: { "pix_fmt", NULL, OFFSET(pix_fmt), AV_OPT_TYPE_PIXEL_FMT, .flags = V }, yading@10: #if FF_API_OLD_FILTER_OPTS yading@10: /* those 4 are for compatibility with the old option passing system where each filter yading@10: * did its own parsing */ yading@10: { "time_base_num", "deprecated, do not use", OFFSET(time_base.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: { "time_base_den", "deprecated, do not use", OFFSET(time_base.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: { "sar_num", "deprecated, do not use", OFFSET(pixel_aspect.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: { "sar_den", "deprecated, do not use", OFFSET(pixel_aspect.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V }, yading@10: #endif yading@10: { "sar", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V }, yading@10: { "pixel_aspect", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V }, yading@10: { "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V }, yading@10: { "frame_rate", NULL, OFFSET(frame_rate), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V }, yading@10: { "sws_param", NULL, OFFSET(sws_param), AV_OPT_TYPE_STRING, .flags = V }, yading@10: { NULL }, yading@10: }; yading@10: yading@10: AVFILTER_DEFINE_CLASS(buffer); yading@10: yading@10: static const AVOption abuffer_options[] = { yading@10: { "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, INT_MAX, A }, yading@10: { "sample_rate", NULL, OFFSET(sample_rate), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A }, yading@10: { "sample_fmt", NULL, OFFSET(sample_fmt_str), AV_OPT_TYPE_STRING, .flags = A }, yading@10: { "channel_layout", NULL, OFFSET(channel_layout_str), AV_OPT_TYPE_STRING, .flags = A }, yading@10: { "channels", NULL, OFFSET(channels), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A }, yading@10: { NULL }, yading@10: }; yading@10: yading@10: AVFILTER_DEFINE_CLASS(abuffer); yading@10: yading@10: static av_cold int init_audio(AVFilterContext *ctx) yading@10: { yading@10: BufferSourceContext *s = ctx->priv; yading@10: int ret = 0; yading@10: yading@10: s->sample_fmt = av_get_sample_fmt(s->sample_fmt_str); yading@10: if (s->sample_fmt == AV_SAMPLE_FMT_NONE) { yading@10: av_log(ctx, AV_LOG_ERROR, "Invalid sample format %s\n", yading@10: s->sample_fmt_str); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: if (s->channel_layout_str) { yading@10: int n; yading@10: /* TODO reindent */ yading@10: s->channel_layout = av_get_channel_layout(s->channel_layout_str); yading@10: if (!s->channel_layout) { yading@10: av_log(ctx, AV_LOG_ERROR, "Invalid channel layout %s.\n", yading@10: s->channel_layout_str); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: n = av_get_channel_layout_nb_channels(s->channel_layout); yading@10: if (s->channels) { yading@10: if (n != s->channels) { yading@10: av_log(ctx, AV_LOG_ERROR, yading@10: "Mismatching channel count %d and layout '%s' " yading@10: "(%d channels)\n", yading@10: s->channels, s->channel_layout_str, n); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: } yading@10: s->channels = n; yading@10: } else if (!s->channels) { yading@10: av_log(ctx, AV_LOG_ERROR, "Neither number of channels nor " yading@10: "channel layout specified\n"); yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: if (!(s->fifo = av_fifo_alloc(sizeof(AVFrame*)))) yading@10: return AVERROR(ENOMEM); yading@10: yading@10: if (!s->time_base.num) yading@10: s->time_base = (AVRational){1, s->sample_rate}; yading@10: yading@10: av_log(ctx, AV_LOG_VERBOSE, yading@10: "tb:%d/%d samplefmt:%s samplerate:%d chlayout:%s\n", yading@10: s->time_base.num, s->time_base.den, s->sample_fmt_str, yading@10: s->sample_rate, s->channel_layout_str); yading@10: s->warning_limit = 100; yading@10: yading@10: return ret; yading@10: } yading@10: yading@10: static av_cold void uninit(AVFilterContext *ctx) yading@10: { yading@10: BufferSourceContext *s = ctx->priv; yading@10: while (s->fifo && av_fifo_size(s->fifo)) { yading@10: AVFrame *frame; yading@10: av_fifo_generic_read(s->fifo, &frame, sizeof(frame), NULL); yading@10: av_frame_free(&frame); yading@10: } yading@10: av_fifo_free(s->fifo); yading@10: s->fifo = NULL; yading@10: } yading@10: yading@10: static int query_formats(AVFilterContext *ctx) yading@10: { yading@10: BufferSourceContext *c = ctx->priv; yading@10: AVFilterChannelLayouts *channel_layouts = NULL; yading@10: AVFilterFormats *formats = NULL; yading@10: AVFilterFormats *samplerates = NULL; yading@10: yading@10: switch (ctx->outputs[0]->type) { yading@10: case AVMEDIA_TYPE_VIDEO: yading@10: ff_add_format(&formats, c->pix_fmt); yading@10: ff_set_common_formats(ctx, formats); yading@10: break; yading@10: case AVMEDIA_TYPE_AUDIO: yading@10: ff_add_format(&formats, c->sample_fmt); yading@10: ff_set_common_formats(ctx, formats); yading@10: yading@10: ff_add_format(&samplerates, c->sample_rate); yading@10: ff_set_common_samplerates(ctx, samplerates); yading@10: yading@10: ff_add_channel_layout(&channel_layouts, yading@10: c->channel_layout ? c->channel_layout : yading@10: FF_COUNT2LAYOUT(c->channels)); yading@10: ff_set_common_channel_layouts(ctx, channel_layouts); yading@10: break; yading@10: default: yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: return 0; yading@10: } yading@10: yading@10: static int config_props(AVFilterLink *link) yading@10: { yading@10: BufferSourceContext *c = link->src->priv; yading@10: yading@10: switch (link->type) { yading@10: case AVMEDIA_TYPE_VIDEO: yading@10: link->w = c->w; yading@10: link->h = c->h; yading@10: link->sample_aspect_ratio = c->pixel_aspect; yading@10: break; yading@10: case AVMEDIA_TYPE_AUDIO: yading@10: if (!c->channel_layout) yading@10: c->channel_layout = link->channel_layout; yading@10: break; yading@10: default: yading@10: return AVERROR(EINVAL); yading@10: } yading@10: yading@10: link->time_base = c->time_base; yading@10: link->frame_rate = c->frame_rate; yading@10: return 0; yading@10: } yading@10: yading@10: static int request_frame(AVFilterLink *link) yading@10: { yading@10: BufferSourceContext *c = link->src->priv; yading@10: AVFrame *frame; yading@10: yading@10: if (!av_fifo_size(c->fifo)) { yading@10: if (c->eof) yading@10: return AVERROR_EOF; yading@10: c->nb_failed_requests++; yading@10: return AVERROR(EAGAIN); yading@10: } yading@10: av_fifo_generic_read(c->fifo, &frame, sizeof(frame), NULL); yading@10: yading@10: return ff_filter_frame(link, frame); yading@10: } yading@10: yading@10: static int poll_frame(AVFilterLink *link) yading@10: { yading@10: BufferSourceContext *c = link->src->priv; yading@10: int size = av_fifo_size(c->fifo); yading@10: if (!size && c->eof) yading@10: return AVERROR_EOF; yading@10: return size/sizeof(AVFrame*); yading@10: } yading@10: yading@10: static const AVFilterPad avfilter_vsrc_buffer_outputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_VIDEO, yading@10: .request_frame = request_frame, yading@10: .poll_frame = poll_frame, yading@10: .config_props = config_props, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_vsrc_buffer = { yading@10: .name = "buffer", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them accessible to the filterchain."), yading@10: .priv_size = sizeof(BufferSourceContext), yading@10: .query_formats = query_formats, yading@10: yading@10: .init = init_video, yading@10: .uninit = uninit, yading@10: yading@10: .inputs = NULL, yading@10: .outputs = avfilter_vsrc_buffer_outputs, yading@10: .priv_class = &buffer_class, yading@10: }; yading@10: yading@10: static const AVFilterPad avfilter_asrc_abuffer_outputs[] = { yading@10: { yading@10: .name = "default", yading@10: .type = AVMEDIA_TYPE_AUDIO, yading@10: .request_frame = request_frame, yading@10: .poll_frame = poll_frame, yading@10: .config_props = config_props, yading@10: }, yading@10: { NULL } yading@10: }; yading@10: yading@10: AVFilter avfilter_asrc_abuffer = { yading@10: .name = "abuffer", yading@10: .description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them accessible to the filterchain."), yading@10: .priv_size = sizeof(BufferSourceContext), yading@10: .query_formats = query_formats, yading@10: yading@10: .init = init_audio, yading@10: .uninit = uninit, yading@10: yading@10: .inputs = NULL, yading@10: .outputs = avfilter_asrc_abuffer_outputs, yading@10: .priv_class = &abuffer_class, yading@10: };