From: Marton Balint <cus@passwd.hu> To: ffmpeg-devel@ffmpeg.org Cc: Marton Balint <cus@passwd.hu> Subject: [FFmpeg-devel] [PATCH 2/4] avfilter/af_afade: factorize functions generating frames Date: Thu, 24 Jul 2025 01:35:58 +0200 Message-ID: <20250723233604.29380-2-cus@passwd.hu> (raw) In-Reply-To: <20250723233604.29380-1-cus@passwd.hu> No change in functionality. Signed-off-by: Marton Balint <cus@passwd.hu> --- libavfilter/af_afade.c | 206 ++++++++++++++++++++++------------------- 1 file changed, 111 insertions(+), 95 deletions(-) diff --git a/libavfilter/af_afade.c b/libavfilter/af_afade.c index a9318c0ac5..b006420111 100644 --- a/libavfilter/af_afade.c +++ b/libavfilter/af_afade.c @@ -25,6 +25,7 @@ #include "config_components.h" +#include "libavutil/avassert.h" #include "libavutil/opt.h" #include "audio.h" #include "avfilter.h" @@ -547,117 +548,132 @@ static int check_input(AVFilterLink *inlink) return ff_inlink_check_available_samples(inlink, queued_samples + 1) == 1; } +static int pass_frame(AVFilterLink *inlink, AVFilterLink *outlink, int64_t *pts) +{ + AVFrame *in; + int ret = ff_inlink_consume_frame(inlink, &in); + if (ret < 0) + return ret; + av_assert1(ret); + in->pts = *pts; + *pts += av_rescale_q(in->nb_samples, + (AVRational){ 1, outlink->sample_rate }, outlink->time_base); + return ff_filter_frame(outlink, in); +} + +static int pass_samples(AVFilterLink *inlink, AVFilterLink *outlink, unsigned nb_samples, int64_t *pts) +{ + AVFrame *in; + int ret = ff_inlink_consume_samples(inlink, nb_samples, nb_samples, &in); + if (ret < 0) + return ret; + av_assert1(ret); + in->pts = *pts; + *pts += av_rescale_q(in->nb_samples, + (AVRational){ 1, outlink->sample_rate }, outlink->time_base); + return ff_filter_frame(outlink, in); +} + +static int pass_crossfade(AVFilterContext *ctx) +{ + AudioFadeContext *s = ctx->priv; + AVFilterLink *outlink = ctx->outputs[0]; + AVFrame *out, *cf[2] = { NULL }; + int ret; + + if (s->overlap) { + out = ff_get_audio_buffer(outlink, s->nb_samples); + if (!out) + return AVERROR(ENOMEM); + + ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]); + if (ret < 0) { + av_frame_free(&out); + return ret; + } + + ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]); + if (ret < 0) { + av_frame_free(&out); + return ret; + } + + s->crossfade_samples(out->extended_data, cf[0]->extended_data, + cf[1]->extended_data, + s->nb_samples, out->ch_layout.nb_channels, + s->curve, s->curve2); + out->pts = s->pts; + s->pts += av_rescale_q(s->nb_samples, + (AVRational){ 1, outlink->sample_rate }, outlink->time_base); + s->passthrough = 1; + av_frame_free(&cf[0]); + av_frame_free(&cf[1]); + return ff_filter_frame(outlink, out); + } else { + out = ff_get_audio_buffer(outlink, s->nb_samples); + if (!out) + return AVERROR(ENOMEM); + + ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]); + if (ret < 0) { + av_frame_free(&out); + return ret; + } + + s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples, + outlink->ch_layout.nb_channels, -1, s->nb_samples - 1, s->nb_samples, s->curve, 0., 1.); + out->pts = s->pts; + s->pts += av_rescale_q(s->nb_samples, + (AVRational){ 1, outlink->sample_rate }, outlink->time_base); + av_frame_free(&cf[0]); + ret = ff_filter_frame(outlink, out); + if (ret < 0) + return ret; + + out = ff_get_audio_buffer(outlink, s->nb_samples); + if (!out) + return AVERROR(ENOMEM); + + ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]); + if (ret < 0) { + av_frame_free(&out); + return ret; + } + + s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples, + outlink->ch_layout.nb_channels, 1, 0, s->nb_samples, s->curve2, 0., 1.); + out->pts = s->pts; + s->pts += av_rescale_q(s->nb_samples, + (AVRational){ 1, outlink->sample_rate }, outlink->time_base); + s->passthrough = 1; + av_frame_free(&cf[1]); + return ff_filter_frame(outlink, out); + } +} + static int activate(AVFilterContext *ctx) { AudioFadeContext *s = ctx->priv; AVFilterLink *outlink = ctx->outputs[0]; - AVFrame *in = NULL, *out, *cf[2] = { NULL }; - int ret = 0, nb_samples, status; - int64_t pts; + int ret = 0, nb_samples; FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx); if (s->passthrough && s->status[0]) { - ret = ff_inlink_consume_frame(ctx->inputs[1], &in); - if (ret > 0) { - in->pts = s->pts; - s->pts += av_rescale_q(in->nb_samples, - (AVRational){ 1, outlink->sample_rate }, outlink->time_base); - return ff_filter_frame(outlink, in); - } else if (ret < 0) { - return ret; - } else if (ff_inlink_acknowledge_status(ctx->inputs[1], &status, &pts)) { - ff_outlink_set_status(outlink, status, pts); - return 0; - } else if (!ret) { - if (ff_outlink_frame_wanted(outlink)) { - ff_inlink_request_frame(ctx->inputs[1]); - return 0; - } - } + if (ff_inlink_queued_frames(ctx->inputs[1])) + return pass_frame(ctx->inputs[1], outlink, &s->pts); + FF_FILTER_FORWARD_STATUS(ctx->inputs[1], outlink); + FF_FILTER_FORWARD_WANTED(outlink, ctx->inputs[1]); } nb_samples = ff_inlink_queued_samples(ctx->inputs[0]); if (nb_samples > s->nb_samples) { nb_samples -= s->nb_samples; s->passthrough = 1; - ret = ff_inlink_consume_samples(ctx->inputs[0], nb_samples, nb_samples, &in); - if (ret < 0) - return ret; - in->pts = s->pts; - s->pts += av_rescale_q(in->nb_samples, - (AVRational){ 1, outlink->sample_rate }, outlink->time_base); - return ff_filter_frame(outlink, in); + return pass_samples(ctx->inputs[0], outlink, nb_samples, &s->pts); } else if (s->status[0] && nb_samples >= s->nb_samples && ff_inlink_queued_samples(ctx->inputs[1]) >= s->nb_samples) { - if (s->overlap) { - out = ff_get_audio_buffer(outlink, s->nb_samples); - if (!out) - return AVERROR(ENOMEM); - - ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]); - if (ret < 0) { - av_frame_free(&out); - return ret; - } - - ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]); - if (ret < 0) { - av_frame_free(&out); - return ret; - } - - s->crossfade_samples(out->extended_data, cf[0]->extended_data, - cf[1]->extended_data, - s->nb_samples, out->ch_layout.nb_channels, - s->curve, s->curve2); - out->pts = s->pts; - s->pts += av_rescale_q(s->nb_samples, - (AVRational){ 1, outlink->sample_rate }, outlink->time_base); - s->passthrough = 1; - av_frame_free(&cf[0]); - av_frame_free(&cf[1]); - return ff_filter_frame(outlink, out); - } else { - out = ff_get_audio_buffer(outlink, s->nb_samples); - if (!out) - return AVERROR(ENOMEM); - - ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]); - if (ret < 0) { - av_frame_free(&out); - return ret; - } - - s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples, - outlink->ch_layout.nb_channels, -1, s->nb_samples - 1, s->nb_samples, s->curve, 0., 1.); - out->pts = s->pts; - s->pts += av_rescale_q(s->nb_samples, - (AVRational){ 1, outlink->sample_rate }, outlink->time_base); - av_frame_free(&cf[0]); - ret = ff_filter_frame(outlink, out); - if (ret < 0) - return ret; - - out = ff_get_audio_buffer(outlink, s->nb_samples); - if (!out) - return AVERROR(ENOMEM); - - ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]); - if (ret < 0) { - av_frame_free(&out); - return ret; - } - - s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples, - outlink->ch_layout.nb_channels, 1, 0, s->nb_samples, s->curve2, 0., 1.); - out->pts = s->pts; - s->pts += av_rescale_q(s->nb_samples, - (AVRational){ 1, outlink->sample_rate }, outlink->time_base); - s->passthrough = 1; - av_frame_free(&cf[1]); - return ff_filter_frame(outlink, out); - } + return pass_crossfade(ctx); } else if (ff_outlink_frame_wanted(outlink)) { if (!s->status[0] && check_input(ctx->inputs[0])) s->status[0] = AVERROR_EOF; -- 2.43.0 _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe".
next prev parent reply other threads:[~2025-07-23 23:36 UTC|newest] Thread overview: 4+ messages / expand[flat|nested] mbox.gz Atom feed top 2025-07-23 23:35 [FFmpeg-devel] [PATCH 1/4] avfilter/trim: consume all available frames and avoid activate reschedule Marton Balint 2025-07-23 23:35 ` Marton Balint [this message] 2025-07-23 23:35 ` [FFmpeg-devel] [PATCH 3/4] avfilter/af_afade: fix check_input for empty streams Marton Balint 2025-07-23 23:36 ` [FFmpeg-devel] [PATCH 4/4] avfilter/af_afade: rework crossfade activate logic Marton Balint
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20250723233604.29380-2-cus@passwd.hu \ --to=cus@passwd.hu \ --cc=ffmpeg-devel@ffmpeg.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel This inbox may be cloned and mirrored by anyone: git clone --mirror https://master.gitmailbox.com/ffmpegdev/0 ffmpegdev/git/0.git # If you have public-inbox 1.1+ installed, you may # initialize and index your mirror using the following commands: public-inbox-init -V2 ffmpegdev ffmpegdev/ https://master.gitmailbox.com/ffmpegdev \ ffmpegdev@gitmailbox.com public-inbox-index ffmpegdev Example config snippet for mirrors. AGPL code for this site: git clone https://public-inbox.org/public-inbox.git