Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
 help / color / mirror / Atom feed
From: Marton Balint <cus@passwd.hu>
To: ffmpeg-devel@ffmpeg.org
Cc: Marton Balint <cus@passwd.hu>
Subject: [FFmpeg-devel] [PATCH 2/4] avfilter/af_afade: factorize functions generating frames
Date: Thu, 24 Jul 2025 01:35:58 +0200
Message-ID: <20250723233604.29380-2-cus@passwd.hu> (raw)
In-Reply-To: <20250723233604.29380-1-cus@passwd.hu>

No change in functionality.

Signed-off-by: Marton Balint <cus@passwd.hu>
---
 libavfilter/af_afade.c | 206 ++++++++++++++++++++++-------------------
 1 file changed, 111 insertions(+), 95 deletions(-)

diff --git a/libavfilter/af_afade.c b/libavfilter/af_afade.c
index a9318c0ac5..b006420111 100644
--- a/libavfilter/af_afade.c
+++ b/libavfilter/af_afade.c
@@ -25,6 +25,7 @@
 
 #include "config_components.h"
 
+#include "libavutil/avassert.h"
 #include "libavutil/opt.h"
 #include "audio.h"
 #include "avfilter.h"
@@ -547,117 +548,132 @@ static int check_input(AVFilterLink *inlink)
     return ff_inlink_check_available_samples(inlink, queued_samples + 1) == 1;
 }
 
+static int pass_frame(AVFilterLink *inlink, AVFilterLink *outlink, int64_t *pts)
+{
+    AVFrame *in;
+    int ret = ff_inlink_consume_frame(inlink, &in);
+    if (ret < 0)
+        return ret;
+    av_assert1(ret);
+    in->pts = *pts;
+    *pts += av_rescale_q(in->nb_samples,
+            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+    return ff_filter_frame(outlink, in);
+}
+
+static int pass_samples(AVFilterLink *inlink, AVFilterLink *outlink, unsigned nb_samples, int64_t *pts)
+{
+    AVFrame *in;
+    int ret = ff_inlink_consume_samples(inlink, nb_samples, nb_samples, &in);
+    if (ret < 0)
+        return ret;
+    av_assert1(ret);
+    in->pts = *pts;
+    *pts += av_rescale_q(in->nb_samples,
+            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+    return ff_filter_frame(outlink, in);
+}
+
+static int pass_crossfade(AVFilterContext *ctx)
+{
+    AudioFadeContext *s = ctx->priv;
+    AVFilterLink *outlink = ctx->outputs[0];
+    AVFrame *out, *cf[2] = { NULL };
+    int ret;
+
+    if (s->overlap) {
+        out = ff_get_audio_buffer(outlink, s->nb_samples);
+        if (!out)
+            return AVERROR(ENOMEM);
+
+        ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
+        if (ret < 0) {
+            av_frame_free(&out);
+            return ret;
+        }
+
+        ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
+        if (ret < 0) {
+            av_frame_free(&out);
+            return ret;
+        }
+
+        s->crossfade_samples(out->extended_data, cf[0]->extended_data,
+                             cf[1]->extended_data,
+                             s->nb_samples, out->ch_layout.nb_channels,
+                             s->curve, s->curve2);
+        out->pts = s->pts;
+        s->pts += av_rescale_q(s->nb_samples,
+            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+        s->passthrough = 1;
+        av_frame_free(&cf[0]);
+        av_frame_free(&cf[1]);
+        return ff_filter_frame(outlink, out);
+    } else {
+        out = ff_get_audio_buffer(outlink, s->nb_samples);
+        if (!out)
+            return AVERROR(ENOMEM);
+
+        ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
+        if (ret < 0) {
+            av_frame_free(&out);
+            return ret;
+        }
+
+        s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples,
+                        outlink->ch_layout.nb_channels, -1, s->nb_samples - 1, s->nb_samples, s->curve, 0., 1.);
+        out->pts = s->pts;
+        s->pts += av_rescale_q(s->nb_samples,
+            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+        av_frame_free(&cf[0]);
+        ret = ff_filter_frame(outlink, out);
+        if (ret < 0)
+            return ret;
+
+        out = ff_get_audio_buffer(outlink, s->nb_samples);
+        if (!out)
+            return AVERROR(ENOMEM);
+
+        ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
+        if (ret < 0) {
+            av_frame_free(&out);
+            return ret;
+        }
+
+        s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples,
+                        outlink->ch_layout.nb_channels, 1, 0, s->nb_samples, s->curve2, 0., 1.);
+        out->pts = s->pts;
+        s->pts += av_rescale_q(s->nb_samples,
+            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
+        s->passthrough = 1;
+        av_frame_free(&cf[1]);
+        return ff_filter_frame(outlink, out);
+    }
+}
+
 static int activate(AVFilterContext *ctx)
 {
     AudioFadeContext *s   = ctx->priv;
     AVFilterLink *outlink = ctx->outputs[0];
-    AVFrame *in = NULL, *out, *cf[2] = { NULL };
-    int ret = 0, nb_samples, status;
-    int64_t pts;
+    int ret = 0, nb_samples;
 
     FF_FILTER_FORWARD_STATUS_BACK_ALL(outlink, ctx);
 
     if (s->passthrough && s->status[0]) {
-        ret = ff_inlink_consume_frame(ctx->inputs[1], &in);
-        if (ret > 0) {
-            in->pts = s->pts;
-            s->pts += av_rescale_q(in->nb_samples,
-                      (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
-            return ff_filter_frame(outlink, in);
-        } else if (ret < 0) {
-            return ret;
-        } else if (ff_inlink_acknowledge_status(ctx->inputs[1], &status, &pts)) {
-            ff_outlink_set_status(outlink, status, pts);
-            return 0;
-        } else if (!ret) {
-            if (ff_outlink_frame_wanted(outlink)) {
-                ff_inlink_request_frame(ctx->inputs[1]);
-                return 0;
-            }
-        }
+        if (ff_inlink_queued_frames(ctx->inputs[1]))
+            return pass_frame(ctx->inputs[1], outlink, &s->pts);
+        FF_FILTER_FORWARD_STATUS(ctx->inputs[1], outlink);
+        FF_FILTER_FORWARD_WANTED(outlink, ctx->inputs[1]);
     }
 
     nb_samples = ff_inlink_queued_samples(ctx->inputs[0]);
     if (nb_samples  > s->nb_samples) {
         nb_samples -= s->nb_samples;
         s->passthrough = 1;
-        ret = ff_inlink_consume_samples(ctx->inputs[0], nb_samples, nb_samples, &in);
-        if (ret < 0)
-            return ret;
-        in->pts = s->pts;
-        s->pts += av_rescale_q(in->nb_samples,
-            (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
-        return ff_filter_frame(outlink, in);
+        return pass_samples(ctx->inputs[0], outlink, nb_samples, &s->pts);
     } else if (s->status[0] && nb_samples >= s->nb_samples &&
                ff_inlink_queued_samples(ctx->inputs[1]) >= s->nb_samples) {
-        if (s->overlap) {
-            out = ff_get_audio_buffer(outlink, s->nb_samples);
-            if (!out)
-                return AVERROR(ENOMEM);
-
-            ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
-            if (ret < 0) {
-                av_frame_free(&out);
-                return ret;
-            }
-
-            ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
-            if (ret < 0) {
-                av_frame_free(&out);
-                return ret;
-            }
-
-            s->crossfade_samples(out->extended_data, cf[0]->extended_data,
-                                 cf[1]->extended_data,
-                                 s->nb_samples, out->ch_layout.nb_channels,
-                                 s->curve, s->curve2);
-            out->pts = s->pts;
-            s->pts += av_rescale_q(s->nb_samples,
-                (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
-            s->passthrough = 1;
-            av_frame_free(&cf[0]);
-            av_frame_free(&cf[1]);
-            return ff_filter_frame(outlink, out);
-        } else {
-            out = ff_get_audio_buffer(outlink, s->nb_samples);
-            if (!out)
-                return AVERROR(ENOMEM);
-
-            ret = ff_inlink_consume_samples(ctx->inputs[0], s->nb_samples, s->nb_samples, &cf[0]);
-            if (ret < 0) {
-                av_frame_free(&out);
-                return ret;
-            }
-
-            s->fade_samples(out->extended_data, cf[0]->extended_data, s->nb_samples,
-                            outlink->ch_layout.nb_channels, -1, s->nb_samples - 1, s->nb_samples, s->curve, 0., 1.);
-            out->pts = s->pts;
-            s->pts += av_rescale_q(s->nb_samples,
-                (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
-            av_frame_free(&cf[0]);
-            ret = ff_filter_frame(outlink, out);
-            if (ret < 0)
-                return ret;
-
-            out = ff_get_audio_buffer(outlink, s->nb_samples);
-            if (!out)
-                return AVERROR(ENOMEM);
-
-            ret = ff_inlink_consume_samples(ctx->inputs[1], s->nb_samples, s->nb_samples, &cf[1]);
-            if (ret < 0) {
-                av_frame_free(&out);
-                return ret;
-            }
-
-            s->fade_samples(out->extended_data, cf[1]->extended_data, s->nb_samples,
-                            outlink->ch_layout.nb_channels, 1, 0, s->nb_samples, s->curve2, 0., 1.);
-            out->pts = s->pts;
-            s->pts += av_rescale_q(s->nb_samples,
-                (AVRational){ 1, outlink->sample_rate }, outlink->time_base);
-            s->passthrough = 1;
-            av_frame_free(&cf[1]);
-            return ff_filter_frame(outlink, out);
-        }
+        return pass_crossfade(ctx);
     } else if (ff_outlink_frame_wanted(outlink)) {
         if (!s->status[0] && check_input(ctx->inputs[0]))
             s->status[0] = AVERROR_EOF;
-- 
2.43.0

_______________________________________________
ffmpeg-devel mailing list
ffmpeg-devel@ffmpeg.org
https://ffmpeg.org/mailman/listinfo/ffmpeg-devel

To unsubscribe, visit link above, or email
ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe".

  reply	other threads:[~2025-07-23 23:36 UTC|newest]

Thread overview: 4+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2025-07-23 23:35 [FFmpeg-devel] [PATCH 1/4] avfilter/trim: consume all available frames and avoid activate reschedule Marton Balint
2025-07-23 23:35 ` Marton Balint [this message]
2025-07-23 23:35 ` [FFmpeg-devel] [PATCH 3/4] avfilter/af_afade: fix check_input for empty streams Marton Balint
2025-07-23 23:36 ` [FFmpeg-devel] [PATCH 4/4] avfilter/af_afade: rework crossfade activate logic Marton Balint

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=20250723233604.29380-2-cus@passwd.hu \
    --to=cus@passwd.hu \
    --cc=ffmpeg-devel@ffmpeg.org \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel

This inbox may be cloned and mirrored by anyone:

	git clone --mirror https://master.gitmailbox.com/ffmpegdev/0 ffmpegdev/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 ffmpegdev ffmpegdev/ https://master.gitmailbox.com/ffmpegdev \
		ffmpegdev@gitmailbox.com
	public-inbox-index ffmpegdev

Example config snippet for mirrors.


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git