Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel
 help / color / mirror / Atom feed
From: "Martin Storsjö" <martin@martin.st>
To: FFmpeg development discussions and patches <ffmpeg-devel@ffmpeg.org>
Cc: Zhao Zhili <zhilizhao@tencent.com>
Subject: Re: [FFmpeg-devel] [PATCH 2/5] swscale/aarch64: Add rgb24 to yuv implementation
Date: Wed, 5 Jun 2024 11:16:24 +0300 (EEST)
Message-ID: <fd4ae3e-6d2d-5412-44ed-d0bfc50bc22@martin.st> (raw)
In-Reply-To: <tencent_C872E1B65BDE3479804A34C93E0156719C09@qq.com>

On Tue, 4 Jun 2024, Zhao Zhili wrote:

> From: Zhao Zhili <zhilizhao@tencent.com>
>
> Test on Apple M1:
>
> rgb24_to_uv_1080_c: 7.2
> rgb24_to_uv_1080_neon: 5.5
> rgb24_to_uv_1280_c: 8.2
> rgb24_to_uv_1280_neon: 6.2
> rgb24_to_uv_1920_c: 12.5
> rgb24_to_uv_1920_neon: 9.5
>
> rgb24_to_uv_half_540_c: 6.5
> rgb24_to_uv_half_540_neon: 3.0
> rgb24_to_uv_half_640_c: 7.5
> rgb24_to_uv_half_640_neon: 3.2
> rgb24_to_uv_half_960_c: 12.5
> rgb24_to_uv_half_960_neon: 6.0
>
> rgb24_to_y_1080_c: 4.5
> rgb24_to_y_1080_neon: 3.5
> rgb24_to_y_1280_c: 5.2
> rgb24_to_y_1280_neon: 4.2
> rgb24_to_y_1920_c: 8.0
> rgb24_to_y_1920_neon: 6.0
>
> Signed-off-by: Zhao Zhili <zhilizhao@tencent.com>
> ---
> libswscale/aarch64/Makefile  |   1 +
> libswscale/aarch64/input.S   | 229 +++++++++++++++++++++++++++++++++++
> libswscale/aarch64/swscale.c |  25 ++++
> 3 files changed, 255 insertions(+)
> create mode 100644 libswscale/aarch64/input.S
>
> diff --git a/libswscale/aarch64/Makefile b/libswscale/aarch64/Makefile
> index da1d909561..adfd90a1b6 100644
> --- a/libswscale/aarch64/Makefile
> +++ b/libswscale/aarch64/Makefile
> @@ -3,6 +3,7 @@ OBJS        += aarch64/rgb2rgb.o                \
>                aarch64/swscale_unscaled.o       \
>
> NEON-OBJS   += aarch64/hscale.o                 \
> +               aarch64/input.o                  \
>                aarch64/output.o                 \
>                aarch64/rgb2rgb_neon.o           \
>                aarch64/yuv2rgb_neon.o           \
> diff --git a/libswscale/aarch64/input.S b/libswscale/aarch64/input.S
> new file mode 100644
> index 0000000000..ee0d223c6e
> --- /dev/null
> +++ b/libswscale/aarch64/input.S
> @@ -0,0 +1,229 @@
> +/*
> + * Copyright (c) 2024 Zhao Zhili <quinkblack@foxmail.com>
> + *
> + * This file is part of FFmpeg.
> + *
> + * FFmpeg is free software; you can redistribute it and/or
> + * modify it under the terms of the GNU Lesser General Public
> + * License as published by the Free Software Foundation; either
> + * version 2.1 of the License, or (at your option) any later version.
> + *
> + * FFmpeg is distributed in the hope that it will be useful,
> + * but WITHOUT ANY WARRANTY; without even the implied warranty of
> + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
> + * Lesser General Public License for more details.
> + *
> + * You should have received a copy of the GNU Lesser General Public
> + * License along with FFmpeg; if not, write to the Free Software
> + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
> + */
> +
> +#include "libavutil/aarch64/asm.S"
> +
> +.macro rgb24_to_yuv_load_rgb, src
> +        ld3             { v16.16b, v17.16b, v18.16b }, [\src]
> +        ushll           v19.8h, v16.8b, #0         // v19: r
> +        ushll           v20.8h, v17.8b, #0         // v20: g
> +        ushll           v21.8h, v18.8b, #0         // v21: b
> +        ushll2          v22.8h, v16.16b, #0        // v22: r
> +        ushll2          v23.8h, v17.16b, #0        // v23: g
> +        ushll2          v24.8h, v18.16b, #0        // v24: b

Doing "ushll #0" is perhaps a bit unusual, the common thing would be 
"uxtl" instead. It doesn't matter in practice though, it assembles to the 
same instruction anyway.

> +.endm
> +
> +.macro rgb24_to_yuv_product, r, g, b, dst1, dst2, dst, coef0, coef1, coef2, right_shift
> +        mov             \dst1\().16b, v6.16b                    // dst1 = const_offset
> +        mov             \dst2\().16b, v6.16b                    // dst2 = const_offset
> +        smlal           \dst1\().4s, \coef0\().4h, \r\().4h     // dst1 += rx * r
> +        smlal2          \dst2\().4s, \coef0\().8h, \r\().8h     // dst2 += rx * r
> +        smlal           \dst1\().4s, \coef1\().4h, \g\().4h     // dst1 += gx * g
> +        smlal2          \dst2\().4s, \coef1\().8h, \g\().8h     // dst2 += gx * g
> +        smlal           \dst1\().4s, \coef2\().4h, \b\().4h     // dst1 += bx * b
> +        smlal2          \dst2\().4s, \coef2\().8h, \b\().8h     // dst2 += bx * b

For sequences like this, the Cortex A53 (and iirc at least the A55 too) 
has got a fastpath; if you do multiple consequent smlal/smlsl (or regular 
mla/mls) into the same register, you actually save a lot of time. E.g. 
instead of this:

     smlal dst1
     smlal dst2
     smlal dst1
     smlal dst2
     smlal dst1
     smlal dst2

Do this:

     smlal dst1
     smlal dst1
     smlal dst1
     smlal dst2
     smlal dst2
     smlal dst2

For in-order cores (with this special fastpath - it is indeed a bit 
non-obvious) this makes a huge difference, and for out of order cores, 
they can reorder it as they prefer anyway (as this is not a very long 
instruction sequence).

This makes a massive difference for the in-order cores.

Before:                  Cortex A53      A72      A73
rgb24_to_y_1920_neon:        6032.7   3385.7   2514.0
After:
rgb24_to_y_1920_neon:        5072.7   3388.2   2522.0

A 19% speedup on A53 with just with this one change, and it makes almost 
no difference for the other cores (mostly within measurement noise).

> +function ff_rgb24ToY_neon, export=1
> +        cmp             w4, #0                  // check width > 0
> +        b.le            4f
> +
> +        ldp             w10, w11, [x5], #8       // w10: ry, w11: gy
> +        dup             v0.8h, w10
> +        dup             v1.8h, w11
> +        ldr             w12, [x5]               // w12: by
> +        dup             v2.8h, w12
> +
> +        mov             w9, #256                // w9 = 1 << (RGB2YUV_SHIFT - 7)
> +        movk            w9, #8, lsl #16         // w9 += 32 << (RGB2YUV_SHIFT - 1)
> +        dup             v6.4s, w9               // w9: const_offset
> +
> +        mov             x2, #0                  // w2: i
> +        and             w3, w4, #0xFFFFFFF0     // w3 = width / 16 * 16
> +        cbz             w3, 3f

This blindly assumes that if width > 0, then width is also >= 16 at least, 
as we always run the SIMD codepath once here. Is this a requirement in 
swscale, or should we check whether width >= 16 here too?

> +1:
> +        rgb24_to_yuv_load_rgb x1
> +        rgb24_to_yuv_product v19, v20, v21, v25, v26, v16, v0, v1, v2, #9
> +        rgb24_to_yuv_product v22, v23, v24, v27, v28, v17, v0, v1, v2, #9
> +        stp             q16, q17, [x0], #32     // store to dst
> +
> +        add             w2, w2, #16             // i += 16
> +        add             x1, x1, #48             // src += 48
> +        cmp             w2, w3                  // i < (width / 16 * 16)
> +        b.lt            1b

For the in-order cores, we can help with scheduling here too. You can do 
the add/add/cmp before the stp. This adds a bit more distance between the 
calculation of q17 and the store of it, and adds more distance between the 
cmp and the b.lt that depends on it.

Secondly, it's a bit uncommon in SIMD like this, to count upwards 
(incrementing i); instead, the common pattern is to keep the width in one 
register and count it down.

Here, you can then do "cmp w2, #16", so you don't need to keep a register 
with the aligned end value. And for the scalar codepath, when approaching 
zero, you can do "subs w2, #1", "b.gt 2b", so you don't need two 
instructions for add+cmp.

> +        b               3f
> +2:
> +        ldrb            w13, [x1]               // w13: r
> +        ldrb            w14, [x1, #1]           // w14: g
> +        ldrb            w15, [x1, #2]           // w15: b
> +
> +        smaddl          x13, w13, w10, x9       // x13 = ry * r + const_offset
> +        smaddl          x13, w14, w11, x13      // x13 += gy * g
> +        smaddl          x13, w15, w12, x13      // x13 += by * b
> +        asr             w13, w13, #9            // x13 >>= 9
> +        strh            w13, [x0], #2           // store to dst
> +
> +        add             w2, w2, #1              // i++
> +        add             x1, x1, #3              // src += 3
> +3:
> +        cmp             w2, w4                  // i < width
> +        b.lt            2b
> +4:
> +        ret
> +endfunc
> +
> +.macro rgb24_load_uv_coeff half
> +        add             x6, x6, #12
> +
> +        ldp             w10, w11, [x6], #8      // w10: ru, w11: gu
> +        dup             v0.8h, w10
> +        dup             v1.8h, w11
> +
> +        ldp             w12, w13, [x6], #8      // w12: bu, w13: rv
> +        dup             v2.8h, w12
> +        dup             v3.8h, w13
> +
> +        ldp             w14, w15, [x6], #8      // w14: gv, w15: bv
> +        dup             v4.8h, w14
> +        dup             v5.8h, w15

As Remi mentioned, scheduling instructions like these can help a bit. But 
here, each "ldp" depends on the updated x6 from the previous one, so it 
maybe doesn't make much difference.

> +
> +    .if \half
> +        mov             w9, #512
> +        movk            w9, #128, lsl #16       // w9: const_offset
> +    .else
> +        mov             w9, #256
> +        movk            w9, #64, lsl #16        // w9: const_offset
> +    .endif
> +        dup             v6.4s, w9
> +.endm
> +
> +function ff_rgb24ToUV_half_neon, export=1
> +        cmp             w5, #0          // check width > 0
> +        b.le            4f
> +
> +        rgb24_load_uv_coeff half=1
> +
> +        mov             x9, #0                  // x9: i
> +        and             w7, w5, #0xFFFFFFF8     // w7 = width / 8 * 8
> +        cbz             w7, 3f
> +1:
> +        ld3             { v16.16b, v17.16b, v18.16b }, [x3]
> +        uaddlp          v19.8h, v16.16b         // v19: r
> +        uaddlp          v20.8h, v17.16b         // v20: g
> +        uaddlp          v21.8h, v18.16b         // v21: b
> +
> +        rgb24_to_yuv_product v19, v20, v21, v22, v23, v16, v0, v1, v2, #10
> +        str             q16, [x0], #16          // store dst_u
> +        rgb24_to_yuv_product v19, v20, v21, v24, v25, v17, v3, v4, v5, #10
> +        str             q17, [x1], #16          // store dst_v
> +
> +        add             w9, w9, #8              // i += 8
> +        add             x3, x3, #48             // src += 48
> +        cmp             w9, w7                  // i < (width * 8 / 8)
> +        b.lt            1b

Here, you can also move the both str out to after the add/cmp, like above.

> +        b               3f
> +2:
> +        ldrb            w2, [x3]                // w2: r1
> +        ldrb            w4, [x3, #3]            // w4: r2
> +        add             w2, w2, w4              // w2 = r1 + r2
> +
> +        ldrb            w4, [x3, #1]            // w4: g1
> +        ldrb            w7, [x3, #4]            // w7: g2
> +        add             w4, w4, w7              // w4 = g1 + g2
> +
> +        ldrb            w7, [x3, #2]            // w7: b1
> +        ldrb            w8, [x3, #5]            // w8: b2
> +        add             w7, w7, w8              // w7 = b1 + b2
> +
> +        umov            w8, v6.s[0]             // dst_u = const_offset
> +        smaddl          x8, w2, w10, x8         // dst_u += ru * r
> +        smaddl          x8, w4, w11, x8         // dst_u += gu * g
> +        smaddl          x8, w7, w12, x8         // dst_u += bu * b
> +        asr             x8, x8, #10             // dst_u >>= 10
> +        strh            w8, [x0], #2            // store dst_u
> +
> +        umov            w8, v6.s[0]             // dst_v = const_offset
> +        smaddl          x8, w2, w13, x8         // dst_v += rv * r
> +        smaddl          x8, w4, w14, x8         // dst_v += gv * g
> +        smaddl          x8, w7, w15, x8         // dst_v += bv * b
> +        asr             x8, x8, #10             // dst_v >>= 10
> +        strh            w8, [x1], #2            // store dst_v
> +
> +        add             w9, w9, #1              // i++
> +        add             x3, x3, #6              // src += 6
> +3:
> +        cmp             w9, w5
> +        b.lt            2b
> +4:
> +        ret
> +endfunc
> +
> +function ff_rgb24ToUV_neon, export=1
> +        cmp             w5, #0                  // check width > 0
> +        b.le            4f
> +
> +        rgb24_load_uv_coeff half=0
> +
> +        mov             x2, #0                  // w2: i
> +        and             w4, w5, #0xFFFFFFF0     // w4: width / 16 * 16
> +        cbz             w4, 3f
> +1:
> +        rgb24_to_yuv_load_rgb x3
> +        rgb24_to_yuv_product v19, v20, v21, v25, v26, v16, v0, v1, v2, #9
> +        rgb24_to_yuv_product v22, v23, v24, v27, v28, v17, v0, v1, v2, #9
> +        stp             q16, q17, [x0], #32      // store to dst_u
> +        rgb24_to_yuv_product v19, v20, v21, v25, v26, v16, v3, v4, v5, #9
> +        rgb24_to_yuv_product v22, v23, v24, v27, v28, v17, v3, v4, v5, #9
> +        stp             q16, q17, [x1], #32      // store to dst_v

If you'd make the second pair of rgb24_to_yuv_product write into v18/v19 
instead of v16/v17, you can move them out to after the add/add/cmp. (It 
doesn't make much of a measurable difference, but is good scheduling in 
general.)

> +
> +        add             w2, w2, #16             // i += 16
> +        add             x3, x3, #48             // src += 48
> +        cmp             w2, w4                  // i < (width / 16 * 16)
> +        b.lt            1b
> +        b               3f
> +2:
> +        ldrb            w16, [x3]               // w16: r
> +        ldrb            w17, [x3, #1]           // w17: g
> +        ldrb            w4, [x3, #2]            // w4: b
> +
> +        umov            w7, v6.s[0]            // w7 = const_offset

SIMD->GPR moves are generally expensive, we shouldn't be doing this 
within the loop, if the value is constant. Instead, you should do this in 
a prologue to the scalar codepath.

But in this case, the value should already be present in w9, if I 
understand the code correctly, so we don't need to fetch it from v6 - and 
this is already what you do in ff_rgb24ToY_neon.

// Martin

_______________________________________________
ffmpeg-devel mailing list
ffmpeg-devel@ffmpeg.org
https://ffmpeg.org/mailman/listinfo/ffmpeg-devel

To unsubscribe, visit link above, or email
ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe".

  parent reply	other threads:[~2024-06-05  8:16 UTC|newest]

Thread overview: 10+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
     [not found] <20240604135504.83169-1-quinkblack@foxmail.com>
2024-06-04 13:55 ` Zhao Zhili
2024-06-05  6:29   ` Rémi Denis-Courmont
2024-06-05  6:53     ` Zhao Zhili
2024-06-05  7:34       ` Rémi Denis-Courmont
2024-06-05  7:34       ` Martin Storsjö
2024-06-05  8:16   ` Martin Storsjö [this message]
2024-06-04 13:55 ` [FFmpeg-devel] [PATCH 3/5] avutil/aarch64: Skip define AV_READ_TIME for apple Zhao Zhili
2024-06-04 13:55 ` [FFmpeg-devel] [PATCH 4/5] avutil/timer: Add clock_gettime as a fallback of AV_READ_TIME Zhao Zhili
2024-06-04 13:55 ` [FFmpeg-devel] [PATCH 5/5] avutil/aarch64: Fallback to clock_gettime as timer on Android Zhao Zhili
2024-06-04 20:25   ` Martin Storsjö

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=fd4ae3e-6d2d-5412-44ed-d0bfc50bc22@martin.st \
    --to=martin@martin.st \
    --cc=ffmpeg-devel@ffmpeg.org \
    --cc=zhilizhao@tencent.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link

Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel

This inbox may be cloned and mirrored by anyone:

	git clone --mirror https://master.gitmailbox.com/ffmpegdev/0 ffmpegdev/git/0.git

	# If you have public-inbox 1.1+ installed, you may
	# initialize and index your mirror using the following commands:
	public-inbox-init -V2 ffmpegdev ffmpegdev/ https://master.gitmailbox.com/ffmpegdev \
		ffmpegdev@gitmailbox.com
	public-inbox-index ffmpegdev

Example config snippet for mirrors.


AGPL code for this site: git clone https://public-inbox.org/public-inbox.git