From: "Martin Storsjö" <martin@martin.st> To: ffmpeg-devel@ffmpeg.org Subject: [FFmpeg-devel] [PATCH 3/4] checkasm: hevc_pel: Use helpers for checking for writes out of bounds Date: Wed, 26 Mar 2025 12:30:14 +0200 Message-ID: <20250326103015.44296-3-martin@martin.st> (raw) In-Reply-To: <20250326103015.44296-1-martin@martin.st> This allows catching whether the functions write outside of the designated rectangle, and if run with "checkasm -v", it also prints out on which side of the rectangle the overwrite was. --- tests/checkasm/hevc_pel.c | 157 ++++++++++++++++++++++---------------- 1 file changed, 90 insertions(+), 67 deletions(-) diff --git a/tests/checkasm/hevc_pel.c b/tests/checkasm/hevc_pel.c index b9417832e5..9a83613915 100644 --- a/tests/checkasm/hevc_pel.c +++ b/tests/checkasm/hevc_pel.c @@ -44,9 +44,14 @@ static const int offsets[] = {0, 255, -1 }; uint32_t r = rnd() & mask; \ AV_WN32A(buf0 + k, r); \ AV_WN32A(buf1 + k, r); \ - if (k >= BUF_SIZE) \ - continue; \ - r = rnd(); \ + } \ + } while (0) + +#define randomize_buffers_dst() \ + do { \ + int k; \ + for (k = 0; k < BUF_SIZE; k += 4) { \ + uint32_t r = rnd(); \ AV_WN32A(dst0 + k, r); \ AV_WN32A(dst1 + k, r); \ } \ @@ -100,6 +105,7 @@ static void checkasm_check_hevc_qpel(void) "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { int16_t *dstw0 = (int16_t *) dst0, *dstw1 = (int16_t *) dst1; randomize_buffers(); + randomize_buffers_dst(); call_ref(dstw0, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); call_new(dstw1, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); checkasm_check(int16_t, dstw0, MAX_PB_SIZE * sizeof(int16_t), @@ -118,8 +124,8 @@ static void checkasm_check_hevc_qpel_uni(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE + SRC_EXTRA]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE + SRC_EXTRA]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); HEVCDSPContext h; int size, bit_depth, i, j; @@ -143,16 +149,18 @@ static void checkasm_check_hevc_qpel_uni(void) if (check_func(h.put_hevc_qpel_uni[size][j][i], "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { randomize_buffers(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); } @@ -167,8 +175,8 @@ static void checkasm_check_hevc_qpel_uni_w(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE + SRC_EXTRA]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE + SRC_EXTRA]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); HEVCDSPContext h; int size, bit_depth, i, j; @@ -196,16 +204,18 @@ static void checkasm_check_hevc_qpel_uni_w(void) for (wx = weights; *wx >= 0; wx++) { for (ox = offsets; *ox >= 0; ox++) { randomize_buffers(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); } @@ -223,8 +233,8 @@ static void checkasm_check_hevc_qpel_bi(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE + SRC_EXTRA]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE + SRC_EXTRA]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); LOCAL_ALIGNED_32(int16_t, ref0, [BUF_SIZE]); LOCAL_ALIGNED_32(int16_t, ref1, [BUF_SIZE]); @@ -251,16 +261,18 @@ static void checkasm_check_hevc_qpel_bi(void) if (check_func(h.put_hevc_qpel_bi[size][j][i], "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { randomize_buffers_ref(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, ref0, sizes[size], i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], i, j, sizes[size]); } @@ -275,8 +287,8 @@ static void checkasm_check_hevc_qpel_bi_w(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE + SRC_EXTRA]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE + SRC_EXTRA]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); LOCAL_ALIGNED_32(int16_t, ref0, [BUF_SIZE]); LOCAL_ALIGNED_32(int16_t, ref1, [BUF_SIZE]); @@ -308,16 +320,18 @@ static void checkasm_check_hevc_qpel_bi_w(void) for (wx = weights; *wx >= 0; wx++) { for (ox = offsets; *ox >= 0; ox++) { randomize_buffers_ref(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, ref0, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); } @@ -364,6 +378,7 @@ static void checkasm_check_hevc_epel(void) "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { int16_t *dstw0 = (int16_t *) dst0, *dstw1 = (int16_t *) dst1; randomize_buffers(); + randomize_buffers_dst(); call_ref(dstw0, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); call_new(dstw1, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); checkasm_check(int16_t, dstw0, MAX_PB_SIZE * sizeof(int16_t), @@ -382,8 +397,8 @@ static void checkasm_check_hevc_epel_uni(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); HEVCDSPContext h; int size, bit_depth, i, j; @@ -407,16 +422,18 @@ static void checkasm_check_hevc_epel_uni(void) if (check_func(h.put_hevc_epel_uni[size][j][i], "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { randomize_buffers(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], i, j, sizes[size]); } @@ -431,8 +448,8 @@ static void checkasm_check_hevc_epel_uni_w(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); HEVCDSPContext h; int size, bit_depth, i, j; @@ -460,16 +477,18 @@ static void checkasm_check_hevc_epel_uni_w(void) for (wx = weights; *wx >= 0; wx++) { for (ox = offsets; *ox >= 0; ox++) { randomize_buffers(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, sizes[size], *denom, *wx, *ox, i, j, sizes[size]); } @@ -487,8 +506,8 @@ static void checkasm_check_hevc_epel_bi(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); LOCAL_ALIGNED_32(int16_t, ref0, [BUF_SIZE]); LOCAL_ALIGNED_32(int16_t, ref1, [BUF_SIZE]); @@ -515,16 +534,18 @@ static void checkasm_check_hevc_epel_bi(void) if (check_func(h.put_hevc_epel_bi[size][j][i], "put_hevc_%s%d_%d", type, sizes[size], bit_depth)) { randomize_buffers_ref(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, ref0, sizes[size], i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], i, j, sizes[size]); } @@ -539,8 +560,8 @@ static void checkasm_check_hevc_epel_bi_w(void) { LOCAL_ALIGNED_32(uint8_t, buf0, [BUF_SIZE]); LOCAL_ALIGNED_32(uint8_t, buf1, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst0, [BUF_SIZE]); - LOCAL_ALIGNED_32(uint8_t, dst1, [BUF_SIZE]); + PIXEL_RECT(dst0, 64, 64); + PIXEL_RECT(dst1, 64, 64); LOCAL_ALIGNED_32(int16_t, ref0, [BUF_SIZE]); LOCAL_ALIGNED_32(int16_t, ref1, [BUF_SIZE]); @@ -572,16 +593,18 @@ static void checkasm_check_hevc_epel_bi_w(void) for (wx = weights; *wx >= 0; wx++) { for (ox = offsets; *ox >= 0; ox++) { randomize_buffers_ref(); - call_ref(dst0, sizes[size] * SIZEOF_PIXEL, + CLEAR_PIXEL_RECT(dst0); + CLEAR_PIXEL_RECT(dst1); + call_ref(dst0, dst0_stride, src0, sizes[size] * SIZEOF_PIXEL, ref0, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); - call_new(dst1, sizes[size] * SIZEOF_PIXEL, + call_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); - checkasm_check_pixel(dst0, sizes[size] * SIZEOF_PIXEL, - dst1, sizes[size] * SIZEOF_PIXEL, - size[sizes], size[sizes], "dst"); - bench_new(dst1, sizes[size] * SIZEOF_PIXEL, + checkasm_check_pixel_padded(dst0, dst0_stride, + dst1, dst1_stride, + size[sizes], size[sizes], "dst"); + bench_new(dst1, dst1_stride, src1, sizes[size] * SIZEOF_PIXEL, ref1, sizes[size], *denom, *wx, *wx, *ox, *ox, i, j, sizes[size]); } -- 2.39.5 (Apple Git-154) _______________________________________________ ffmpeg-devel mailing list ffmpeg-devel@ffmpeg.org https://ffmpeg.org/mailman/listinfo/ffmpeg-devel To unsubscribe, visit link above, or email ffmpeg-devel-request@ffmpeg.org with subject "unsubscribe".
next prev parent reply other threads:[~2025-03-26 10:30 UTC|newest] Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top 2025-03-26 10:30 [FFmpeg-devel] [PATCH 1/4] checkasm: Make checkasm_fail_func return whether we should print verbosely Martin Storsjö 2025-03-26 10:30 ` [FFmpeg-devel] [PATCH 2/4] checkasm: Implement helpers for defining and checking padded rects Martin Storsjö 2025-03-29 0:15 ` Michael Niedermayer 2025-03-31 13:05 ` Martin Storsjö 2025-04-01 0:28 ` Michael Niedermayer 2025-04-01 9:38 ` Martin Storsjö 2025-04-01 15:38 ` Martin Storsjö 2025-03-26 10:30 ` Martin Storsjö [this message] 2025-04-01 9:35 ` [FFmpeg-devel] [PATCH 3/4] checkasm: hevc_pel: Use helpers for checking for writes out of bounds Martin Storsjö 2025-04-01 12:54 ` yinshiyou-hf 2025-03-26 10:30 ` [FFmpeg-devel] [PATCH 4/4] checkasm: vp8dsp: Use checkasm_check_padded in check_mc Martin Storsjö
Reply instructions: You may reply publicly to this message via plain-text email using any one of the following methods: * Save the following mbox file, import it into your mail client, and reply-to-all from there: mbox Avoid top-posting and favor interleaved quoting: https://en.wikipedia.org/wiki/Posting_style#Interleaved_style * Reply using the --to, --cc, and --in-reply-to switches of git-send-email(1): git send-email \ --in-reply-to=20250326103015.44296-3-martin@martin.st \ --to=martin@martin.st \ --cc=ffmpeg-devel@ffmpeg.org \ /path/to/YOUR_REPLY https://kernel.org/pub/software/scm/git/docs/git-send-email.html * If your mail client supports setting the In-Reply-To header via mailto: links, try the mailto: link
Git Inbox Mirror of the ffmpeg-devel mailing list - see https://ffmpeg.org/mailman/listinfo/ffmpeg-devel This inbox may be cloned and mirrored by anyone: git clone --mirror https://master.gitmailbox.com/ffmpegdev/0 ffmpegdev/git/0.git # If you have public-inbox 1.1+ installed, you may # initialize and index your mirror using the following commands: public-inbox-init -V2 ffmpegdev ffmpegdev/ https://master.gitmailbox.com/ffmpegdev \ ffmpegdev@gitmailbox.com public-inbox-index ffmpegdev Example config snippet for mirrors. AGPL code for this site: git clone https://public-inbox.org/public-inbox.git