diff options
Diffstat (limited to 'tests/checkasm/vp9dsp.c')
-rw-r--r-- | tests/checkasm/vp9dsp.c | 609 |
1 files changed, 477 insertions, 132 deletions
diff --git a/tests/checkasm/vp9dsp.c b/tests/checkasm/vp9dsp.c index f0d93725eb..a91577b94e 100644 --- a/tests/checkasm/vp9dsp.c +++ b/tests/checkasm/vp9dsp.c @@ -1,56 +1,399 @@ /* * Copyright (c) 2015 Ronald S. Bultje <rsbultje@gmail.com> * - * This file is part of Libav. + * This file is part of FFmpeg. * - * Libav is free software; you can redistribute it and/or modify + * FFmpeg is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * - * Libav is distributed in the hope that it will be useful, + * FFmpeg is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License along - * with Libav; if not, write to the Free Software Foundation, Inc., + * with FFmpeg; if not, write to the Free Software Foundation, Inc., * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. */ +#include <math.h> #include <string.h> - +#include "checkasm.h" +#include "libavcodec/vp9data.h" +#include "libavcodec/vp9.h" #include "libavutil/common.h" #include "libavutil/internal.h" #include "libavutil/intreadwrite.h" +#include "libavutil/mathematics.h" -#include "libavcodec/vp9.h" +static const uint32_t pixel_mask[3] = { 0xffffffff, 0x03ff03ff, 0x0fff0fff }; +#define SIZEOF_PIXEL ((bit_depth + 7) / 8) -#include "checkasm.h" +#define randomize_buffers() \ + do { \ + uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \ + int k; \ + for (k = -4; k < SIZEOF_PIXEL * FFMAX(8, size); k += 4) { \ + uint32_t r = rnd() & mask; \ + AV_WN32A(a + k, r); \ + } \ + for (k = 0; k < size * SIZEOF_PIXEL; k += 4) { \ + uint32_t r = rnd() & mask; \ + AV_WN32A(l + k, r); \ + } \ + } while (0) -static const uint32_t pixel_mask[3] = { 0xffffffff, 0x03ff03ff, 0x0fff0fff }; +static void check_ipred(void) +{ + LOCAL_ALIGNED_32(uint8_t, a_buf, [64 * 2]); + uint8_t *a = &a_buf[32 * 2]; + LOCAL_ALIGNED_32(uint8_t, l, [32 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst0, [32 * 32 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst1, [32 * 32 * 2]); + VP9DSPContext dsp; + int tx, mode, bit_depth; + declare_func_emms(AV_CPU_FLAG_MMX | AV_CPU_FLAG_MMXEXT, void, uint8_t *dst, ptrdiff_t stride, + const uint8_t *left, const uint8_t *top); + static const char *const mode_names[N_INTRA_PRED_MODES] = { + [VERT_PRED] = "vert", + [HOR_PRED] = "hor", + [DC_PRED] = "dc", + [DIAG_DOWN_LEFT_PRED] = "diag_downleft", + [DIAG_DOWN_RIGHT_PRED] = "diag_downright", + [VERT_RIGHT_PRED] = "vert_right", + [HOR_DOWN_PRED] = "hor_down", + [VERT_LEFT_PRED] = "vert_left", + [HOR_UP_PRED] = "hor_up", + [TM_VP8_PRED] = "tm", + [LEFT_DC_PRED] = "dc_left", + [TOP_DC_PRED] = "dc_top", + [DC_128_PRED] = "dc_128", + [DC_127_PRED] = "dc_127", + [DC_129_PRED] = "dc_129", + }; -#define BIT_DEPTH 8 -#define SIZEOF_PIXEL ((BIT_DEPTH + 7) / 8) + for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) { + ff_vp9dsp_init(&dsp, bit_depth, 0); + for (tx = 0; tx < 4; tx++) { + int size = 4 << tx; + + for (mode = 0; mode < N_INTRA_PRED_MODES; mode++) { + if (check_func(dsp.intra_pred[tx][mode], "vp9_%s_%dx%d_%dbpp", + mode_names[mode], size, size, bit_depth)) { + randomize_buffers(); + call_ref(dst0, size * SIZEOF_PIXEL, l, a); + call_new(dst1, size * SIZEOF_PIXEL, l, a); + if (memcmp(dst0, dst1, size * size * SIZEOF_PIXEL)) + fail(); + bench_new(dst1, size * SIZEOF_PIXEL,l, a); + } + } + } + } + report("ipred"); +} + +#undef randomize_buffers + +#define randomize_buffers() \ + do { \ + uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \ + for (y = 0; y < sz; y++) { \ + for (x = 0; x < sz * SIZEOF_PIXEL; x += 4) { \ + uint32_t r = rnd() & mask; \ + AV_WN32A(dst + y * sz * SIZEOF_PIXEL + x, r); \ + AV_WN32A(src + y * sz * SIZEOF_PIXEL + x, rnd() & mask); \ + } \ + for (x = 0; x < sz; x++) { \ + if (bit_depth == 8) { \ + coef[y * sz + x] = src[y * sz + x] - dst[y * sz + x]; \ + } else { \ + ((int32_t *) coef)[y * sz + x] = \ + ((uint16_t *) src)[y * sz + x] - \ + ((uint16_t *) dst)[y * sz + x]; \ + } \ + } \ + } \ + } while(0) + +// wht function copied from libvpx +static void fwht_1d(double *out, const double *in, int sz) +{ + double t0 = in[0] + in[1]; + double t3 = in[3] - in[2]; + double t4 = trunc((t0 - t3) * 0.5); + double t1 = t4 - in[1]; + double t2 = t4 - in[2]; + + out[0] = t0 - t2; + out[1] = t2; + out[2] = t3 + t1; + out[3] = t1; +} + +// standard DCT-II +static void fdct_1d(double *out, const double *in, int sz) +{ + int k, n; + + for (k = 0; k < sz; k++) { + out[k] = 0.0; + for (n = 0; n < sz; n++) + out[k] += in[n] * cos(M_PI * (2 * n + 1) * k / (sz * 2.0)); + } + out[0] *= M_SQRT1_2; +} + +// see "Towards jointly optimal spatial prediction and adaptive transform in +// video/image coding", by J. Han, A. Saxena, and K. Rose +// IEEE Proc. ICASSP, pp. 726-729, Mar. 2010. +static void fadst4_1d(double *out, const double *in, int sz) +{ + int k, n; + + for (k = 0; k < sz; k++) { + out[k] = 0.0; + for (n = 0; n < sz; n++) + out[k] += in[n] * sin(M_PI * (n + 1) * (2 * k + 1) / (sz * 2.0 + 1.0)); + } +} + +// see "A Butterfly Structured Design of The Hybrid Transform Coding Scheme", +// by Jingning Han, Yaowu Xu, and Debargha Mukherjee +// http://static.googleusercontent.com/media/research.google.com/en//pubs/archive/41418.pdf +static void fadst_1d(double *out, const double *in, int sz) +{ + int k, n; + + for (k = 0; k < sz; k++) { + out[k] = 0.0; + for (n = 0; n < sz; n++) + out[k] += in[n] * sin(M_PI * (2 * n + 1) * (2 * k + 1) / (sz * 4.0)); + } +} + +typedef void (*ftx1d_fn)(double *out, const double *in, int sz); +static void ftx_2d(double *out, const double *in, enum TxfmMode tx, + enum TxfmType txtp, int sz) +{ + static const double scaling_factors[5][4] = { + { 4.0, 16.0 * M_SQRT1_2 / 3.0, 16.0 * M_SQRT1_2 / 3.0, 32.0 / 9.0 }, + { 2.0, 2.0, 2.0, 2.0 }, + { 1.0, 1.0, 1.0, 1.0 }, + { 0.25 }, + { 4.0 } + }; + static const ftx1d_fn ftx1d_tbl[5][4][2] = { + { + { fdct_1d, fdct_1d }, + { fadst4_1d, fdct_1d }, + { fdct_1d, fadst4_1d }, + { fadst4_1d, fadst4_1d }, + }, { + { fdct_1d, fdct_1d }, + { fadst_1d, fdct_1d }, + { fdct_1d, fadst_1d }, + { fadst_1d, fadst_1d }, + }, { + { fdct_1d, fdct_1d }, + { fadst_1d, fdct_1d }, + { fdct_1d, fadst_1d }, + { fadst_1d, fadst_1d }, + }, { + { fdct_1d, fdct_1d }, + }, { + { fwht_1d, fwht_1d }, + }, + }; + double temp[1024]; + double scaling_factor = scaling_factors[tx][txtp]; + int i, j; + + // cols + for (i = 0; i < sz; ++i) { + double temp_out[32]; + + ftx1d_tbl[tx][txtp][0](temp_out, &in[i * sz], sz); + // scale and transpose + for (j = 0; j < sz; ++j) + temp[j * sz + i] = temp_out[j] * scaling_factor; + } + + // rows + for (i = 0; i < sz; i++) + ftx1d_tbl[tx][txtp][1](&out[i * sz], &temp[i * sz], sz); +} + +static void ftx(int16_t *buf, enum TxfmMode tx, + enum TxfmType txtp, int sz, int bit_depth) +{ + double ind[1024], outd[1024]; + int n; + + emms_c(); + for (n = 0; n < sz * sz; n++) { + if (bit_depth == 8) + ind[n] = buf[n]; + else + ind[n] = ((int32_t *) buf)[n]; + } + ftx_2d(outd, ind, tx, txtp, sz); + for (n = 0; n < sz * sz; n++) { + if (bit_depth == 8) + buf[n] = lrint(outd[n]); + else + ((int32_t *) buf)[n] = lrint(outd[n]); + } +} + +static int copy_subcoefs(int16_t *out, const int16_t *in, enum TxfmMode tx, + enum TxfmType txtp, int sz, int sub, int bit_depth) +{ + // copy the topleft coefficients such that the return value (being the + // coefficient scantable index for the eob token) guarantees that only + // the topleft $sub out of $sz (where $sz >= $sub) coefficients in both + // dimensions are non-zero. This leads to braching to specific optimized + // simd versions (e.g. dc-only) so that we get full asm coverage in this + // test + + int n; + const int16_t *scan = ff_vp9_scans[tx][txtp]; + int eob; + + for (n = 0; n < sz * sz; n++) { + int rc = scan[n], rcx = rc % sz, rcy = rc / sz; + + // find eob for this sub-idct + if (rcx >= sub || rcy >= sub) + break; + + // copy coef + if (bit_depth == 8) { + out[rc] = in[rc]; + } else { + AV_COPY32(&out[rc * 2], &in[rc * 2]); + } + } + + eob = n; + + for (; n < sz * sz; n++) { + int rc = scan[n]; + + // zero + if (bit_depth == 8) { + out[rc] = 0; + } else { + AV_ZERO32(&out[rc * 2]); + } + } + + return eob; +} + +static int iszero(const int16_t *c, int sz) +{ + int n; + + for (n = 0; n < sz / sizeof(int16_t); n += 2) + if (AV_RN32A(&c[n])) + return 0; + + return 1; +} + +#define SIZEOF_COEF (2 * ((bit_depth + 7) / 8)) + +static void check_itxfm(void) +{ + LOCAL_ALIGNED_32(uint8_t, src, [32 * 32 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst, [32 * 32 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst0, [32 * 32 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst1, [32 * 32 * 2]); + LOCAL_ALIGNED_32(int16_t, coef, [32 * 32 * 2]); + LOCAL_ALIGNED_32(int16_t, subcoef0, [32 * 32 * 2]); + LOCAL_ALIGNED_32(int16_t, subcoef1, [32 * 32 * 2]); + declare_func_emms(AV_CPU_FLAG_MMX | AV_CPU_FLAG_MMXEXT, void, uint8_t *dst, ptrdiff_t stride, int16_t *block, int eob); + VP9DSPContext dsp; + int y, x, tx, txtp, bit_depth, sub; + static const char *const txtp_types[N_TXFM_TYPES] = { + [DCT_DCT] = "dct_dct", [DCT_ADST] = "adst_dct", + [ADST_DCT] = "dct_adst", [ADST_ADST] = "adst_adst" + }; + + for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) { + ff_vp9dsp_init(&dsp, bit_depth, 0); + + for (tx = TX_4X4; tx <= N_TXFM_SIZES /* 4 = lossless */; tx++) { + int sz = 4 << (tx & 3); + int n_txtps = tx < TX_32X32 ? N_TXFM_TYPES : 1; + + for (txtp = 0; txtp < n_txtps; txtp++) { + // skip testing sub-IDCTs for WHT or ADST since they don't + // implement it in any of the SIMD functions. If they do, + // consider changing this to ensure we have complete test + // coverage. Test sub=1 for dc-only, then 2, 4, 8, 12, etc, + // since the arm version can distinguish them at that level. + for (sub = (txtp == 0 && tx < 4) ? 1 : sz; sub <= sz; + sub < 4 ? (sub <<= 1) : (sub += 4)) { + if (check_func(dsp.itxfm_add[tx][txtp], + "vp9_inv_%s_%dx%d_sub%d_add_%d", + tx == 4 ? "wht_wht" : txtp_types[txtp], + sz, sz, sub, bit_depth)) { + int eob; + + randomize_buffers(); + ftx(coef, tx, txtp, sz, bit_depth); + + if (sub < sz) { + eob = copy_subcoefs(subcoef0, coef, tx, txtp, + sz, sub, bit_depth); + } else { + eob = sz * sz; + memcpy(subcoef0, coef, sz * sz * SIZEOF_COEF); + } + + memcpy(dst0, dst, sz * sz * SIZEOF_PIXEL); + memcpy(dst1, dst, sz * sz * SIZEOF_PIXEL); + memcpy(subcoef1, subcoef0, sz * sz * SIZEOF_COEF); + call_ref(dst0, sz * SIZEOF_PIXEL, subcoef0, eob); + call_new(dst1, sz * SIZEOF_PIXEL, subcoef1, eob); + if (memcmp(dst0, dst1, sz * sz * SIZEOF_PIXEL) || + !iszero(subcoef0, sz * sz * SIZEOF_COEF) || + !iszero(subcoef1, sz * sz * SIZEOF_COEF)) + fail(); + + bench_new(dst, sz * SIZEOF_PIXEL, coef, eob); + } + } + } + } + } + report("itxfm"); +} + +#undef randomize_buffers #define setpx(a,b,c) \ do { \ if (SIZEOF_PIXEL == 1) { \ buf0[(a) + (b) * jstride] = av_clip_uint8(c); \ } else { \ - ((uint16_t *)buf0)[(a) + (b) * jstride] = av_clip_uintp2(c, BIT_DEPTH); \ + ((uint16_t *)buf0)[(a) + (b) * jstride] = av_clip_uintp2(c, bit_depth); \ } \ } while (0) -#define setdx(a,b,c,d) setpx(a,b,c-(d)+(rnd()%((d)*2+1))) -#define setsx(a,b,c,d) setdx(a,b,c,(d) << (BIT_DEPTH - 8)) +// c can be an assignment and must not be put under () +#define setdx(a,b,c,d) setpx(a,b,c-(d)+(rnd()%((d)*2+1))) +#define setsx(a,b,c,d) setdx(a,b,c,(d) << (bit_depth - 8)) static void randomize_loopfilter_buffers(int bidx, int lineoff, int str, - int bit_depth, int dir, - const int *E, const int *F, - const int *H, const int *I, + int bit_depth, int dir, const int *E, + const int *F, const int *H, const int *I, uint8_t *buf0, uint8_t *buf1) { - uint32_t mask = (1 << BIT_DEPTH) - 1; + uint32_t mask = (1 << bit_depth) - 1; int off = dir ? lineoff : lineoff * 16; int istride = dir ? 1 : 16; int jstride = dir ? str : 1; @@ -100,84 +443,83 @@ static void randomize_loopfilter_buffers(int bidx, int lineoff, int str, } } } - -#define randomize_buffers(bidx, lineoff, str) \ - randomize_loopfilter_buffers(bidx, lineoff, str, BIT_DEPTH, dir, \ - E, F, H, I, buf0, buf1) +#define randomize_buffers(bidx, lineoff, str) \ + randomize_loopfilter_buffers(bidx, lineoff, str, bit_depth, dir, \ + E, F, H, I, buf0, buf1) static void check_loopfilter(void) { LOCAL_ALIGNED_32(uint8_t, base0, [32 + 16 * 16 * 2]); LOCAL_ALIGNED_32(uint8_t, base1, [32 + 16 * 16 * 2]); VP9DSPContext dsp; - int dir, wd, wd2; + int dir, wd, wd2, bit_depth; static const char *const dir_name[2] = { "h", "v" }; static const int E[2] = { 20, 28 }, I[2] = { 10, 16 }; - static const int H[2] = { 7, 11 }, F[2] = { 1, 1 }; - declare_func(void, uint8_t *dst, ptrdiff_t stride, int E, int I, int H); - - ff_vp9dsp_init(&dsp); - - for (dir = 0; dir < 2; dir++) { - uint8_t *buf0, *buf1; - int midoff = (dir ? 8 * 8 : 8) * SIZEOF_PIXEL; - int midoff_aligned = (dir ? 8 * 8 : 16) * SIZEOF_PIXEL; - - buf0 = base0 + midoff_aligned; - buf1 = base1 + midoff_aligned; - - for (wd = 0; wd < 3; wd++) { - // 4/8/16wd_8px - if (check_func(dsp.loop_filter_8[wd][dir], - "vp9_loop_filter_%s_%d_8", - dir_name[dir], 4 << wd)) { - randomize_buffers(0, 0, 8); - memcpy(buf1 - midoff, buf0 - midoff, - 16 * 8 * SIZEOF_PIXEL); - call_ref(buf0, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); - call_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); - if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 8 * SIZEOF_PIXEL)) - fail(); - bench_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); - } - } + static const int H[2] = { 7, 11 }, F[2] = { 1, 1 }; + declare_func_emms(AV_CPU_FLAG_MMX | AV_CPU_FLAG_MMXEXT, void, uint8_t *dst, ptrdiff_t stride, int E, int I, int H); - midoff = (dir ? 16 * 8 : 8) * SIZEOF_PIXEL; - midoff_aligned = (dir ? 16 * 8 : 16) * SIZEOF_PIXEL; - - buf0 = base0 + midoff_aligned; - buf1 = base1 + midoff_aligned; - - // 16wd_16px loopfilter - if (check_func(dsp.loop_filter_16[dir], - "vp9_loop_filter_%s_16_16", - dir_name[dir])) { - randomize_buffers(0, 0, 16); - randomize_buffers(0, 8, 16); - memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL); - call_ref(buf0, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); - call_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); - if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL)) - fail(); - bench_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); - } + for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) { + ff_vp9dsp_init(&dsp, bit_depth, 0); + + for (dir = 0; dir < 2; dir++) { + int midoff = (dir ? 8 * 8 : 8) * SIZEOF_PIXEL; + int midoff_aligned = (dir ? 8 * 8 : 16) * SIZEOF_PIXEL; + uint8_t *buf0 = base0 + midoff_aligned; + uint8_t *buf1 = base1 + midoff_aligned; - for (wd = 0; wd < 2; wd++) { - for (wd2 = 0; wd2 < 2; wd2++) { - // mix2 loopfilter - if (check_func(dsp.loop_filter_mix2[wd][wd2][dir], - "vp9_loop_filter_mix2_%s_%d%d_16", - dir_name[dir], 4 << wd, 4 << wd2)) { - randomize_buffers(0, 0, 16); - randomize_buffers(1, 8, 16); - memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL); -#define M(a) ((a[1] << 8) | a[0]) - call_ref(buf0, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); - call_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); - if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL)) + for (wd = 0; wd < 3; wd++) { + // 4/8/16wd_8px + if (check_func(dsp.loop_filter_8[wd][dir], + "vp9_loop_filter_%s_%d_8_%dbpp", + dir_name[dir], 4 << wd, bit_depth)) { + randomize_buffers(0, 0, 8); + memcpy(buf1 - midoff, buf0 - midoff, + 16 * 8 * SIZEOF_PIXEL); + call_ref(buf0, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); + call_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); + if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 8 * SIZEOF_PIXEL)) fail(); - bench_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); + bench_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]); + } + } + + midoff = (dir ? 16 * 8 : 8) * SIZEOF_PIXEL; + midoff_aligned = (dir ? 16 * 8 : 16) * SIZEOF_PIXEL; + + buf0 = base0 + midoff_aligned; + buf1 = base1 + midoff_aligned; + + // 16wd_16px loopfilter + if (check_func(dsp.loop_filter_16[dir], + "vp9_loop_filter_%s_16_16_%dbpp", + dir_name[dir], bit_depth)) { + randomize_buffers(0, 0, 16); + randomize_buffers(0, 8, 16); + memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL); + call_ref(buf0, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); + call_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); + if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL)) + fail(); + bench_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]); + } + + for (wd = 0; wd < 2; wd++) { + for (wd2 = 0; wd2 < 2; wd2++) { + // mix2 loopfilter + if (check_func(dsp.loop_filter_mix2[wd][wd2][dir], + "vp9_loop_filter_mix2_%s_%d%d_16_%dbpp", + dir_name[dir], 4 << wd, 4 << wd2, bit_depth)) { + randomize_buffers(0, 0, 16); + randomize_buffers(1, 8, 16); + memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL); +#define M(a) (((a)[1] << 8) | (a)[0]) + call_ref(buf0, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); + call_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); + if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL)) + fail(); + bench_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H)); #undef M + } } } } @@ -197,7 +539,7 @@ static void check_loopfilter(void) #define randomize_buffers() \ do { \ - uint32_t mask = pixel_mask[(BIT_DEPTH - 8) >> 1]; \ + uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \ int k; \ for (k = 0; k < SRC_BUF_SIZE; k += 4) { \ uint32_t r = rnd() & mask; \ @@ -214,62 +556,63 @@ static void check_loopfilter(void) static void check_mc(void) { + LOCAL_ALIGNED_32(uint8_t, buf, [72 * 72 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst0, [64 * 64 * 2]); + LOCAL_ALIGNED_32(uint8_t, dst1, [64 * 64 * 2]); + VP9DSPContext dsp; + int op, hsize, bit_depth, filter, dx, dy; + declare_func_emms(AV_CPU_FLAG_MMX | AV_CPU_FLAG_MMXEXT, void, uint8_t *dst, ptrdiff_t dst_stride, + const uint8_t *ref, ptrdiff_t ref_stride, + int h, int mx, int my); static const char *const filter_names[4] = { "8tap_smooth", "8tap_regular", "8tap_sharp", "bilin" }; static const char *const subpel_names[2][2] = { { "", "h" }, { "v", "hv" } }; static const char *const op_names[2] = { "put", "avg" }; - - LOCAL_ALIGNED_32(uint8_t, buf, [72 * 72 * 2]); - LOCAL_ALIGNED_32(uint8_t, dst0, [64 * 64 * 2]); - LOCAL_ALIGNED_32(uint8_t, dst1, [64 * 64 * 2]); char str[256]; - VP9DSPContext dsp; - int op, hsize, filter, dx, dy; - - declare_func_emms(AV_CPU_FLAG_MMX | AV_CPU_FLAG_MMXEXT, - void, uint8_t *dst, const uint8_t *ref, - ptrdiff_t dst_stride, ptrdiff_t ref_stride, - int h, int mx, int my); for (op = 0; op < 2; op++) { - ff_vp9dsp_init(&dsp); - for (hsize = 0; hsize < 5; hsize++) { - int size = 64 >> hsize; - - for (filter = 0; filter < 4; filter++) { - for (dx = 0; dx < 2; dx++) { - for (dy = 0; dy < 2; dy++) { - if (dx || dy) { - snprintf(str, sizeof(str), "%s_%s_%d%s", op_names[op], - filter_names[filter], size, - subpel_names[dy][dx]); - } else { - snprintf(str, sizeof(str), "%s%d", op_names[op], size); - } - if (check_func(dsp.mc[hsize][filter][op][dx][dy], - "vp9_%s", str)) { - int mx = dx ? 1 + (rnd() % 14) : 0; - int my = dy ? 1 + (rnd() % 14) : 0; - randomize_buffers(); - call_ref(dst0, src, - size * SIZEOF_PIXEL, - SRC_BUF_STRIDE * SIZEOF_PIXEL, - size, mx, my); - call_new(dst1, src, - size * SIZEOF_PIXEL, - SRC_BUF_STRIDE * SIZEOF_PIXEL, - size, mx, my); - if (memcmp(dst0, dst1, DST_BUF_SIZE)) - fail(); - - // SIMD implementations for each filter of subpel - // functions are identical - if (filter >= 1 && filter <= 2) continue; - - bench_new(dst1, src, size * SIZEOF_PIXEL, - SRC_BUF_STRIDE * SIZEOF_PIXEL, - size, mx, my); + for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) { + ff_vp9dsp_init(&dsp, bit_depth, 0); + for (hsize = 0; hsize < 5; hsize++) { + int size = 64 >> hsize; + + for (filter = 0; filter < 4; filter++) { + for (dx = 0; dx < 2; dx++) { + for (dy = 0; dy < 2; dy++) { + if (dx || dy) { + snprintf(str, sizeof(str), + "%s_%s_%d%s", op_names[op], + filter_names[filter], size, + subpel_names[dy][dx]); + } else { + snprintf(str, sizeof(str), + "%s%d", op_names[op], size); + } + if (check_func(dsp.mc[hsize][filter][op][dx][dy], + "vp9_%s_%dbpp", str, bit_depth)) { + int mx = dx ? 1 + (rnd() % 14) : 0; + int my = dy ? 1 + (rnd() % 14) : 0; + randomize_buffers(); + call_ref(dst0, size * SIZEOF_PIXEL, + src, SRC_BUF_STRIDE * SIZEOF_PIXEL, + size, mx, my); + call_new(dst1, size * SIZEOF_PIXEL, + src, SRC_BUF_STRIDE * SIZEOF_PIXEL, + size, mx, my); + if (memcmp(dst0, dst1, DST_BUF_SIZE)) + fail(); + + // simd implementations for each filter of subpel + // functions are identical + if (filter >= 1 && filter <= 2) continue; + // 10/12 bpp for bilin are identical + if (bit_depth == 12 && filter == 3) continue; + + bench_new(dst1, size * SIZEOF_PIXEL, + src, SRC_BUF_STRIDE * SIZEOF_PIXEL, + size, mx, my); + } } } } @@ -281,6 +624,8 @@ static void check_mc(void) void checkasm_check_vp9dsp(void) { + check_ipred(); + check_itxfm(); check_loopfilter(); check_mc(); } |