You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

341 lines
14KB

  1. /*
  2. * Copyright (c) 2015 Ronald S. Bultje <rsbultje@gmail.com>
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or modify
  7. * it under the terms of the GNU General Public License as published by
  8. * the Free Software Foundation; either version 2 of the License, or
  9. * (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
  14. * GNU General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU General Public License along
  17. * with FFmpeg; if not, write to the Free Software Foundation, Inc.,
  18. * 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
  19. */
  20. #include <string.h>
  21. #include "checkasm.h"
  22. #include "libavcodec/vp9dsp.h"
  23. #include "libavutil/common.h"
  24. #include "libavutil/internal.h"
  25. #include "libavutil/intreadwrite.h"
  26. static const uint32_t pixel_mask[3] = { 0xffffffff, 0x03ff03ff, 0x0fff0fff };
  27. #define SIZEOF_PIXEL ((bit_depth + 7) / 8)
  28. #define randomize_buffers() \
  29. do { \
  30. uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \
  31. int k; \
  32. for (k = -4; k < SIZEOF_PIXEL * FFMAX(8, size); k += 4) { \
  33. uint32_t r = rnd() & mask; \
  34. AV_WN32A(a + k, r); \
  35. } \
  36. for (k = 0; k < size * SIZEOF_PIXEL; k += 4) { \
  37. uint32_t r = rnd() & mask; \
  38. AV_WN32A(l + k, r); \
  39. } \
  40. } while (0)
  41. static void check_ipred(void)
  42. {
  43. LOCAL_ALIGNED_32(uint8_t, a_buf, [64 * 2]);
  44. uint8_t *a = &a_buf[32 * 2];
  45. LOCAL_ALIGNED_32(uint8_t, l, [32 * 2]);
  46. LOCAL_ALIGNED_32(uint8_t, dst0, [32 * 32 * 2]);
  47. LOCAL_ALIGNED_32(uint8_t, dst1, [32 * 32 * 2]);
  48. VP9DSPContext dsp;
  49. int tx, mode, bit_depth;
  50. declare_func(void, uint8_t *dst, ptrdiff_t stride,
  51. const uint8_t *left, const uint8_t *top);
  52. static const char *const mode_names[N_INTRA_PRED_MODES] = {
  53. [VERT_PRED] = "vert",
  54. [HOR_PRED] = "hor",
  55. [DC_PRED] = "dc",
  56. [DIAG_DOWN_LEFT_PRED] = "diag_downleft",
  57. [DIAG_DOWN_RIGHT_PRED] = "diag_downright",
  58. [VERT_RIGHT_PRED] = "vert_right",
  59. [HOR_DOWN_PRED] = "hor_down",
  60. [VERT_LEFT_PRED] = "vert_left",
  61. [HOR_UP_PRED] = "hor_up",
  62. [TM_VP8_PRED] = "tm",
  63. [LEFT_DC_PRED] = "dc_left",
  64. [TOP_DC_PRED] = "dc_top",
  65. [DC_128_PRED] = "dc_128",
  66. [DC_127_PRED] = "dc_127",
  67. [DC_129_PRED] = "dc_129",
  68. };
  69. for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
  70. ff_vp9dsp_init(&dsp, bit_depth, 0);
  71. for (tx = 0; tx < 4; tx++) {
  72. int size = 4 << tx;
  73. for (mode = 0; mode < N_INTRA_PRED_MODES; mode++) {
  74. if (check_func(dsp.intra_pred[tx][mode], "vp9_%s_%dx%d_%dbpp",
  75. mode_names[mode], size, size, bit_depth)) {
  76. randomize_buffers();
  77. call_ref(dst0, size * SIZEOF_PIXEL, l, a);
  78. call_new(dst1, size * SIZEOF_PIXEL, l, a);
  79. if (memcmp(dst0, dst1, size * size * SIZEOF_PIXEL))
  80. fail();
  81. bench_new(dst1, size * SIZEOF_PIXEL,l, a);
  82. }
  83. }
  84. }
  85. }
  86. report("ipred");
  87. }
  88. #undef randomize_buffers
  89. #define setpx(a,b,c) \
  90. do { \
  91. if (SIZEOF_PIXEL == 1) { \
  92. buf0[(a) + (b) * jstride] = c; \
  93. } else { \
  94. ((uint16_t *)buf0)[(a) + (b) * jstride] = c; \
  95. } \
  96. } while (0)
  97. #define setdx(a,b,c,d) setpx(a,b,(c)-(d)+(rnd()%((d)*2+1)))
  98. #define setsx(a,b,c,d) setdx(a,b,c,(d) << (bit_depth - 8))
  99. #define randomize_buffers(bidx, lineoff, str) \
  100. do { \
  101. uint32_t mask = (1 << bit_depth) - 1; \
  102. int off = dir ? lineoff : lineoff * 16; \
  103. int istride = dir ? 1 : 16; \
  104. int jstride = dir ? str : 1; \
  105. int i, j; \
  106. for (i = 0; i < 2; i++) /* flat16 */ { \
  107. int idx = off + i * istride, p0, q0; \
  108. setpx(idx, 0, q0 = rnd() & mask); \
  109. setsx(idx, -1, p0 = q0, E[bidx] >> 2); \
  110. for (j = 1; j < 8; j++) { \
  111. setsx(idx, -1 - j, p0, F[bidx]); \
  112. setsx(idx, j, q0, F[bidx]); \
  113. } \
  114. } \
  115. for (i = 2; i < 4; i++) /* flat8 */ { \
  116. int idx = off + i * istride, p0, q0; \
  117. setpx(idx, 0, q0 = rnd() & mask); \
  118. setsx(idx, -1, p0 = q0, E[bidx] >> 2); \
  119. for (j = 1; j < 4; j++) { \
  120. setsx(idx, -1 - j, p0, F[bidx]); \
  121. setsx(idx, j, q0, F[bidx]); \
  122. } \
  123. for (j = 4; j < 8; j++) { \
  124. setpx(idx, -1 - j, rnd() & mask); \
  125. setpx(idx, j, rnd() & mask); \
  126. } \
  127. } \
  128. for (i = 4; i < 6; i++) /* regular */ { \
  129. int idx = off + i * istride, p2, p1, p0, q0, q1, q2; \
  130. setpx(idx, 0, q0 = rnd() & mask); \
  131. setsx(idx, 1, q1 = q0, I[bidx]); \
  132. setsx(idx, 2, q2 = q1, I[bidx]); \
  133. setsx(idx, 3, q2, I[bidx]); \
  134. setsx(idx, -1, p0 = q0, E[bidx] >> 2); \
  135. setsx(idx, -2, p1 = p0, I[bidx]); \
  136. setsx(idx, -3, p2 = p1, I[bidx]); \
  137. setsx(idx, -4, p2, I[bidx]); \
  138. for (j = 4; j < 8; j++) { \
  139. setpx(idx, -1 - j, rnd() & mask); \
  140. setpx(idx, j, rnd() & mask); \
  141. } \
  142. } \
  143. for (i = 6; i < 8; i++) /* off */ { \
  144. int idx = off + i * istride; \
  145. for (j = 0; j < 8; j++) { \
  146. setpx(idx, -1 - j, rnd() & mask); \
  147. setpx(idx, j, rnd() & mask); \
  148. } \
  149. } \
  150. } while (0)
  151. static void check_loopfilter(void)
  152. {
  153. LOCAL_ALIGNED_32(uint8_t, base0, [32 + 16 * 16 * 2]);
  154. LOCAL_ALIGNED_32(uint8_t, base1, [32 + 16 * 16 * 2]);
  155. VP9DSPContext dsp;
  156. int dir, wd, wd2, bit_depth;
  157. static const char *const dir_name[2] = { "h", "v" };
  158. static const int E[2] = { 20, 28 }, I[2] = { 10, 16 };
  159. static const int H[2] = { 7, 11 }, F[2] = { 1, 1 };
  160. declare_func(void, uint8_t *dst, ptrdiff_t stride, int E, int I, int H);
  161. for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
  162. ff_vp9dsp_init(&dsp, bit_depth, 0);
  163. for (dir = 0; dir < 2; dir++) {
  164. int midoff = (dir ? 8 * 8 : 8) * SIZEOF_PIXEL;
  165. int midoff_aligned = (dir ? 8 * 8 : 16) * SIZEOF_PIXEL;
  166. uint8_t *buf0 = base0 + midoff_aligned;
  167. uint8_t *buf1 = base1 + midoff_aligned;
  168. for (wd = 0; wd < 3; wd++) {
  169. // 4/8/16wd_8px
  170. if (check_func(dsp.loop_filter_8[wd][dir],
  171. "vp9_loop_filter_%s_%d_8_%dbpp",
  172. dir_name[dir], 4 << wd, bit_depth)) {
  173. randomize_buffers(0, 0, 8);
  174. memcpy(buf1 - midoff, buf0 - midoff,
  175. 16 * 8 * SIZEOF_PIXEL);
  176. call_ref(buf0, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]);
  177. call_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]);
  178. if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 8 * SIZEOF_PIXEL))
  179. fail();
  180. bench_new(buf1, 16 * SIZEOF_PIXEL >> dir, E[0], I[0], H[0]);
  181. }
  182. }
  183. midoff = (dir ? 16 * 8 : 8) * SIZEOF_PIXEL;
  184. midoff_aligned = (dir ? 16 * 8 : 16) * SIZEOF_PIXEL;
  185. buf0 = base0 + midoff_aligned;
  186. buf1 = base1 + midoff_aligned;
  187. // 16wd_16px loopfilter
  188. if (check_func(dsp.loop_filter_16[dir],
  189. "vp9_loop_filter_%s_16_16_%dbpp",
  190. dir_name[dir], bit_depth)) {
  191. randomize_buffers(0, 0, 16);
  192. randomize_buffers(0, 8, 16);
  193. memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL);
  194. call_ref(buf0, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]);
  195. call_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]);
  196. if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL))
  197. fail();
  198. bench_new(buf1, 16 * SIZEOF_PIXEL, E[0], I[0], H[0]);
  199. }
  200. for (wd = 0; wd < 2; wd++) {
  201. for (wd2 = 0; wd2 < 2; wd2++) {
  202. // mix2 loopfilter
  203. if (check_func(dsp.loop_filter_mix2[wd][wd2][dir],
  204. "vp9_loop_filter_mix2_%s_%d%d_16_%dbpp",
  205. dir_name[dir], 4 << wd, 4 << wd2, bit_depth)) {
  206. randomize_buffers(0, 0, 16);
  207. randomize_buffers(1, 8, 16);
  208. memcpy(buf1 - midoff, buf0 - midoff, 16 * 16 * SIZEOF_PIXEL);
  209. #define M(a) (((a)[1] << 8) | (a)[0])
  210. call_ref(buf0, 16 * SIZEOF_PIXEL, M(E), M(I), M(H));
  211. call_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H));
  212. if (memcmp(buf0 - midoff, buf1 - midoff, 16 * 16 * SIZEOF_PIXEL))
  213. fail();
  214. bench_new(buf1, 16 * SIZEOF_PIXEL, M(E), M(I), M(H));
  215. #undef M
  216. }
  217. }
  218. }
  219. }
  220. }
  221. report("loopfilter");
  222. }
  223. #undef setsx
  224. #undef setpx
  225. #undef setdx
  226. #undef randomize_buffers
  227. #define DST_BUF_SIZE (size * size * SIZEOF_PIXEL)
  228. #define SRC_BUF_STRIDE 72
  229. #define SRC_BUF_SIZE ((size + 7) * SRC_BUF_STRIDE * SIZEOF_PIXEL)
  230. #define src (buf + 3 * SIZEOF_PIXEL * (SRC_BUF_STRIDE + 1))
  231. #define randomize_buffers() \
  232. do { \
  233. uint32_t mask = pixel_mask[(bit_depth - 8) >> 1]; \
  234. int k; \
  235. for (k = 0; k < SRC_BUF_SIZE; k += 4) { \
  236. uint32_t r = rnd() & mask; \
  237. AV_WN32A(buf + k, r); \
  238. } \
  239. if (op == 1) { \
  240. for (k = 0; k < DST_BUF_SIZE; k += 4) { \
  241. uint32_t r = rnd() & mask; \
  242. AV_WN32A(dst0 + k, r); \
  243. AV_WN32A(dst1 + k, r); \
  244. } \
  245. } \
  246. } while (0)
  247. static void check_mc(void)
  248. {
  249. LOCAL_ALIGNED_32(uint8_t, buf, [72 * 72 * 2]);
  250. LOCAL_ALIGNED_32(uint8_t, dst0, [64 * 64 * 2]);
  251. LOCAL_ALIGNED_32(uint8_t, dst1, [64 * 64 * 2]);
  252. VP9DSPContext dsp;
  253. int op, hsize, bit_depth, filter, dx, dy;
  254. declare_func(void, uint8_t *dst, ptrdiff_t dst_stride,
  255. const uint8_t *ref, ptrdiff_t ref_stride,
  256. int h, int mx, int my);
  257. static const char *const filter_names[4] = {
  258. "8tap_smooth", "8tap_regular", "8tap_sharp", "bilin"
  259. };
  260. static const char *const subpel_names[2][2] = { { "", "h" }, { "v", "hv" } };
  261. static const char *const op_names[2] = { "put", "avg" };
  262. char str[256];
  263. for (op = 0; op < 2; op++) {
  264. for (bit_depth = 8; bit_depth <= 12; bit_depth += 2) {
  265. ff_vp9dsp_init(&dsp, bit_depth, 0);
  266. for (hsize = 0; hsize < 5; hsize++) {
  267. int size = 64 >> hsize;
  268. for (filter = 0; filter < 4; filter++) {
  269. for (dx = 0; dx < 2; dx++) {
  270. for (dy = 0; dy < 2; dy++) {
  271. if (dx || dy) {
  272. snprintf(str, sizeof(str),
  273. "%s_%s_%d%s", op_names[op],
  274. filter_names[filter], size,
  275. subpel_names[dy][dx]);
  276. } else {
  277. snprintf(str, sizeof(str),
  278. "%s%d", op_names[op], size);
  279. }
  280. if (check_func(dsp.mc[hsize][filter][op][dx][dy],
  281. "vp9_%s_%dbpp", str, bit_depth)) {
  282. int mx = dx ? 1 + (rnd() % 14) : 0;
  283. int my = dy ? 1 + (rnd() % 14) : 0;
  284. randomize_buffers();
  285. call_ref(dst0, size * SIZEOF_PIXEL,
  286. src, SRC_BUF_STRIDE * SIZEOF_PIXEL,
  287. size, mx, my);
  288. call_new(dst1, size * SIZEOF_PIXEL,
  289. src, SRC_BUF_STRIDE * SIZEOF_PIXEL,
  290. size, mx, my);
  291. if (memcmp(dst0, dst1, DST_BUF_SIZE))
  292. fail();
  293. // simd implementations for each filter of subpel
  294. // functions are identical
  295. if (filter >= 1 && filter <= 2) continue;
  296. // 10/12 bpp for bilin are identical
  297. if (bit_depth == 12 && filter == 3) continue;
  298. bench_new(dst1, size * SIZEOF_PIXEL,
  299. src, SRC_BUF_STRIDE * SIZEOF_PIXEL,
  300. size, mx, my);
  301. }
  302. }
  303. }
  304. }
  305. }
  306. }
  307. }
  308. report("mc");
  309. }
  310. void checkasm_check_vp9dsp(void)
  311. {
  312. check_ipred();
  313. check_loopfilter();
  314. check_mc();
  315. }