You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

2242 lines
76KB

  1. /*
  2. * VP9 compatible video decoder
  3. *
  4. * Copyright (C) 2013 Ronald S. Bultje <rsbultje gmail com>
  5. * Copyright (C) 2013 Clément Bœsch <u pkh me>
  6. *
  7. * This file is part of FFmpeg.
  8. *
  9. * FFmpeg is free software; you can redistribute it and/or
  10. * modify it under the terms of the GNU Lesser General Public
  11. * License as published by the Free Software Foundation; either
  12. * version 2.1 of the License, or (at your option) any later version.
  13. *
  14. * FFmpeg is distributed in the hope that it will be useful,
  15. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  16. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  17. * Lesser General Public License for more details.
  18. *
  19. * You should have received a copy of the GNU Lesser General Public
  20. * License along with FFmpeg; if not, write to the Free Software
  21. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  22. */
  23. #include "libavutil/common.h"
  24. #include "libavutil/intreadwrite.h"
  25. #include "vp9dsp.h"
  26. #include "rnd_avg.h"
  27. // FIXME see whether we can merge parts of this (perhaps at least 4x4 and 8x8)
  28. // back with h264pred.[ch]
  29. static void vert_4x4_c(uint8_t *dst, ptrdiff_t stride,
  30. const uint8_t *left, const uint8_t *top)
  31. {
  32. unsigned p4 = AV_RN32A(top);
  33. AV_WN32A(dst + stride * 0, p4);
  34. AV_WN32A(dst + stride * 1, p4);
  35. AV_WN32A(dst + stride * 2, p4);
  36. AV_WN32A(dst + stride * 3, p4);
  37. }
  38. static void vert_8x8_c(uint8_t *dst, ptrdiff_t stride,
  39. const uint8_t *left, const uint8_t *top)
  40. {
  41. uint64_t p8 = AV_RN64A(top);
  42. int y;
  43. for (y = 0; y < 8; y++) {
  44. AV_WN64A(dst, p8);
  45. dst += stride;
  46. }
  47. }
  48. static void vert_16x16_c(uint8_t *dst, ptrdiff_t stride,
  49. const uint8_t *left, const uint8_t *top)
  50. {
  51. uint64_t p8a = AV_RN64A(top + 0), p8b = AV_RN64A(top + 8);
  52. int y;
  53. for (y = 0; y < 16; y++) {
  54. AV_WN64A(dst + 0, p8a);
  55. AV_WN64A(dst + 8, p8b);
  56. dst += stride;
  57. }
  58. }
  59. static void vert_32x32_c(uint8_t *dst, ptrdiff_t stride,
  60. const uint8_t *left, const uint8_t *top)
  61. {
  62. uint64_t p8a = AV_RN64A(top + 0), p8b = AV_RN64A(top + 8),
  63. p8c = AV_RN64A(top + 16), p8d = AV_RN64A(top + 24);
  64. int y;
  65. for (y = 0; y < 32; y++) {
  66. AV_WN64A(dst + 0, p8a);
  67. AV_WN64A(dst + 8, p8b);
  68. AV_WN64A(dst + 16, p8c);
  69. AV_WN64A(dst + 24, p8d);
  70. dst += stride;
  71. }
  72. }
  73. static void hor_4x4_c(uint8_t *dst, ptrdiff_t stride,
  74. const uint8_t *left, const uint8_t *top)
  75. {
  76. AV_WN32A(dst + stride * 0, left[3] * 0x01010101U);
  77. AV_WN32A(dst + stride * 1, left[2] * 0x01010101U);
  78. AV_WN32A(dst + stride * 2, left[1] * 0x01010101U);
  79. AV_WN32A(dst + stride * 3, left[0] * 0x01010101U);
  80. }
  81. static void hor_8x8_c(uint8_t *dst, ptrdiff_t stride,
  82. const uint8_t *left, const uint8_t *top)
  83. {
  84. int y;
  85. for (y = 0; y < 8; y++) {
  86. AV_WN64A(dst, left[7 - y] * 0x0101010101010101ULL);
  87. dst += stride;
  88. }
  89. }
  90. static void hor_16x16_c(uint8_t *dst, ptrdiff_t stride,
  91. const uint8_t *left, const uint8_t *top)
  92. {
  93. int y;
  94. for (y = 0; y < 16; y++) {
  95. uint64_t p8 = left[15 - y] * 0x0101010101010101ULL;
  96. AV_WN64A(dst + 0, p8);
  97. AV_WN64A(dst + 8, p8);
  98. dst += stride;
  99. }
  100. }
  101. static void hor_32x32_c(uint8_t *dst, ptrdiff_t stride,
  102. const uint8_t *left, const uint8_t *top)
  103. {
  104. int y;
  105. for (y = 0; y < 32; y++) {
  106. uint64_t p8 = left[31 - y] * 0x0101010101010101ULL;
  107. AV_WN64A(dst + 0, p8);
  108. AV_WN64A(dst + 8, p8);
  109. AV_WN64A(dst + 16, p8);
  110. AV_WN64A(dst + 24, p8);
  111. dst += stride;
  112. }
  113. }
  114. static void tm_4x4_c(uint8_t *dst, ptrdiff_t stride,
  115. const uint8_t *left, const uint8_t *top)
  116. {
  117. int y, tl = top[-1];
  118. for (y = 0; y < 4; y++) {
  119. int l_m_tl = left[3 - y] - tl;
  120. dst[0] = av_clip_uint8(top[0] + l_m_tl);
  121. dst[1] = av_clip_uint8(top[1] + l_m_tl);
  122. dst[2] = av_clip_uint8(top[2] + l_m_tl);
  123. dst[3] = av_clip_uint8(top[3] + l_m_tl);
  124. dst += stride;
  125. }
  126. }
  127. static void tm_8x8_c(uint8_t *dst, ptrdiff_t stride,
  128. const uint8_t *left, const uint8_t *top)
  129. {
  130. int y, tl = top[-1];
  131. for (y = 0; y < 8; y++) {
  132. int l_m_tl = left[7 - y] - tl;
  133. dst[0] = av_clip_uint8(top[0] + l_m_tl);
  134. dst[1] = av_clip_uint8(top[1] + l_m_tl);
  135. dst[2] = av_clip_uint8(top[2] + l_m_tl);
  136. dst[3] = av_clip_uint8(top[3] + l_m_tl);
  137. dst[4] = av_clip_uint8(top[4] + l_m_tl);
  138. dst[5] = av_clip_uint8(top[5] + l_m_tl);
  139. dst[6] = av_clip_uint8(top[6] + l_m_tl);
  140. dst[7] = av_clip_uint8(top[7] + l_m_tl);
  141. dst += stride;
  142. }
  143. }
  144. static void tm_16x16_c(uint8_t *dst, ptrdiff_t stride,
  145. const uint8_t *left, const uint8_t *top)
  146. {
  147. int y, tl = top[-1];
  148. for (y = 0; y < 16; y++) {
  149. int l_m_tl = left[15 - y] - tl;
  150. dst[ 0] = av_clip_uint8(top[ 0] + l_m_tl);
  151. dst[ 1] = av_clip_uint8(top[ 1] + l_m_tl);
  152. dst[ 2] = av_clip_uint8(top[ 2] + l_m_tl);
  153. dst[ 3] = av_clip_uint8(top[ 3] + l_m_tl);
  154. dst[ 4] = av_clip_uint8(top[ 4] + l_m_tl);
  155. dst[ 5] = av_clip_uint8(top[ 5] + l_m_tl);
  156. dst[ 6] = av_clip_uint8(top[ 6] + l_m_tl);
  157. dst[ 7] = av_clip_uint8(top[ 7] + l_m_tl);
  158. dst[ 8] = av_clip_uint8(top[ 8] + l_m_tl);
  159. dst[ 9] = av_clip_uint8(top[ 9] + l_m_tl);
  160. dst[10] = av_clip_uint8(top[10] + l_m_tl);
  161. dst[11] = av_clip_uint8(top[11] + l_m_tl);
  162. dst[12] = av_clip_uint8(top[12] + l_m_tl);
  163. dst[13] = av_clip_uint8(top[13] + l_m_tl);
  164. dst[14] = av_clip_uint8(top[14] + l_m_tl);
  165. dst[15] = av_clip_uint8(top[15] + l_m_tl);
  166. dst += stride;
  167. }
  168. }
  169. static void tm_32x32_c(uint8_t *dst, ptrdiff_t stride,
  170. const uint8_t *left, const uint8_t *top)
  171. {
  172. int y, tl = top[-1];
  173. for (y = 0; y < 32; y++) {
  174. int l_m_tl = left[31 - y] - tl;
  175. dst[ 0] = av_clip_uint8(top[ 0] + l_m_tl);
  176. dst[ 1] = av_clip_uint8(top[ 1] + l_m_tl);
  177. dst[ 2] = av_clip_uint8(top[ 2] + l_m_tl);
  178. dst[ 3] = av_clip_uint8(top[ 3] + l_m_tl);
  179. dst[ 4] = av_clip_uint8(top[ 4] + l_m_tl);
  180. dst[ 5] = av_clip_uint8(top[ 5] + l_m_tl);
  181. dst[ 6] = av_clip_uint8(top[ 6] + l_m_tl);
  182. dst[ 7] = av_clip_uint8(top[ 7] + l_m_tl);
  183. dst[ 8] = av_clip_uint8(top[ 8] + l_m_tl);
  184. dst[ 9] = av_clip_uint8(top[ 9] + l_m_tl);
  185. dst[10] = av_clip_uint8(top[10] + l_m_tl);
  186. dst[11] = av_clip_uint8(top[11] + l_m_tl);
  187. dst[12] = av_clip_uint8(top[12] + l_m_tl);
  188. dst[13] = av_clip_uint8(top[13] + l_m_tl);
  189. dst[14] = av_clip_uint8(top[14] + l_m_tl);
  190. dst[15] = av_clip_uint8(top[15] + l_m_tl);
  191. dst[16] = av_clip_uint8(top[16] + l_m_tl);
  192. dst[17] = av_clip_uint8(top[17] + l_m_tl);
  193. dst[18] = av_clip_uint8(top[18] + l_m_tl);
  194. dst[19] = av_clip_uint8(top[19] + l_m_tl);
  195. dst[20] = av_clip_uint8(top[20] + l_m_tl);
  196. dst[21] = av_clip_uint8(top[21] + l_m_tl);
  197. dst[22] = av_clip_uint8(top[22] + l_m_tl);
  198. dst[23] = av_clip_uint8(top[23] + l_m_tl);
  199. dst[24] = av_clip_uint8(top[24] + l_m_tl);
  200. dst[25] = av_clip_uint8(top[25] + l_m_tl);
  201. dst[26] = av_clip_uint8(top[26] + l_m_tl);
  202. dst[27] = av_clip_uint8(top[27] + l_m_tl);
  203. dst[28] = av_clip_uint8(top[28] + l_m_tl);
  204. dst[29] = av_clip_uint8(top[29] + l_m_tl);
  205. dst[30] = av_clip_uint8(top[30] + l_m_tl);
  206. dst[31] = av_clip_uint8(top[31] + l_m_tl);
  207. dst += stride;
  208. }
  209. }
  210. static void dc_4x4_c(uint8_t *dst, ptrdiff_t stride,
  211. const uint8_t *left, const uint8_t *top)
  212. {
  213. unsigned dc = 0x01010101U * ((left[0] + left[1] + left[2] + left[3] +
  214. top[0] + top[1] + top[2] + top[3] + 4) >> 3);
  215. AV_WN32A(dst + stride * 0, dc);
  216. AV_WN32A(dst + stride * 1, dc);
  217. AV_WN32A(dst + stride * 2, dc);
  218. AV_WN32A(dst + stride * 3, dc);
  219. }
  220. static void dc_8x8_c(uint8_t *dst, ptrdiff_t stride,
  221. const uint8_t *left, const uint8_t *top)
  222. {
  223. uint64_t dc = 0x0101010101010101ULL *
  224. ((left[0] + left[1] + left[2] + left[3] + left[4] + left[5] +
  225. left[6] + left[7] + top[0] + top[1] + top[2] + top[3] +
  226. top[4] + top[5] + top[6] + top[7] + 8) >> 4);
  227. int y;
  228. for (y = 0; y < 8; y++) {
  229. AV_WN64A(dst, dc);
  230. dst += stride;
  231. }
  232. }
  233. static void dc_16x16_c(uint8_t *dst, ptrdiff_t stride,
  234. const uint8_t *left, const uint8_t *top)
  235. {
  236. uint64_t dc = 0x0101010101010101ULL *
  237. ((left[0] + left[1] + left[2] + left[3] + left[4] + left[5] + left[6] +
  238. left[7] + left[8] + left[9] + left[10] + left[11] + left[12] +
  239. left[13] + left[14] + left[15] + top[0] + top[1] + top[2] + top[3] +
  240. top[4] + top[5] + top[6] + top[7] + top[8] + top[9] + top[10] +
  241. top[11] + top[12] + top[13] + top[14] + top[15] + 16) >> 5);
  242. int y;
  243. for (y = 0; y < 16; y++) {
  244. AV_WN64A(dst + 0, dc);
  245. AV_WN64A(dst + 8, dc);
  246. dst += stride;
  247. }
  248. }
  249. static void dc_32x32_c(uint8_t *dst, ptrdiff_t stride,
  250. const uint8_t *left, const uint8_t *top)
  251. {
  252. uint64_t dc = 0x0101010101010101ULL *
  253. ((left[0] + left[1] + left[2] + left[3] + left[4] + left[5] + left[6] +
  254. left[7] + left[8] + left[9] + left[10] + left[11] + left[12] +
  255. left[13] + left[14] + left[15] + left[16] + left[17] + left[18] +
  256. left[19] + left[20] + left[21] + left[22] + left[23] + left[24] +
  257. left[25] + left[26] + left[27] + left[28] + left[29] + left[30] +
  258. left[31] + top[0] + top[1] + top[2] + top[3] + top[4] + top[5] +
  259. top[6] + top[7] + top[8] + top[9] + top[10] + top[11] + top[12] +
  260. top[13] + top[14] + top[15] + top[16] + top[17] + top[18] + top[19] +
  261. top[20] + top[21] + top[22] + top[23] + top[24] + top[25] + top[26] +
  262. top[27] + top[28] + top[29] + top[30] + top[31] + 32) >> 6);
  263. int y;
  264. for (y = 0; y < 32; y++) {
  265. AV_WN64A(dst + 0, dc);
  266. AV_WN64A(dst + 8, dc);
  267. AV_WN64A(dst + 16, dc);
  268. AV_WN64A(dst + 24, dc);
  269. dst += stride;
  270. }
  271. }
  272. static void dc_left_4x4_c(uint8_t *dst, ptrdiff_t stride,
  273. const uint8_t *left, const uint8_t *top)
  274. {
  275. unsigned dc = 0x01010101U * ((left[0] + left[1] + left[2] + left[3] + 2) >> 2);
  276. AV_WN32A(dst + stride * 0, dc);
  277. AV_WN32A(dst + stride * 1, dc);
  278. AV_WN32A(dst + stride * 2, dc);
  279. AV_WN32A(dst + stride * 3, dc);
  280. }
  281. static void dc_left_8x8_c(uint8_t *dst, ptrdiff_t stride,
  282. const uint8_t *left, const uint8_t *top)
  283. {
  284. uint64_t dc = 0x0101010101010101ULL *
  285. ((left[0] + left[1] + left[2] + left[3] +
  286. left[4] + left[5] + left[6] + left[7] + 4) >> 3);
  287. int y;
  288. for (y = 0; y < 8; y++) {
  289. AV_WN64A(dst, dc);
  290. dst += stride;
  291. }
  292. }
  293. static void dc_left_16x16_c(uint8_t *dst, ptrdiff_t stride,
  294. const uint8_t *left, const uint8_t *top)
  295. {
  296. uint64_t dc = 0x0101010101010101ULL *
  297. ((left[0] + left[1] + left[2] + left[3] + left[4] + left[5] +
  298. left[6] + left[7] + left[8] + left[9] + left[10] + left[11] +
  299. left[12] + left[13] + left[14] + left[15] + 8) >> 4);
  300. int y;
  301. for (y = 0; y < 16; y++) {
  302. AV_WN64A(dst + 0, dc);
  303. AV_WN64A(dst + 8, dc);
  304. dst += stride;
  305. }
  306. }
  307. static void dc_left_32x32_c(uint8_t *dst, ptrdiff_t stride,
  308. const uint8_t *left, const uint8_t *top)
  309. {
  310. uint64_t dc = 0x0101010101010101ULL *
  311. ((left[0] + left[1] + left[2] + left[3] + left[4] + left[5] +
  312. left[6] + left[7] + left[8] + left[9] + left[10] + left[11] +
  313. left[12] + left[13] + left[14] + left[15] + left[16] + left[17] +
  314. left[18] + left[19] + left[20] + left[21] + left[22] + left[23] +
  315. left[24] + left[25] + left[26] + left[27] + left[28] + left[29] +
  316. left[30] + left[31] + 16) >> 5);
  317. int y;
  318. for (y = 0; y < 32; y++) {
  319. AV_WN64A(dst + 0, dc);
  320. AV_WN64A(dst + 8, dc);
  321. AV_WN64A(dst + 16, dc);
  322. AV_WN64A(dst + 24, dc);
  323. dst += stride;
  324. }
  325. }
  326. static void dc_top_4x4_c(uint8_t *dst, ptrdiff_t stride,
  327. const uint8_t *left, const uint8_t *top)
  328. {
  329. unsigned dc = 0x01010101U * ((top[0] + top[1] + top[2] + top[3] + 2) >> 2);
  330. AV_WN32A(dst + stride * 0, dc);
  331. AV_WN32A(dst + stride * 1, dc);
  332. AV_WN32A(dst + stride * 2, dc);
  333. AV_WN32A(dst + stride * 3, dc);
  334. }
  335. static void dc_top_8x8_c(uint8_t *dst, ptrdiff_t stride,
  336. const uint8_t *left, const uint8_t *top)
  337. {
  338. uint64_t dc = 0x0101010101010101ULL *
  339. ((top[0] + top[1] + top[2] + top[3] +
  340. top[4] + top[5] + top[6] + top[7] + 4) >> 3);
  341. int y;
  342. for (y = 0; y < 8; y++) {
  343. AV_WN64A(dst, dc);
  344. dst += stride;
  345. }
  346. }
  347. static void dc_top_16x16_c(uint8_t *dst, ptrdiff_t stride,
  348. const uint8_t *left, const uint8_t *top)
  349. {
  350. uint64_t dc = 0x0101010101010101ULL *
  351. ((top[0] + top[1] + top[2] + top[3] + top[4] + top[5] +
  352. top[6] + top[7] + top[8] + top[9] + top[10] + top[11] +
  353. top[12] + top[13] + top[14] + top[15] + 8) >> 4);
  354. int y;
  355. for (y = 0; y < 16; y++) {
  356. AV_WN64A(dst + 0, dc);
  357. AV_WN64A(dst + 8, dc);
  358. dst += stride;
  359. }
  360. }
  361. static void dc_top_32x32_c(uint8_t *dst, ptrdiff_t stride,
  362. const uint8_t *left, const uint8_t *top)
  363. {
  364. uint64_t dc = 0x0101010101010101ULL *
  365. ((top[0] + top[1] + top[2] + top[3] + top[4] + top[5] +
  366. top[6] + top[7] + top[8] + top[9] + top[10] + top[11] +
  367. top[12] + top[13] + top[14] + top[15] + top[16] + top[17] +
  368. top[18] + top[19] + top[20] + top[21] + top[22] + top[23] +
  369. top[24] + top[25] + top[26] + top[27] + top[28] + top[29] +
  370. top[30] + top[31] + 16) >> 5);
  371. int y;
  372. for (y = 0; y < 32; y++) {
  373. AV_WN64A(dst + 0, dc);
  374. AV_WN64A(dst + 8, dc);
  375. AV_WN64A(dst + 16, dc);
  376. AV_WN64A(dst + 24, dc);
  377. dst += stride;
  378. }
  379. }
  380. static void dc_128_4x4_c(uint8_t *dst, ptrdiff_t stride,
  381. const uint8_t *left, const uint8_t *top)
  382. {
  383. AV_WN32A(dst + stride * 0, 0x80808080U);
  384. AV_WN32A(dst + stride * 1, 0x80808080U);
  385. AV_WN32A(dst + stride * 2, 0x80808080U);
  386. AV_WN32A(dst + stride * 3, 0x80808080U);
  387. }
  388. static void dc_128_8x8_c(uint8_t *dst, ptrdiff_t stride,
  389. const uint8_t *left, const uint8_t *top)
  390. {
  391. int y;
  392. for (y = 0; y < 8; y++) {
  393. AV_WN64A(dst, 0x8080808080808080ULL);
  394. dst += stride;
  395. }
  396. }
  397. static void dc_128_16x16_c(uint8_t *dst, ptrdiff_t stride,
  398. const uint8_t *left, const uint8_t *top)
  399. {
  400. int y;
  401. for (y = 0; y < 16; y++) {
  402. AV_WN64A(dst + 0, 0x8080808080808080ULL);
  403. AV_WN64A(dst + 8, 0x8080808080808080ULL);
  404. dst += stride;
  405. }
  406. }
  407. static void dc_128_32x32_c(uint8_t *dst, ptrdiff_t stride,
  408. const uint8_t *left, const uint8_t *top)
  409. {
  410. int y;
  411. for (y = 0; y < 32; y++) {
  412. AV_WN64A(dst + 0, 0x8080808080808080ULL);
  413. AV_WN64A(dst + 8, 0x8080808080808080ULL);
  414. AV_WN64A(dst + 16, 0x8080808080808080ULL);
  415. AV_WN64A(dst + 24, 0x8080808080808080ULL);
  416. dst += stride;
  417. }
  418. }
  419. static void dc_127_4x4_c(uint8_t *dst, ptrdiff_t stride,
  420. const uint8_t *left, const uint8_t *top)
  421. {
  422. AV_WN32A(dst + stride * 0, 0x7F7F7F7FU);
  423. AV_WN32A(dst + stride * 1, 0x7F7F7F7FU);
  424. AV_WN32A(dst + stride * 2, 0x7F7F7F7FU);
  425. AV_WN32A(dst + stride * 3, 0x7F7F7F7FU);
  426. }
  427. static void dc_127_8x8_c(uint8_t *dst, ptrdiff_t stride,
  428. const uint8_t *left, const uint8_t *top)
  429. {
  430. int y;
  431. for (y = 0; y < 8; y++) {
  432. AV_WN64A(dst, 0x7F7F7F7F7F7F7F7FULL);
  433. dst += stride;
  434. }
  435. }
  436. static void dc_127_16x16_c(uint8_t *dst, ptrdiff_t stride,
  437. const uint8_t *left, const uint8_t *top)
  438. {
  439. int y;
  440. for (y = 0; y < 16; y++) {
  441. AV_WN64A(dst + 0, 0x7F7F7F7F7F7F7F7FULL);
  442. AV_WN64A(dst + 8, 0x7F7F7F7F7F7F7F7FULL);
  443. dst += stride;
  444. }
  445. }
  446. static void dc_127_32x32_c(uint8_t *dst, ptrdiff_t stride,
  447. const uint8_t *left, const uint8_t *top)
  448. {
  449. int y;
  450. for (y = 0; y < 32; y++) {
  451. AV_WN64A(dst + 0, 0x7F7F7F7F7F7F7F7FULL);
  452. AV_WN64A(dst + 8, 0x7F7F7F7F7F7F7F7FULL);
  453. AV_WN64A(dst + 16, 0x7F7F7F7F7F7F7F7FULL);
  454. AV_WN64A(dst + 24, 0x7F7F7F7F7F7F7F7FULL);
  455. dst += stride;
  456. }
  457. }
  458. static void dc_129_4x4_c(uint8_t *dst, ptrdiff_t stride,
  459. const uint8_t *left, const uint8_t *top)
  460. {
  461. AV_WN32A(dst + stride * 0, 0x81818181U);
  462. AV_WN32A(dst + stride * 1, 0x81818181U);
  463. AV_WN32A(dst + stride * 2, 0x81818181U);
  464. AV_WN32A(dst + stride * 3, 0x81818181U);
  465. }
  466. static void dc_129_8x8_c(uint8_t *dst, ptrdiff_t stride,
  467. const uint8_t *left, const uint8_t *top)
  468. {
  469. int y;
  470. for (y = 0; y < 8; y++) {
  471. AV_WN64A(dst, 0x8181818181818181ULL);
  472. dst += stride;
  473. }
  474. }
  475. static void dc_129_16x16_c(uint8_t *dst, ptrdiff_t stride,
  476. const uint8_t *left, const uint8_t *top)
  477. {
  478. int y;
  479. for (y = 0; y < 16; y++) {
  480. AV_WN64A(dst + 0, 0x8181818181818181ULL);
  481. AV_WN64A(dst + 8, 0x8181818181818181ULL);
  482. dst += stride;
  483. }
  484. }
  485. static void dc_129_32x32_c(uint8_t *dst, ptrdiff_t stride,
  486. const uint8_t *left, const uint8_t *top)
  487. {
  488. int y;
  489. for (y = 0; y < 32; y++) {
  490. AV_WN64A(dst + 0, 0x8181818181818181ULL);
  491. AV_WN64A(dst + 8, 0x8181818181818181ULL);
  492. AV_WN64A(dst + 16, 0x8181818181818181ULL);
  493. AV_WN64A(dst + 24, 0x8181818181818181ULL);
  494. dst += stride;
  495. }
  496. }
  497. #define DST(x, y) dst[(x) + (y) * stride]
  498. static void diag_downleft_4x4_c(uint8_t *dst, ptrdiff_t stride,
  499. const uint8_t *left, const uint8_t *top)
  500. {
  501. int a0 = top[0], a1 = top[1], a2 = top[2], a3 = top[3],
  502. a4 = top[4], a5 = top[5], a6 = top[6], a7 = top[7];
  503. DST(0,0) = (a0 + a1 * 2 + a2 + 2) >> 2;
  504. DST(1,0) = DST(0,1) = (a1 + a2 * 2 + a3 + 2) >> 2;
  505. DST(2,0) = DST(1,1) = DST(0,2) = (a2 + a3 * 2 + a4 + 2) >> 2;
  506. DST(3,0) = DST(2,1) = DST(1,2) = DST(0,3) = (a3 + a4 * 2 + a5 + 2) >> 2;
  507. DST(3,1) = DST(2,2) = DST(1,3) = (a4 + a5 * 2 + a6 + 2) >> 2;
  508. DST(3,2) = DST(2,3) = (a5 + a6 * 2 + a7 + 2) >> 2;
  509. DST(3,3) = a7; // note: this is different from vp8 and such
  510. }
  511. #define def_diag_downleft(size) \
  512. static void diag_downleft_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  513. const uint8_t *left, const uint8_t *top) \
  514. { \
  515. int i, j; \
  516. uint8_t v[size - 1]; \
  517. \
  518. for (i = 0; i < size - 2; i++) \
  519. v[i] = (top[i] + top[i + 1] * 2 + top[i + 2] + 2) >> 2; \
  520. v[size - 2] = (top[size - 2] + top[size - 1] * 3 + 2) >> 2; \
  521. \
  522. for (j = 0; j < size; j++) { \
  523. memcpy(dst + j*stride, v + j, size - 1 - j); \
  524. memset(dst + j*stride + size - 1 - j, top[size - 1], j + 1); \
  525. } \
  526. }
  527. def_diag_downleft(8)
  528. def_diag_downleft(16)
  529. def_diag_downleft(32)
  530. static void diag_downright_4x4_c(uint8_t *dst, ptrdiff_t stride,
  531. const uint8_t *left, const uint8_t *top)
  532. {
  533. int tl = top[-1], a0 = top[0], a1 = top[1], a2 = top[2], a3 = top[3],
  534. l0 = left[3], l1 = left[2], l2 = left[1], l3 = left[0];
  535. DST(0,3) = (l1 + l2 * 2 + l3 + 2) >> 2;
  536. DST(0,2) = DST(1,3) = (l0 + l1 * 2 + l2 + 2) >> 2;
  537. DST(0,1) = DST(1,2) = DST(2,3) = (tl + l0 * 2 + l1 + 2) >> 2;
  538. DST(0,0) = DST(1,1) = DST(2,2) = DST(3,3) = (l0 + tl * 2 + a0 + 2) >> 2;
  539. DST(1,0) = DST(2,1) = DST(3,2) = (tl + a0 * 2 + a1 + 2) >> 2;
  540. DST(2,0) = DST(3,1) = (a0 + a1 * 2 + a2 + 2) >> 2;
  541. DST(3,0) = (a1 + a2 * 2 + a3 + 2) >> 2;
  542. }
  543. #define def_diag_downright(size) \
  544. static void diag_downright_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  545. const uint8_t *left, const uint8_t *top) \
  546. { \
  547. int i, j; \
  548. uint8_t v[size + size - 1]; \
  549. \
  550. for (i = 0; i < size - 2; i++) { \
  551. v[i ] = (left[i] + left[i + 1] * 2 + left[i + 2] + 2) >> 2; \
  552. v[size + 1 + i] = (top[i] + top[i + 1] * 2 + top[i + 2] + 2) >> 2; \
  553. } \
  554. v[size - 2] = (left[size - 2] + left[size - 1] * 2 + top[-1] + 2) >> 2; \
  555. v[size - 1] = (left[size - 1] + top[-1] * 2 + top[ 0] + 2) >> 2; \
  556. v[size ] = (top[-1] + top[0] * 2 + top[ 1] + 2) >> 2; \
  557. \
  558. for (j = 0; j < size; j++) \
  559. memcpy(dst + j*stride, v + size - 1 - j, size); \
  560. }
  561. def_diag_downright(8)
  562. def_diag_downright(16)
  563. def_diag_downright(32)
  564. static void vert_right_4x4_c(uint8_t *dst, ptrdiff_t stride,
  565. const uint8_t *left, const uint8_t *top)
  566. {
  567. int tl = top[-1], a0 = top[0], a1 = top[1], a2 = top[2], a3 = top[3],
  568. l0 = left[3], l1 = left[2], l2 = left[1];
  569. DST(0,3) = (l0 + l1 * 2 + l2 + 2) >> 2;
  570. DST(0,2) = (tl + l0 * 2 + l1 + 2) >> 2;
  571. DST(0,0) = DST(1,2) = (tl + a0 + 1) >> 1;
  572. DST(0,1) = DST(1,3) = (l0 + tl * 2 + a0 + 2) >> 2;
  573. DST(1,0) = DST(2,2) = (a0 + a1 + 1) >> 1;
  574. DST(1,1) = DST(2,3) = (tl + a0 * 2 + a1 + 2) >> 2;
  575. DST(2,0) = DST(3,2) = (a1 + a2 + 1) >> 1;
  576. DST(2,1) = DST(3,3) = (a0 + a1 * 2 + a2 + 2) >> 2;
  577. DST(3,0) = (a2 + a3 + 1) >> 1;
  578. DST(3,1) = (a1 + a2 * 2 + a3 + 2) >> 2;
  579. }
  580. #define def_vert_right(size) \
  581. static void vert_right_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  582. const uint8_t *left, const uint8_t *top) \
  583. { \
  584. int i, j; \
  585. uint8_t ve[size + size/2 - 1], vo[size + size/2 - 1]; \
  586. \
  587. for (i = 0; i < size/2 - 2; i++) { \
  588. vo[i] = (left[i*2 + 3] + left[i*2 + 2] * 2 + left[i*2 + 1] + 2) >> 2; \
  589. ve[i] = (left[i*2 + 4] + left[i*2 + 3] * 2 + left[i*2 + 2] + 2) >> 2; \
  590. } \
  591. vo[size/2 - 2] = (left[size - 1] + left[size - 2] * 2 + left[size - 3] + 2) >> 2; \
  592. ve[size/2 - 2] = (top[-1] + left[size - 1] * 2 + left[size - 2] + 2) >> 2; \
  593. \
  594. ve[size/2 - 1] = (top[-1] + top[0] + 1) >> 1; \
  595. vo[size/2 - 1] = (left[size - 1] + top[-1] * 2 + top[0] + 2) >> 2; \
  596. for (i = 0; i < size - 1; i++) { \
  597. ve[size/2 + i] = (top[i] + top[i + 1] + 1) >> 1; \
  598. vo[size/2 + i] = (top[i - 1] + top[i] * 2 + top[i + 1] + 2) >> 2; \
  599. } \
  600. \
  601. for (j = 0; j < size / 2; j++) { \
  602. memcpy(dst + j*2 *stride, ve + size/2 - 1 - j, size); \
  603. memcpy(dst + (j*2 + 1)*stride, vo + size/2 - 1 - j, size); \
  604. } \
  605. }
  606. def_vert_right(8)
  607. def_vert_right(16)
  608. def_vert_right(32)
  609. static void hor_down_4x4_c(uint8_t *dst, ptrdiff_t stride,
  610. const uint8_t *left, const uint8_t *top)
  611. {
  612. int l0 = left[3], l1 = left[2], l2 = left[1], l3 = left[0],
  613. tl = top[-1], a0 = top[0], a1 = top[1], a2 = top[2];
  614. DST(2,0) = (tl + a0 * 2 + a1 + 2) >> 2;
  615. DST(3,0) = (a0 + a1 * 2 + a2 + 2) >> 2;
  616. DST(0,0) = DST(2,1) = (tl + l0 + 1) >> 1;
  617. DST(1,0) = DST(3,1) = (a0 + tl * 2 + l0 + 2) >> 2;
  618. DST(0,1) = DST(2,2) = (l0 + l1 + 1) >> 1;
  619. DST(1,1) = DST(3,2) = (tl + l0 * 2 + l1 + 2) >> 2;
  620. DST(0,2) = DST(2,3) = (l1 + l2 + 1) >> 1;
  621. DST(1,2) = DST(3,3) = (l0 + l1 * 2 + l2 + 2) >> 2;
  622. DST(0,3) = (l2 + l3 + 1) >> 1;
  623. DST(1,3) = (l1 + l2 * 2 + l3 + 2) >> 2;
  624. }
  625. #define def_hor_down(size) \
  626. static void hor_down_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  627. const uint8_t *left, const uint8_t *top) \
  628. { \
  629. int i, j; \
  630. uint8_t v[size * 3 - 2]; \
  631. \
  632. for (i = 0; i < size - 2; i++) { \
  633. v[i*2 ] = (left[i + 1] + left[i + 0] + 1) >> 1; \
  634. v[i*2 + 1] = (left[i + 2] + left[i + 1] * 2 + left[i + 0] + 2) >> 2; \
  635. v[size*2 + i] = (top[i - 1] + top[i] * 2 + top[i + 1] + 2) >> 2; \
  636. } \
  637. v[size*2 - 2] = (top[-1] + left[size - 1] + 1) >> 1; \
  638. v[size*2 - 4] = (left[size - 1] + left[size - 2] + 1) >> 1; \
  639. v[size*2 - 1] = (top[0] + top[-1] * 2 + left[size - 1] + 2) >> 2; \
  640. v[size*2 - 3] = (top[-1] + left[size - 1] * 2 + left[size - 2] + 2) >> 2; \
  641. \
  642. for (j = 0; j < size; j++) \
  643. memcpy(dst + j*stride, v + size*2 - 2 - j*2, size); \
  644. }
  645. def_hor_down(8)
  646. def_hor_down(16)
  647. def_hor_down(32)
  648. static void vert_left_4x4_c(uint8_t *dst, ptrdiff_t stride,
  649. const uint8_t *left, const uint8_t *top)
  650. {
  651. int a0 = top[0], a1 = top[1], a2 = top[2], a3 = top[3],
  652. a4 = top[4], a5 = top[5], a6 = top[6];
  653. DST(0,0) = (a0 + a1 + 1) >> 1;
  654. DST(0,1) = (a0 + a1 * 2 + a2 + 2) >> 2;
  655. DST(1,0) = DST(0,2) = (a1 + a2 + 1) >> 1;
  656. DST(1,1) = DST(0,3) = (a1 + a2 * 2 + a3 + 2) >> 2;
  657. DST(2,0) = DST(1,2) = (a2 + a3 + 1) >> 1;
  658. DST(2,1) = DST(1,3) = (a2 + a3 * 2 + a4 + 2) >> 2;
  659. DST(3,0) = DST(2,2) = (a3 + a4 + 1) >> 1;
  660. DST(3,1) = DST(2,3) = (a3 + a4 * 2 + a5 + 2) >> 2;
  661. DST(3,2) = (a4 + a5 + 1) >> 1;
  662. DST(3,3) = (a4 + a5 * 2 + a6 + 2) >> 2;
  663. }
  664. #define def_vert_left(size) \
  665. static void vert_left_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  666. const uint8_t *left, const uint8_t *top) \
  667. { \
  668. int i, j; \
  669. uint8_t ve[size - 1], vo[size - 1]; \
  670. \
  671. for (i = 0; i < size - 2; i++) { \
  672. ve[i] = (top[i] + top[i + 1] + 1) >> 1; \
  673. vo[i] = (top[i] + top[i + 1] * 2 + top[i + 2] + 2) >> 2; \
  674. } \
  675. ve[size - 2] = (top[size - 2] + top[size - 1] + 1) >> 1; \
  676. vo[size - 2] = (top[size - 2] + top[size - 1] * 3 + 2) >> 2; \
  677. \
  678. for (j = 0; j < size / 2; j++) { \
  679. memcpy(dst + j*2 * stride, ve + j, size - j - 1); \
  680. memset(dst + j*2 * stride + size - j - 1, top[size - 1], j + 1); \
  681. memcpy(dst + (j*2 + 1) * stride, vo + j, size - j - 1); \
  682. memset(dst + (j*2 + 1) * stride + size - j - 1, top[size - 1], j + 1); \
  683. } \
  684. }
  685. def_vert_left(8)
  686. def_vert_left(16)
  687. def_vert_left(32)
  688. static void hor_up_4x4_c(uint8_t *dst, ptrdiff_t stride,
  689. const uint8_t *left, const uint8_t *top)
  690. {
  691. int l0 = left[3], l1 = left[2], l2 = left[1], l3 = left[0];
  692. DST(0,0) = (l0 + l1 + 1) >> 1;
  693. DST(1,0) = (l0 + l1 * 2 + l2 + 2) >> 2;
  694. DST(0,1) = DST(2,0) = (l1 + l2 + 1) >> 1;
  695. DST(1,1) = DST(3,0) = (l1 + l2 * 2 + l3 + 2) >> 2;
  696. DST(0,2) = DST(2,1) = (l2 + l3 + 1) >> 1;
  697. DST(1,2) = DST(3,1) = (l2 + l3 * 3 + 2) >> 2;
  698. DST(0,3) = DST(1,3) = DST(2,2) = DST(2,3) = DST(3,2) = DST(3,3) = l3;
  699. }
  700. #define def_hor_up(size) \
  701. static void hor_up_##size##x##size##_c(uint8_t *dst, ptrdiff_t stride, \
  702. const uint8_t *left, const uint8_t *top) \
  703. { \
  704. int i, j; \
  705. uint8_t v[size*2 - 2]; \
  706. \
  707. for (i = 0; i < size - 2; i++) { \
  708. v[i*2 ] = (left[size - i - 1] + left[size - i - 2] + 1) >> 1; \
  709. v[i*2 + 1] = (left[size - i - 1] + left[size - i - 2] * 2 + left[size - i - 3] + 2) >> 2; \
  710. } \
  711. v[size*2 - 4] = (left[1] + left[0] + 1) >> 1; \
  712. v[size*2 - 3] = (left[1] + left[0] * 3 + 2) >> 2; \
  713. \
  714. for (j = 0; j < size / 2; j++) \
  715. memcpy(dst + j*stride, v + j*2, size); \
  716. for (j = size / 2; j < size; j++) { \
  717. memcpy(dst + j*stride, v + j*2, size*2 - 2 - j*2); \
  718. memset(dst + j*stride + size*2 - 2 - j*2, left[0], \
  719. 2 + j*2 - size); \
  720. } \
  721. }
  722. def_hor_up(8)
  723. def_hor_up(16)
  724. def_hor_up(32)
  725. #undef DST
  726. static av_cold void vp9dsp_intrapred_init(VP9DSPContext *dsp)
  727. {
  728. #define init_intra_pred(tx, sz) \
  729. dsp->intra_pred[tx][VERT_PRED] = vert_##sz##_c; \
  730. dsp->intra_pred[tx][HOR_PRED] = hor_##sz##_c; \
  731. dsp->intra_pred[tx][DC_PRED] = dc_##sz##_c; \
  732. dsp->intra_pred[tx][DIAG_DOWN_LEFT_PRED] = diag_downleft_##sz##_c; \
  733. dsp->intra_pred[tx][DIAG_DOWN_RIGHT_PRED] = diag_downright_##sz##_c; \
  734. dsp->intra_pred[tx][VERT_RIGHT_PRED] = vert_right_##sz##_c; \
  735. dsp->intra_pred[tx][HOR_DOWN_PRED] = hor_down_##sz##_c; \
  736. dsp->intra_pred[tx][VERT_LEFT_PRED] = vert_left_##sz##_c; \
  737. dsp->intra_pred[tx][HOR_UP_PRED] = hor_up_##sz##_c; \
  738. dsp->intra_pred[tx][TM_VP8_PRED] = tm_##sz##_c; \
  739. dsp->intra_pred[tx][LEFT_DC_PRED] = dc_left_##sz##_c; \
  740. dsp->intra_pred[tx][TOP_DC_PRED] = dc_top_##sz##_c; \
  741. dsp->intra_pred[tx][DC_128_PRED] = dc_128_##sz##_c; \
  742. dsp->intra_pred[tx][DC_127_PRED] = dc_127_##sz##_c; \
  743. dsp->intra_pred[tx][DC_129_PRED] = dc_129_##sz##_c
  744. init_intra_pred(TX_4X4, 4x4);
  745. init_intra_pred(TX_8X8, 8x8);
  746. init_intra_pred(TX_16X16, 16x16);
  747. init_intra_pred(TX_32X32, 32x32);
  748. #undef init_intra_pred
  749. }
  750. #define itxfm_wrapper(type_a, type_b, sz, bits, has_dconly) \
  751. static void type_a##_##type_b##_##sz##x##sz##_add_c(uint8_t *dst, \
  752. ptrdiff_t stride, \
  753. int16_t *block, int eob) \
  754. { \
  755. int i, j; \
  756. int16_t tmp[sz * sz], out[sz]; \
  757. \
  758. if (has_dconly && eob == 1) { \
  759. const int t = (((block[0] * 11585 + (1 << 13)) >> 14) \
  760. * 11585 + (1 << 13)) >> 14; \
  761. block[0] = 0; \
  762. for (i = 0; i < sz; i++) { \
  763. for (j = 0; j < sz; j++) \
  764. dst[j * stride] = av_clip_uint8(dst[j * stride] + \
  765. (bits ? \
  766. (t + (1 << (bits - 1))) >> bits : \
  767. t)); \
  768. dst++; \
  769. } \
  770. return; \
  771. } \
  772. \
  773. for (i = 0; i < sz; i++) \
  774. type_a##sz##_1d(block + i, sz, tmp + i * sz, 0); \
  775. memset(block, 0, sz * sz * sizeof(*block)); \
  776. for (i = 0; i < sz; i++) { \
  777. type_b##sz##_1d(tmp + i, sz, out, 1); \
  778. for (j = 0; j < sz; j++) \
  779. dst[j * stride] = av_clip_uint8(dst[j * stride] + \
  780. (bits ? \
  781. (out[j] + (1 << (bits - 1))) >> bits : \
  782. out[j])); \
  783. dst++; \
  784. } \
  785. }
  786. #define itxfm_wrap(sz, bits) \
  787. itxfm_wrapper(idct, idct, sz, bits, 1) \
  788. itxfm_wrapper(iadst, idct, sz, bits, 0) \
  789. itxfm_wrapper(idct, iadst, sz, bits, 0) \
  790. itxfm_wrapper(iadst, iadst, sz, bits, 0)
  791. #define IN(x) in[(x) * stride]
  792. static av_always_inline void idct4_1d(const int16_t *in, ptrdiff_t stride,
  793. int16_t *out, int pass)
  794. {
  795. int t0, t1, t2, t3;
  796. t0 = ((IN(0) + IN(2)) * 11585 + (1 << 13)) >> 14;
  797. t1 = ((IN(0) - IN(2)) * 11585 + (1 << 13)) >> 14;
  798. t2 = (IN(1) * 6270 - IN(3) * 15137 + (1 << 13)) >> 14;
  799. t3 = (IN(1) * 15137 + IN(3) * 6270 + (1 << 13)) >> 14;
  800. out[0] = t0 + t3;
  801. out[1] = t1 + t2;
  802. out[2] = t1 - t2;
  803. out[3] = t0 - t3;
  804. }
  805. static av_always_inline void iadst4_1d(const int16_t *in, ptrdiff_t stride,
  806. int16_t *out, int pass)
  807. {
  808. int t0, t1, t2, t3;
  809. t0 = 5283 * IN(0) + 15212 * IN(2) + 9929 * IN(3);
  810. t1 = 9929 * IN(0) - 5283 * IN(2) - 15212 * IN(3);
  811. t2 = 13377 * (IN(0) - IN(2) + IN(3));
  812. t3 = 13377 * IN(1);
  813. out[0] = (t0 + t3 + (1 << 13)) >> 14;
  814. out[1] = (t1 + t3 + (1 << 13)) >> 14;
  815. out[2] = (t2 + (1 << 13)) >> 14;
  816. out[3] = (t0 + t1 - t3 + (1 << 13)) >> 14;
  817. }
  818. itxfm_wrap(4, 4)
  819. static av_always_inline void idct8_1d(const int16_t *in, ptrdiff_t stride,
  820. int16_t *out, int pass)
  821. {
  822. int t0, t0a, t1, t1a, t2, t2a, t3, t3a, t4, t4a, t5, t5a, t6, t6a, t7, t7a;
  823. t0a = ((IN(0) + IN(4)) * 11585 + (1 << 13)) >> 14;
  824. t1a = ((IN(0) - IN(4)) * 11585 + (1 << 13)) >> 14;
  825. t2a = (IN(2) * 6270 - IN(6) * 15137 + (1 << 13)) >> 14;
  826. t3a = (IN(2) * 15137 + IN(6) * 6270 + (1 << 13)) >> 14;
  827. t4a = (IN(1) * 3196 - IN(7) * 16069 + (1 << 13)) >> 14;
  828. t5a = (IN(5) * 13623 - IN(3) * 9102 + (1 << 13)) >> 14;
  829. t6a = (IN(5) * 9102 + IN(3) * 13623 + (1 << 13)) >> 14;
  830. t7a = (IN(1) * 16069 + IN(7) * 3196 + (1 << 13)) >> 14;
  831. t0 = t0a + t3a;
  832. t1 = t1a + t2a;
  833. t2 = t1a - t2a;
  834. t3 = t0a - t3a;
  835. t4 = t4a + t5a;
  836. t5a = t4a - t5a;
  837. t7 = t7a + t6a;
  838. t6a = t7a - t6a;
  839. t5 = ((t6a - t5a) * 11585 + (1 << 13)) >> 14;
  840. t6 = ((t6a + t5a) * 11585 + (1 << 13)) >> 14;
  841. out[0] = t0 + t7;
  842. out[1] = t1 + t6;
  843. out[2] = t2 + t5;
  844. out[3] = t3 + t4;
  845. out[4] = t3 - t4;
  846. out[5] = t2 - t5;
  847. out[6] = t1 - t6;
  848. out[7] = t0 - t7;
  849. }
  850. static av_always_inline void iadst8_1d(const int16_t *in, ptrdiff_t stride,
  851. int16_t *out, int pass)
  852. {
  853. int t0, t0a, t1, t1a, t2, t2a, t3, t3a, t4, t4a, t5, t5a, t6, t6a, t7, t7a;
  854. t0a = 16305 * IN(7) + 1606 * IN(0);
  855. t1a = 1606 * IN(7) - 16305 * IN(0);
  856. t2a = 14449 * IN(5) + 7723 * IN(2);
  857. t3a = 7723 * IN(5) - 14449 * IN(2);
  858. t4a = 10394 * IN(3) + 12665 * IN(4);
  859. t5a = 12665 * IN(3) - 10394 * IN(4);
  860. t6a = 4756 * IN(1) + 15679 * IN(6);
  861. t7a = 15679 * IN(1) - 4756 * IN(6);
  862. t0 = (t0a + t4a + (1 << 13)) >> 14;
  863. t1 = (t1a + t5a + (1 << 13)) >> 14;
  864. t2 = (t2a + t6a + (1 << 13)) >> 14;
  865. t3 = (t3a + t7a + (1 << 13)) >> 14;
  866. t4 = (t0a - t4a + (1 << 13)) >> 14;
  867. t5 = (t1a - t5a + (1 << 13)) >> 14;
  868. t6 = (t2a - t6a + (1 << 13)) >> 14;
  869. t7 = (t3a - t7a + (1 << 13)) >> 14;
  870. t4a = 15137 * t4 + 6270 * t5;
  871. t5a = 6270 * t4 - 15137 * t5;
  872. t6a = 15137 * t7 - 6270 * t6;
  873. t7a = 6270 * t7 + 15137 * t6;
  874. out[0] = t0 + t2;
  875. out[7] = -(t1 + t3);
  876. t2 = t0 - t2;
  877. t3 = t1 - t3;
  878. out[1] = -((t4a + t6a + (1 << 13)) >> 14);
  879. out[6] = (t5a + t7a + (1 << 13)) >> 14;
  880. t6 = (t4a - t6a + (1 << 13)) >> 14;
  881. t7 = (t5a - t7a + (1 << 13)) >> 14;
  882. out[3] = -(((t2 + t3) * 11585 + (1 << 13)) >> 14);
  883. out[4] = ((t2 - t3) * 11585 + (1 << 13)) >> 14;
  884. out[2] = ((t6 + t7) * 11585 + (1 << 13)) >> 14;
  885. out[5] = -(((t6 - t7) * 11585 + (1 << 13)) >> 14);
  886. }
  887. itxfm_wrap(8, 5)
  888. static av_always_inline void idct16_1d(const int16_t *in, ptrdiff_t stride,
  889. int16_t *out, int pass)
  890. {
  891. int t0, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, t11, t12, t13, t14, t15;
  892. int t0a, t1a, t2a, t3a, t4a, t5a, t6a, t7a;
  893. int t8a, t9a, t10a, t11a, t12a, t13a, t14a, t15a;
  894. t0a = ((IN(0) + IN(8)) * 11585 + (1 << 13)) >> 14;
  895. t1a = ((IN(0) - IN(8)) * 11585 + (1 << 13)) >> 14;
  896. t2a = (IN(4) * 6270 - IN(12) * 15137 + (1 << 13)) >> 14;
  897. t3a = (IN(4) * 15137 + IN(12) * 6270 + (1 << 13)) >> 14;
  898. t4a = (IN(2) * 3196 - IN(14) * 16069 + (1 << 13)) >> 14;
  899. t7a = (IN(2) * 16069 + IN(14) * 3196 + (1 << 13)) >> 14;
  900. t5a = (IN(10) * 13623 - IN(6) * 9102 + (1 << 13)) >> 14;
  901. t6a = (IN(10) * 9102 + IN(6) * 13623 + (1 << 13)) >> 14;
  902. t8a = (IN(1) * 1606 - IN(15) * 16305 + (1 << 13)) >> 14;
  903. t15a = (IN(1) * 16305 + IN(15) * 1606 + (1 << 13)) >> 14;
  904. t9a = (IN(9) * 12665 - IN(7) * 10394 + (1 << 13)) >> 14;
  905. t14a = (IN(9) * 10394 + IN(7) * 12665 + (1 << 13)) >> 14;
  906. t10a = (IN(5) * 7723 - IN(11) * 14449 + (1 << 13)) >> 14;
  907. t13a = (IN(5) * 14449 + IN(11) * 7723 + (1 << 13)) >> 14;
  908. t11a = (IN(13) * 15679 - IN(3) * 4756 + (1 << 13)) >> 14;
  909. t12a = (IN(13) * 4756 + IN(3) * 15679 + (1 << 13)) >> 14;
  910. t0 = t0a + t3a;
  911. t1 = t1a + t2a;
  912. t2 = t1a - t2a;
  913. t3 = t0a - t3a;
  914. t4 = t4a + t5a;
  915. t5 = t4a - t5a;
  916. t6 = t7a - t6a;
  917. t7 = t7a + t6a;
  918. t8 = t8a + t9a;
  919. t9 = t8a - t9a;
  920. t10 = t11a - t10a;
  921. t11 = t11a + t10a;
  922. t12 = t12a + t13a;
  923. t13 = t12a - t13a;
  924. t14 = t15a - t14a;
  925. t15 = t15a + t14a;
  926. t5a = ((t6 - t5) * 11585 + (1 << 13)) >> 14;
  927. t6a = ((t6 + t5) * 11585 + (1 << 13)) >> 14;
  928. t9a = ( t14 * 6270 - t9 * 15137 + (1 << 13)) >> 14;
  929. t14a = ( t14 * 15137 + t9 * 6270 + (1 << 13)) >> 14;
  930. t10a = (-(t13 * 15137 + t10 * 6270) + (1 << 13)) >> 14;
  931. t13a = ( t13 * 6270 - t10 * 15137 + (1 << 13)) >> 14;
  932. t0a = t0 + t7;
  933. t1a = t1 + t6a;
  934. t2a = t2 + t5a;
  935. t3a = t3 + t4;
  936. t4 = t3 - t4;
  937. t5 = t2 - t5a;
  938. t6 = t1 - t6a;
  939. t7 = t0 - t7;
  940. t8a = t8 + t11;
  941. t9 = t9a + t10a;
  942. t10 = t9a - t10a;
  943. t11a = t8 - t11;
  944. t12a = t15 - t12;
  945. t13 = t14a - t13a;
  946. t14 = t14a + t13a;
  947. t15a = t15 + t12;
  948. t10a = ((t13 - t10) * 11585 + (1 << 13)) >> 14;
  949. t13a = ((t13 + t10) * 11585 + (1 << 13)) >> 14;
  950. t11 = ((t12a - t11a) * 11585 + (1 << 13)) >> 14;
  951. t12 = ((t12a + t11a) * 11585 + (1 << 13)) >> 14;
  952. out[ 0] = t0a + t15a;
  953. out[ 1] = t1a + t14;
  954. out[ 2] = t2a + t13a;
  955. out[ 3] = t3a + t12;
  956. out[ 4] = t4 + t11;
  957. out[ 5] = t5 + t10a;
  958. out[ 6] = t6 + t9;
  959. out[ 7] = t7 + t8a;
  960. out[ 8] = t7 - t8a;
  961. out[ 9] = t6 - t9;
  962. out[10] = t5 - t10a;
  963. out[11] = t4 - t11;
  964. out[12] = t3a - t12;
  965. out[13] = t2a - t13a;
  966. out[14] = t1a - t14;
  967. out[15] = t0a - t15a;
  968. }
  969. static av_always_inline void iadst16_1d(const int16_t *in, ptrdiff_t stride,
  970. int16_t *out, int pass)
  971. {
  972. int t0, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, t11, t12, t13, t14, t15;
  973. int t0a, t1a, t2a, t3a, t4a, t5a, t6a, t7a;
  974. int t8a, t9a, t10a, t11a, t12a, t13a, t14a, t15a;
  975. t0 = IN(15) * 16364 + IN(0) * 804;
  976. t1 = IN(15) * 804 - IN(0) * 16364;
  977. t2 = IN(13) * 15893 + IN(2) * 3981;
  978. t3 = IN(13) * 3981 - IN(2) * 15893;
  979. t4 = IN(11) * 14811 + IN(4) * 7005;
  980. t5 = IN(11) * 7005 - IN(4) * 14811;
  981. t6 = IN(9) * 13160 + IN(6) * 9760;
  982. t7 = IN(9) * 9760 - IN(6) * 13160;
  983. t8 = IN(7) * 11003 + IN(8) * 12140;
  984. t9 = IN(7) * 12140 - IN(8) * 11003;
  985. t10 = IN(5) * 8423 + IN(10) * 14053;
  986. t11 = IN(5) * 14053 - IN(10) * 8423;
  987. t12 = IN(3) * 5520 + IN(12) * 15426;
  988. t13 = IN(3) * 15426 - IN(12) * 5520;
  989. t14 = IN(1) * 2404 + IN(14) * 16207;
  990. t15 = IN(1) * 16207 - IN(14) * 2404;
  991. t0a = (t0 + t8 + (1 << 13)) >> 14;
  992. t1a = (t1 + t9 + (1 << 13)) >> 14;
  993. t2a = (t2 + t10 + (1 << 13)) >> 14;
  994. t3a = (t3 + t11 + (1 << 13)) >> 14;
  995. t4a = (t4 + t12 + (1 << 13)) >> 14;
  996. t5a = (t5 + t13 + (1 << 13)) >> 14;
  997. t6a = (t6 + t14 + (1 << 13)) >> 14;
  998. t7a = (t7 + t15 + (1 << 13)) >> 14;
  999. t8a = (t0 - t8 + (1 << 13)) >> 14;
  1000. t9a = (t1 - t9 + (1 << 13)) >> 14;
  1001. t10a = (t2 - t10 + (1 << 13)) >> 14;
  1002. t11a = (t3 - t11 + (1 << 13)) >> 14;
  1003. t12a = (t4 - t12 + (1 << 13)) >> 14;
  1004. t13a = (t5 - t13 + (1 << 13)) >> 14;
  1005. t14a = (t6 - t14 + (1 << 13)) >> 14;
  1006. t15a = (t7 - t15 + (1 << 13)) >> 14;
  1007. t8 = t8a * 16069 + t9a * 3196;
  1008. t9 = t8a * 3196 - t9a * 16069;
  1009. t10 = t10a * 9102 + t11a * 13623;
  1010. t11 = t10a * 13623 - t11a * 9102;
  1011. t12 = t13a * 16069 - t12a * 3196;
  1012. t13 = t13a * 3196 + t12a * 16069;
  1013. t14 = t15a * 9102 - t14a * 13623;
  1014. t15 = t15a * 13623 + t14a * 9102;
  1015. t0 = t0a + t4a;
  1016. t1 = t1a + t5a;
  1017. t2 = t2a + t6a;
  1018. t3 = t3a + t7a;
  1019. t4 = t0a - t4a;
  1020. t5 = t1a - t5a;
  1021. t6 = t2a - t6a;
  1022. t7 = t3a - t7a;
  1023. t8a = (t8 + t12 + (1 << 13)) >> 14;
  1024. t9a = (t9 + t13 + (1 << 13)) >> 14;
  1025. t10a = (t10 + t14 + (1 << 13)) >> 14;
  1026. t11a = (t11 + t15 + (1 << 13)) >> 14;
  1027. t12a = (t8 - t12 + (1 << 13)) >> 14;
  1028. t13a = (t9 - t13 + (1 << 13)) >> 14;
  1029. t14a = (t10 - t14 + (1 << 13)) >> 14;
  1030. t15a = (t11 - t15 + (1 << 13)) >> 14;
  1031. t4a = t4 * 15137 + t5 * 6270;
  1032. t5a = t4 * 6270 - t5 * 15137;
  1033. t6a = t7 * 15137 - t6 * 6270;
  1034. t7a = t7 * 6270 + t6 * 15137;
  1035. t12 = t12a * 15137 + t13a * 6270;
  1036. t13 = t12a * 6270 - t13a * 15137;
  1037. t14 = t15a * 15137 - t14a * 6270;
  1038. t15 = t15a * 6270 + t14a * 15137;
  1039. out[ 0] = t0 + t2;
  1040. out[15] = -(t1 + t3);
  1041. t2a = t0 - t2;
  1042. t3a = t1 - t3;
  1043. out[ 3] = -((t4a + t6a + (1 << 13)) >> 14);
  1044. out[12] = (t5a + t7a + (1 << 13)) >> 14;
  1045. t6 = (t4a - t6a + (1 << 13)) >> 14;
  1046. t7 = (t5a - t7a + (1 << 13)) >> 14;
  1047. out[ 1] = -(t8a + t10a);
  1048. out[14] = t9a + t11a;
  1049. t10 = t8a - t10a;
  1050. t11 = t9a - t11a;
  1051. out[ 2] = (t12 + t14 + (1 << 13)) >> 14;
  1052. out[13] = -((t13 + t15 + (1 << 13)) >> 14);
  1053. t14a = (t12 - t14 + (1 << 13)) >> 14;
  1054. t15a = (t13 - t15 + (1 << 13)) >> 14;
  1055. out[ 7] = ((t2a + t3a) * -11585 + (1 << 13)) >> 14;
  1056. out[ 8] = ((t2a - t3a) * 11585 + (1 << 13)) >> 14;
  1057. out[ 4] = ((t7 + t6) * 11585 + (1 << 13)) >> 14;
  1058. out[11] = ((t7 - t6) * 11585 + (1 << 13)) >> 14;
  1059. out[ 6] = ((t11 + t10) * 11585 + (1 << 13)) >> 14;
  1060. out[ 9] = ((t11 - t10) * 11585 + (1 << 13)) >> 14;
  1061. out[ 5] = ((t14a + t15a) * -11585 + (1 << 13)) >> 14;
  1062. out[10] = ((t14a - t15a) * 11585 + (1 << 13)) >> 14;
  1063. }
  1064. itxfm_wrap(16, 6)
  1065. static av_always_inline void idct32_1d(const int16_t *in, ptrdiff_t stride,
  1066. int16_t *out, int pass)
  1067. {
  1068. int t0a = ((IN(0) + IN(16)) * 11585 + (1 << 13)) >> 14;
  1069. int t1a = ((IN(0) - IN(16)) * 11585 + (1 << 13)) >> 14;
  1070. int t2a = (IN( 8) * 6270 - IN(24) * 15137 + (1 << 13)) >> 14;
  1071. int t3a = (IN( 8) * 15137 + IN(24) * 6270 + (1 << 13)) >> 14;
  1072. int t4a = (IN( 4) * 3196 - IN(28) * 16069 + (1 << 13)) >> 14;
  1073. int t7a = (IN( 4) * 16069 + IN(28) * 3196 + (1 << 13)) >> 14;
  1074. int t5a = (IN(20) * 13623 - IN(12) * 9102 + (1 << 13)) >> 14;
  1075. int t6a = (IN(20) * 9102 + IN(12) * 13623 + (1 << 13)) >> 14;
  1076. int t8a = (IN( 2) * 1606 - IN(30) * 16305 + (1 << 13)) >> 14;
  1077. int t15a = (IN( 2) * 16305 + IN(30) * 1606 + (1 << 13)) >> 14;
  1078. int t9a = (IN(18) * 12665 - IN(14) * 10394 + (1 << 13)) >> 14;
  1079. int t14a = (IN(18) * 10394 + IN(14) * 12665 + (1 << 13)) >> 14;
  1080. int t10a = (IN(10) * 7723 - IN(22) * 14449 + (1 << 13)) >> 14;
  1081. int t13a = (IN(10) * 14449 + IN(22) * 7723 + (1 << 13)) >> 14;
  1082. int t11a = (IN(26) * 15679 - IN( 6) * 4756 + (1 << 13)) >> 14;
  1083. int t12a = (IN(26) * 4756 + IN( 6) * 15679 + (1 << 13)) >> 14;
  1084. int t16a = (IN( 1) * 804 - IN(31) * 16364 + (1 << 13)) >> 14;
  1085. int t31a = (IN( 1) * 16364 + IN(31) * 804 + (1 << 13)) >> 14;
  1086. int t17a = (IN(17) * 12140 - IN(15) * 11003 + (1 << 13)) >> 14;
  1087. int t30a = (IN(17) * 11003 + IN(15) * 12140 + (1 << 13)) >> 14;
  1088. int t18a = (IN( 9) * 7005 - IN(23) * 14811 + (1 << 13)) >> 14;
  1089. int t29a = (IN( 9) * 14811 + IN(23) * 7005 + (1 << 13)) >> 14;
  1090. int t19a = (IN(25) * 15426 - IN( 7) * 5520 + (1 << 13)) >> 14;
  1091. int t28a = (IN(25) * 5520 + IN( 7) * 15426 + (1 << 13)) >> 14;
  1092. int t20a = (IN( 5) * 3981 - IN(27) * 15893 + (1 << 13)) >> 14;
  1093. int t27a = (IN( 5) * 15893 + IN(27) * 3981 + (1 << 13)) >> 14;
  1094. int t21a = (IN(21) * 14053 - IN(11) * 8423 + (1 << 13)) >> 14;
  1095. int t26a = (IN(21) * 8423 + IN(11) * 14053 + (1 << 13)) >> 14;
  1096. int t22a = (IN(13) * 9760 - IN(19) * 13160 + (1 << 13)) >> 14;
  1097. int t25a = (IN(13) * 13160 + IN(19) * 9760 + (1 << 13)) >> 14;
  1098. int t23a = (IN(29) * 16207 - IN( 3) * 2404 + (1 << 13)) >> 14;
  1099. int t24a = (IN(29) * 2404 + IN( 3) * 16207 + (1 << 13)) >> 14;
  1100. int t0 = t0a + t3a;
  1101. int t1 = t1a + t2a;
  1102. int t2 = t1a - t2a;
  1103. int t3 = t0a - t3a;
  1104. int t4 = t4a + t5a;
  1105. int t5 = t4a - t5a;
  1106. int t6 = t7a - t6a;
  1107. int t7 = t7a + t6a;
  1108. int t8 = t8a + t9a;
  1109. int t9 = t8a - t9a;
  1110. int t10 = t11a - t10a;
  1111. int t11 = t11a + t10a;
  1112. int t12 = t12a + t13a;
  1113. int t13 = t12a - t13a;
  1114. int t14 = t15a - t14a;
  1115. int t15 = t15a + t14a;
  1116. int t16 = t16a + t17a;
  1117. int t17 = t16a - t17a;
  1118. int t18 = t19a - t18a;
  1119. int t19 = t19a + t18a;
  1120. int t20 = t20a + t21a;
  1121. int t21 = t20a - t21a;
  1122. int t22 = t23a - t22a;
  1123. int t23 = t23a + t22a;
  1124. int t24 = t24a + t25a;
  1125. int t25 = t24a - t25a;
  1126. int t26 = t27a - t26a;
  1127. int t27 = t27a + t26a;
  1128. int t28 = t28a + t29a;
  1129. int t29 = t28a - t29a;
  1130. int t30 = t31a - t30a;
  1131. int t31 = t31a + t30a;
  1132. t5a = ((t6 - t5) * 11585 + (1 << 13)) >> 14;
  1133. t6a = ((t6 + t5) * 11585 + (1 << 13)) >> 14;
  1134. t9a = ( t14 * 6270 - t9 * 15137 + (1 << 13)) >> 14;
  1135. t14a = ( t14 * 15137 + t9 * 6270 + (1 << 13)) >> 14;
  1136. t10a = (-(t13 * 15137 + t10 * 6270) + (1 << 13)) >> 14;
  1137. t13a = ( t13 * 6270 - t10 * 15137 + (1 << 13)) >> 14;
  1138. t17a = ( t30 * 3196 - t17 * 16069 + (1 << 13)) >> 14;
  1139. t30a = ( t30 * 16069 + t17 * 3196 + (1 << 13)) >> 14;
  1140. t18a = (-(t29 * 16069 + t18 * 3196) + (1 << 13)) >> 14;
  1141. t29a = ( t29 * 3196 - t18 * 16069 + (1 << 13)) >> 14;
  1142. t21a = ( t26 * 13623 - t21 * 9102 + (1 << 13)) >> 14;
  1143. t26a = ( t26 * 9102 + t21 * 13623 + (1 << 13)) >> 14;
  1144. t22a = (-(t25 * 9102 + t22 * 13623) + (1 << 13)) >> 14;
  1145. t25a = ( t25 * 13623 - t22 * 9102 + (1 << 13)) >> 14;
  1146. t0a = t0 + t7;
  1147. t1a = t1 + t6a;
  1148. t2a = t2 + t5a;
  1149. t3a = t3 + t4;
  1150. t4a = t3 - t4;
  1151. t5 = t2 - t5a;
  1152. t6 = t1 - t6a;
  1153. t7a = t0 - t7;
  1154. t8a = t8 + t11;
  1155. t9 = t9a + t10a;
  1156. t10 = t9a - t10a;
  1157. t11a = t8 - t11;
  1158. t12a = t15 - t12;
  1159. t13 = t14a - t13a;
  1160. t14 = t14a + t13a;
  1161. t15a = t15 + t12;
  1162. t16a = t16 + t19;
  1163. t17 = t17a + t18a;
  1164. t18 = t17a - t18a;
  1165. t19a = t16 - t19;
  1166. t20a = t23 - t20;
  1167. t21 = t22a - t21a;
  1168. t22 = t22a + t21a;
  1169. t23a = t23 + t20;
  1170. t24a = t24 + t27;
  1171. t25 = t25a + t26a;
  1172. t26 = t25a - t26a;
  1173. t27a = t24 - t27;
  1174. t28a = t31 - t28;
  1175. t29 = t30a - t29a;
  1176. t30 = t30a + t29a;
  1177. t31a = t31 + t28;
  1178. t10a = ((t13 - t10) * 11585 + (1 << 13)) >> 14;
  1179. t13a = ((t13 + t10) * 11585 + (1 << 13)) >> 14;
  1180. t11 = ((t12a - t11a) * 11585 + (1 << 13)) >> 14;
  1181. t12 = ((t12a + t11a) * 11585 + (1 << 13)) >> 14;
  1182. t18a = ( t29 * 6270 - t18 * 15137 + (1 << 13)) >> 14;
  1183. t29a = ( t29 * 15137 + t18 * 6270 + (1 << 13)) >> 14;
  1184. t19 = ( t28a * 6270 - t19a * 15137 + (1 << 13)) >> 14;
  1185. t28 = ( t28a * 15137 + t19a * 6270 + (1 << 13)) >> 14;
  1186. t20 = (-(t27a * 15137 + t20a * 6270) + (1 << 13)) >> 14;
  1187. t27 = ( t27a * 6270 - t20a * 15137 + (1 << 13)) >> 14;
  1188. t21a = (-(t26 * 15137 + t21 * 6270) + (1 << 13)) >> 14;
  1189. t26a = ( t26 * 6270 - t21 * 15137 + (1 << 13)) >> 14;
  1190. t0 = t0a + t15a;
  1191. t1 = t1a + t14;
  1192. t2 = t2a + t13a;
  1193. t3 = t3a + t12;
  1194. t4 = t4a + t11;
  1195. t5a = t5 + t10a;
  1196. t6a = t6 + t9;
  1197. t7 = t7a + t8a;
  1198. t8 = t7a - t8a;
  1199. t9a = t6 - t9;
  1200. t10 = t5 - t10a;
  1201. t11a = t4a - t11;
  1202. t12a = t3a - t12;
  1203. t13 = t2a - t13a;
  1204. t14a = t1a - t14;
  1205. t15 = t0a - t15a;
  1206. t16 = t16a + t23a;
  1207. t17a = t17 + t22;
  1208. t18 = t18a + t21a;
  1209. t19a = t19 + t20;
  1210. t20a = t19 - t20;
  1211. t21 = t18a - t21a;
  1212. t22a = t17 - t22;
  1213. t23 = t16a - t23a;
  1214. t24 = t31a - t24a;
  1215. t25a = t30 - t25;
  1216. t26 = t29a - t26a;
  1217. t27a = t28 - t27;
  1218. t28a = t28 + t27;
  1219. t29 = t29a + t26a;
  1220. t30a = t30 + t25;
  1221. t31 = t31a + t24a;
  1222. t20 = ((t27a - t20a) * 11585 + (1 << 13)) >> 14;
  1223. t27 = ((t27a + t20a) * 11585 + (1 << 13)) >> 14;
  1224. t21a = ((t26 - t21 ) * 11585 + (1 << 13)) >> 14;
  1225. t26a = ((t26 + t21 ) * 11585 + (1 << 13)) >> 14;
  1226. t22 = ((t25a - t22a) * 11585 + (1 << 13)) >> 14;
  1227. t25 = ((t25a + t22a) * 11585 + (1 << 13)) >> 14;
  1228. t23a = ((t24 - t23 ) * 11585 + (1 << 13)) >> 14;
  1229. t24a = ((t24 + t23 ) * 11585 + (1 << 13)) >> 14;
  1230. out[ 0] = t0 + t31;
  1231. out[ 1] = t1 + t30a;
  1232. out[ 2] = t2 + t29;
  1233. out[ 3] = t3 + t28a;
  1234. out[ 4] = t4 + t27;
  1235. out[ 5] = t5a + t26a;
  1236. out[ 6] = t6a + t25;
  1237. out[ 7] = t7 + t24a;
  1238. out[ 8] = t8 + t23a;
  1239. out[ 9] = t9a + t22;
  1240. out[10] = t10 + t21a;
  1241. out[11] = t11a + t20;
  1242. out[12] = t12a + t19a;
  1243. out[13] = t13 + t18;
  1244. out[14] = t14a + t17a;
  1245. out[15] = t15 + t16;
  1246. out[16] = t15 - t16;
  1247. out[17] = t14a - t17a;
  1248. out[18] = t13 - t18;
  1249. out[19] = t12a - t19a;
  1250. out[20] = t11a - t20;
  1251. out[21] = t10 - t21a;
  1252. out[22] = t9a - t22;
  1253. out[23] = t8 - t23a;
  1254. out[24] = t7 - t24a;
  1255. out[25] = t6a - t25;
  1256. out[26] = t5a - t26a;
  1257. out[27] = t4 - t27;
  1258. out[28] = t3 - t28a;
  1259. out[29] = t2 - t29;
  1260. out[30] = t1 - t30a;
  1261. out[31] = t0 - t31;
  1262. }
  1263. itxfm_wrapper(idct, idct, 32, 6, 1)
  1264. static av_always_inline void iwht4_1d(const int16_t *in, ptrdiff_t stride,
  1265. int16_t *out, int pass)
  1266. {
  1267. int t0, t1, t2, t3, t4;
  1268. if (pass == 0) {
  1269. t0 = IN(0) >> 2;
  1270. t1 = IN(3) >> 2;
  1271. t2 = IN(1) >> 2;
  1272. t3 = IN(2) >> 2;
  1273. } else {
  1274. t0 = IN(0);
  1275. t1 = IN(3);
  1276. t2 = IN(1);
  1277. t3 = IN(2);
  1278. }
  1279. t0 += t2;
  1280. t3 -= t1;
  1281. t4 = (t0 - t3) >> 1;
  1282. t1 = t4 - t1;
  1283. t2 = t4 - t2;
  1284. t0 -= t1;
  1285. t3 += t2;
  1286. out[0] = t0;
  1287. out[1] = t1;
  1288. out[2] = t2;
  1289. out[3] = t3;
  1290. }
  1291. itxfm_wrapper(iwht, iwht, 4, 0, 0)
  1292. #undef IN
  1293. #undef itxfm_wrapper
  1294. #undef itxfm_wrap
  1295. static av_cold void vp9dsp_itxfm_init(VP9DSPContext *dsp)
  1296. {
  1297. #define init_itxfm(tx, sz) \
  1298. dsp->itxfm_add[tx][DCT_DCT] = idct_idct_##sz##_add_c; \
  1299. dsp->itxfm_add[tx][DCT_ADST] = iadst_idct_##sz##_add_c; \
  1300. dsp->itxfm_add[tx][ADST_DCT] = idct_iadst_##sz##_add_c; \
  1301. dsp->itxfm_add[tx][ADST_ADST] = iadst_iadst_##sz##_add_c
  1302. #define init_idct(tx, nm) \
  1303. dsp->itxfm_add[tx][DCT_DCT] = \
  1304. dsp->itxfm_add[tx][ADST_DCT] = \
  1305. dsp->itxfm_add[tx][DCT_ADST] = \
  1306. dsp->itxfm_add[tx][ADST_ADST] = nm##_add_c
  1307. init_itxfm(TX_4X4, 4x4);
  1308. init_itxfm(TX_8X8, 8x8);
  1309. init_itxfm(TX_16X16, 16x16);
  1310. init_idct(TX_32X32, idct_idct_32x32);
  1311. init_idct(4 /* lossless */, iwht_iwht_4x4);
  1312. #undef init_itxfm
  1313. #undef init_idct
  1314. }
  1315. static av_always_inline void loop_filter(uint8_t *dst, int E, int I, int H,
  1316. ptrdiff_t stridea, ptrdiff_t strideb,
  1317. int wd)
  1318. {
  1319. int i;
  1320. for (i = 0; i < 8; i++, dst += stridea) {
  1321. int p7, p6, p5, p4;
  1322. int p3 = dst[strideb * -4], p2 = dst[strideb * -3];
  1323. int p1 = dst[strideb * -2], p0 = dst[strideb * -1];
  1324. int q0 = dst[strideb * +0], q1 = dst[strideb * +1];
  1325. int q2 = dst[strideb * +2], q3 = dst[strideb * +3];
  1326. int q4, q5, q6, q7;
  1327. int fm = FFABS(p3 - p2) <= I && FFABS(p2 - p1) <= I &&
  1328. FFABS(p1 - p0) <= I && FFABS(q1 - q0) <= I &&
  1329. FFABS(q2 - q1) <= I && FFABS(q3 - q2) <= I &&
  1330. FFABS(p0 - q0) * 2 + (FFABS(p1 - q1) >> 1) <= E;
  1331. int flat8out, flat8in;
  1332. if (!fm)
  1333. continue;
  1334. if (wd >= 16) {
  1335. p7 = dst[strideb * -8];
  1336. p6 = dst[strideb * -7];
  1337. p5 = dst[strideb * -6];
  1338. p4 = dst[strideb * -5];
  1339. q4 = dst[strideb * +4];
  1340. q5 = dst[strideb * +5];
  1341. q6 = dst[strideb * +6];
  1342. q7 = dst[strideb * +7];
  1343. flat8out = FFABS(p7 - p0) <= 1 && FFABS(p6 - p0) <= 1 &&
  1344. FFABS(p5 - p0) <= 1 && FFABS(p4 - p0) <= 1 &&
  1345. FFABS(q4 - q0) <= 1 && FFABS(q5 - q0) <= 1 &&
  1346. FFABS(q6 - q0) <= 1 && FFABS(q7 - q0) <= 1;
  1347. }
  1348. if (wd >= 8)
  1349. flat8in = FFABS(p3 - p0) <= 1 && FFABS(p2 - p0) <= 1 &&
  1350. FFABS(p1 - p0) <= 1 && FFABS(q1 - q0) <= 1 &&
  1351. FFABS(q2 - q0) <= 1 && FFABS(q3 - q0) <= 1;
  1352. if (wd >= 16 && flat8out && flat8in) {
  1353. dst[strideb * -7] = (p7 + p7 + p7 + p7 + p7 + p7 + p7 + p6 * 2 +
  1354. p5 + p4 + p3 + p2 + p1 + p0 + q0 + 8) >> 4;
  1355. dst[strideb * -6] = (p7 + p7 + p7 + p7 + p7 + p7 + p6 + p5 * 2 +
  1356. p4 + p3 + p2 + p1 + p0 + q0 + q1 + 8) >> 4;
  1357. dst[strideb * -5] = (p7 + p7 + p7 + p7 + p7 + p6 + p5 + p4 * 2 +
  1358. p3 + p2 + p1 + p0 + q0 + q1 + q2 + 8) >> 4;
  1359. dst[strideb * -4] = (p7 + p7 + p7 + p7 + p6 + p5 + p4 + p3 * 2 +
  1360. p2 + p1 + p0 + q0 + q1 + q2 + q3 + 8) >> 4;
  1361. dst[strideb * -3] = (p7 + p7 + p7 + p6 + p5 + p4 + p3 + p2 * 2 +
  1362. p1 + p0 + q0 + q1 + q2 + q3 + q4 + 8) >> 4;
  1363. dst[strideb * -2] = (p7 + p7 + p6 + p5 + p4 + p3 + p2 + p1 * 2 +
  1364. p0 + q0 + q1 + q2 + q3 + q4 + q5 + 8) >> 4;
  1365. dst[strideb * -1] = (p7 + p6 + p5 + p4 + p3 + p2 + p1 + p0 * 2 +
  1366. q0 + q1 + q2 + q3 + q4 + q5 + q6 + 8) >> 4;
  1367. dst[strideb * +0] = (p6 + p5 + p4 + p3 + p2 + p1 + p0 + q0 * 2 +
  1368. q1 + q2 + q3 + q4 + q5 + q6 + q7 + 8) >> 4;
  1369. dst[strideb * +1] = (p5 + p4 + p3 + p2 + p1 + p0 + q0 + q1 * 2 +
  1370. q2 + q3 + q4 + q5 + q6 + q7 + q7 + 8) >> 4;
  1371. dst[strideb * +2] = (p4 + p3 + p2 + p1 + p0 + q0 + q1 + q2 * 2 +
  1372. q3 + q4 + q5 + q6 + q7 + q7 + q7 + 8) >> 4;
  1373. dst[strideb * +3] = (p3 + p2 + p1 + p0 + q0 + q1 + q2 + q3 * 2 +
  1374. q4 + q5 + q6 + q7 + q7 + q7 + q7 + 8) >> 4;
  1375. dst[strideb * +4] = (p2 + p1 + p0 + q0 + q1 + q2 + q3 + q4 * 2 +
  1376. q5 + q6 + q7 + q7 + q7 + q7 + q7 + 8) >> 4;
  1377. dst[strideb * +5] = (p1 + p0 + q0 + q1 + q2 + q3 + q4 + q5 * 2 +
  1378. q6 + q7 + q7 + q7 + q7 + q7 + q7 + 8) >> 4;
  1379. dst[strideb * +6] = (p0 + q0 + q1 + q2 + q3 + q4 + q5 + q6 * 2 +
  1380. q7 + q7 + q7 + q7 + q7 + q7 + q7 + 8) >> 4;
  1381. } else if (wd >= 8 && flat8in) {
  1382. dst[strideb * -3] = (p3 + p3 + p3 + 2 * p2 + p1 + p0 + q0 + 4) >> 3;
  1383. dst[strideb * -2] = (p3 + p3 + p2 + 2 * p1 + p0 + q0 + q1 + 4) >> 3;
  1384. dst[strideb * -1] = (p3 + p2 + p1 + 2 * p0 + q0 + q1 + q2 + 4) >> 3;
  1385. dst[strideb * +0] = (p2 + p1 + p0 + 2 * q0 + q1 + q2 + q3 + 4) >> 3;
  1386. dst[strideb * +1] = (p1 + p0 + q0 + 2 * q1 + q2 + q3 + q3 + 4) >> 3;
  1387. dst[strideb * +2] = (p0 + q0 + q1 + 2 * q2 + q3 + q3 + q3 + 4) >> 3;
  1388. } else {
  1389. int hev = FFABS(p1 - p0) > H || FFABS(q1 - q0) > H;
  1390. if (hev) {
  1391. int f = av_clip_int8(3 * (q0 - p0) + av_clip_int8(p1 - q1)), f1, f2;
  1392. f1 = FFMIN(f + 4, 127) >> 3;
  1393. f2 = FFMIN(f + 3, 127) >> 3;
  1394. dst[strideb * -1] = av_clip_uint8(p0 + f2);
  1395. dst[strideb * +0] = av_clip_uint8(q0 - f1);
  1396. } else {
  1397. int f = av_clip_int8(3 * (q0 - p0)), f1, f2;
  1398. f1 = FFMIN(f + 4, 127) >> 3;
  1399. f2 = FFMIN(f + 3, 127) >> 3;
  1400. dst[strideb * -1] = av_clip_uint8(p0 + f2);
  1401. dst[strideb * +0] = av_clip_uint8(q0 - f1);
  1402. f = (f1 + 1) >> 1;
  1403. dst[strideb * -2] = av_clip_uint8(p1 + f);
  1404. dst[strideb * +1] = av_clip_uint8(q1 - f);
  1405. }
  1406. }
  1407. }
  1408. }
  1409. #define lf_8_fn(dir, wd, stridea, strideb) \
  1410. static void loop_filter_##dir##_##wd##_8_c(uint8_t *dst, \
  1411. ptrdiff_t stride, \
  1412. int E, int I, int H) \
  1413. { \
  1414. loop_filter(dst, E, I, H, stridea, strideb, wd); \
  1415. }
  1416. #define lf_8_fns(wd) \
  1417. lf_8_fn(h, wd, stride, 1) \
  1418. lf_8_fn(v, wd, 1, stride)
  1419. lf_8_fns(4)
  1420. lf_8_fns(8)
  1421. lf_8_fns(16)
  1422. #undef lf_8_fn
  1423. #undef lf_8_fns
  1424. #define lf_16_fn(dir, stridea) \
  1425. static void loop_filter_##dir##_16_16_c(uint8_t *dst, \
  1426. ptrdiff_t stride, \
  1427. int E, int I, int H) \
  1428. { \
  1429. loop_filter_##dir##_16_8_c(dst, stride, E, I, H); \
  1430. loop_filter_##dir##_16_8_c(dst + 8 * stridea, stride, E, I, H); \
  1431. }
  1432. lf_16_fn(h, stride)
  1433. lf_16_fn(v, 1)
  1434. #undef lf_16_fn
  1435. #define lf_mix_fn(dir, wd1, wd2, stridea) \
  1436. static void loop_filter_##dir##_##wd1##wd2##_16_c(uint8_t *dst, \
  1437. ptrdiff_t stride, \
  1438. int E, int I, int H) \
  1439. { \
  1440. loop_filter_##dir##_##wd1##_8_c(dst, stride, E & 0xff, I & 0xff, H & 0xff); \
  1441. loop_filter_##dir##_##wd2##_8_c(dst + 8 * stridea, stride, E >> 8, I >> 8, H >> 8); \
  1442. }
  1443. #define lf_mix_fns(wd1, wd2) \
  1444. lf_mix_fn(h, wd1, wd2, stride) \
  1445. lf_mix_fn(v, wd1, wd2, 1)
  1446. lf_mix_fns(4, 4)
  1447. lf_mix_fns(4, 8)
  1448. lf_mix_fns(8, 4)
  1449. lf_mix_fns(8, 8)
  1450. #undef lf_mix_fn
  1451. #undef lf_mix_fns
  1452. static av_cold void vp9dsp_loopfilter_init(VP9DSPContext *dsp)
  1453. {
  1454. dsp->loop_filter_8[0][0] = loop_filter_h_4_8_c;
  1455. dsp->loop_filter_8[0][1] = loop_filter_v_4_8_c;
  1456. dsp->loop_filter_8[1][0] = loop_filter_h_8_8_c;
  1457. dsp->loop_filter_8[1][1] = loop_filter_v_8_8_c;
  1458. dsp->loop_filter_8[2][0] = loop_filter_h_16_8_c;
  1459. dsp->loop_filter_8[2][1] = loop_filter_v_16_8_c;
  1460. dsp->loop_filter_16[0] = loop_filter_h_16_16_c;
  1461. dsp->loop_filter_16[1] = loop_filter_v_16_16_c;
  1462. dsp->loop_filter_mix2[0][0][0] = loop_filter_h_44_16_c;
  1463. dsp->loop_filter_mix2[0][0][1] = loop_filter_v_44_16_c;
  1464. dsp->loop_filter_mix2[0][1][0] = loop_filter_h_48_16_c;
  1465. dsp->loop_filter_mix2[0][1][1] = loop_filter_v_48_16_c;
  1466. dsp->loop_filter_mix2[1][0][0] = loop_filter_h_84_16_c;
  1467. dsp->loop_filter_mix2[1][0][1] = loop_filter_v_84_16_c;
  1468. dsp->loop_filter_mix2[1][1][0] = loop_filter_h_88_16_c;
  1469. dsp->loop_filter_mix2[1][1][1] = loop_filter_v_88_16_c;
  1470. }
  1471. static av_always_inline void copy_c(uint8_t *dst, ptrdiff_t dst_stride,
  1472. const uint8_t *src, ptrdiff_t src_stride,
  1473. int w, int h)
  1474. {
  1475. do {
  1476. memcpy(dst, src, w);
  1477. dst += dst_stride;
  1478. src += src_stride;
  1479. } while (--h);
  1480. }
  1481. static av_always_inline void avg_c(uint8_t *dst, ptrdiff_t dst_stride,
  1482. const uint8_t *src, ptrdiff_t src_stride,
  1483. int w, int h)
  1484. {
  1485. do {
  1486. int x;
  1487. for (x = 0; x < w; x += 4)
  1488. AV_WN32A(&dst[x], rnd_avg32(AV_RN32A(&dst[x]), AV_RN32(&src[x])));
  1489. dst += dst_stride;
  1490. src += src_stride;
  1491. } while (--h);
  1492. }
  1493. #define fpel_fn(type, sz) \
  1494. static void type##sz##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1495. const uint8_t *src, ptrdiff_t src_stride, \
  1496. int h, int mx, int my) \
  1497. { \
  1498. type##_c(dst, dst_stride, src, src_stride, sz, h); \
  1499. }
  1500. #define copy_avg_fn(sz) \
  1501. fpel_fn(copy, sz) \
  1502. fpel_fn(avg, sz)
  1503. copy_avg_fn(64)
  1504. copy_avg_fn(32)
  1505. copy_avg_fn(16)
  1506. copy_avg_fn(8)
  1507. copy_avg_fn(4)
  1508. #undef fpel_fn
  1509. #undef copy_avg_fn
  1510. static const int16_t vp9_subpel_filters[3][16][8] = {
  1511. [FILTER_8TAP_REGULAR] = {
  1512. { 0, 0, 0, 128, 0, 0, 0, 0 },
  1513. { 0, 1, -5, 126, 8, -3, 1, 0 },
  1514. { -1, 3, -10, 122, 18, -6, 2, 0 },
  1515. { -1, 4, -13, 118, 27, -9, 3, -1 },
  1516. { -1, 4, -16, 112, 37, -11, 4, -1 },
  1517. { -1, 5, -18, 105, 48, -14, 4, -1 },
  1518. { -1, 5, -19, 97, 58, -16, 5, -1 },
  1519. { -1, 6, -19, 88, 68, -18, 5, -1 },
  1520. { -1, 6, -19, 78, 78, -19, 6, -1 },
  1521. { -1, 5, -18, 68, 88, -19, 6, -1 },
  1522. { -1, 5, -16, 58, 97, -19, 5, -1 },
  1523. { -1, 4, -14, 48, 105, -18, 5, -1 },
  1524. { -1, 4, -11, 37, 112, -16, 4, -1 },
  1525. { -1, 3, -9, 27, 118, -13, 4, -1 },
  1526. { 0, 2, -6, 18, 122, -10, 3, -1 },
  1527. { 0, 1, -3, 8, 126, -5, 1, 0 },
  1528. }, [FILTER_8TAP_SHARP] = {
  1529. { 0, 0, 0, 128, 0, 0, 0, 0 },
  1530. { -1, 3, -7, 127, 8, -3, 1, 0 },
  1531. { -2, 5, -13, 125, 17, -6, 3, -1 },
  1532. { -3, 7, -17, 121, 27, -10, 5, -2 },
  1533. { -4, 9, -20, 115, 37, -13, 6, -2 },
  1534. { -4, 10, -23, 108, 48, -16, 8, -3 },
  1535. { -4, 10, -24, 100, 59, -19, 9, -3 },
  1536. { -4, 11, -24, 90, 70, -21, 10, -4 },
  1537. { -4, 11, -23, 80, 80, -23, 11, -4 },
  1538. { -4, 10, -21, 70, 90, -24, 11, -4 },
  1539. { -3, 9, -19, 59, 100, -24, 10, -4 },
  1540. { -3, 8, -16, 48, 108, -23, 10, -4 },
  1541. { -2, 6, -13, 37, 115, -20, 9, -4 },
  1542. { -2, 5, -10, 27, 121, -17, 7, -3 },
  1543. { -1, 3, -6, 17, 125, -13, 5, -2 },
  1544. { 0, 1, -3, 8, 127, -7, 3, -1 },
  1545. }, [FILTER_8TAP_SMOOTH] = {
  1546. { 0, 0, 0, 128, 0, 0, 0, 0 },
  1547. { -3, -1, 32, 64, 38, 1, -3, 0 },
  1548. { -2, -2, 29, 63, 41, 2, -3, 0 },
  1549. { -2, -2, 26, 63, 43, 4, -4, 0 },
  1550. { -2, -3, 24, 62, 46, 5, -4, 0 },
  1551. { -2, -3, 21, 60, 49, 7, -4, 0 },
  1552. { -1, -4, 18, 59, 51, 9, -4, 0 },
  1553. { -1, -4, 16, 57, 53, 12, -4, -1 },
  1554. { -1, -4, 14, 55, 55, 14, -4, -1 },
  1555. { -1, -4, 12, 53, 57, 16, -4, -1 },
  1556. { 0, -4, 9, 51, 59, 18, -4, -1 },
  1557. { 0, -4, 7, 49, 60, 21, -3, -2 },
  1558. { 0, -4, 5, 46, 62, 24, -3, -2 },
  1559. { 0, -4, 4, 43, 63, 26, -2, -2 },
  1560. { 0, -3, 2, 41, 63, 29, -2, -2 },
  1561. { 0, -3, 1, 38, 64, 32, -1, -3 },
  1562. }
  1563. };
  1564. #define FILTER_8TAP(src, x, F, stride) \
  1565. av_clip_uint8((F[0] * src[x + -3 * stride] + \
  1566. F[1] * src[x + -2 * stride] + \
  1567. F[2] * src[x + -1 * stride] + \
  1568. F[3] * src[x + +0 * stride] + \
  1569. F[4] * src[x + +1 * stride] + \
  1570. F[5] * src[x + +2 * stride] + \
  1571. F[6] * src[x + +3 * stride] + \
  1572. F[7] * src[x + +4 * stride] + 64) >> 7)
  1573. static av_always_inline void do_8tap_1d_c(uint8_t *dst, ptrdiff_t dst_stride,
  1574. const uint8_t *src, ptrdiff_t src_stride,
  1575. int w, int h, ptrdiff_t ds,
  1576. const int16_t *filter, int avg)
  1577. {
  1578. do {
  1579. int x;
  1580. for (x = 0; x < w; x++)
  1581. if (avg) {
  1582. dst[x] = (dst[x] + FILTER_8TAP(src, x, filter, ds) + 1) >> 1;
  1583. } else {
  1584. dst[x] = FILTER_8TAP(src, x, filter, ds);
  1585. }
  1586. dst += dst_stride;
  1587. src += src_stride;
  1588. } while (--h);
  1589. }
  1590. #define filter_8tap_1d_fn(opn, opa, dir, ds) \
  1591. static av_noinline void opn##_8tap_1d_##dir##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1592. const uint8_t *src, ptrdiff_t src_stride, \
  1593. int w, int h, const int16_t *filter) \
  1594. { \
  1595. do_8tap_1d_c(dst, dst_stride, src, src_stride, w, h, ds, filter, opa); \
  1596. }
  1597. filter_8tap_1d_fn(put, 0, v, src_stride)
  1598. filter_8tap_1d_fn(put, 0, h, 1)
  1599. filter_8tap_1d_fn(avg, 1, v, src_stride)
  1600. filter_8tap_1d_fn(avg, 1, h, 1)
  1601. #undef filter_8tap_1d_fn
  1602. static av_always_inline void do_8tap_2d_c(uint8_t *dst, ptrdiff_t dst_stride,
  1603. const uint8_t *src, ptrdiff_t src_stride,
  1604. int w, int h, const int16_t *filterx,
  1605. const int16_t *filtery, int avg)
  1606. {
  1607. int tmp_h = h + 7;
  1608. uint8_t tmp[64 * 71], *tmp_ptr = tmp;
  1609. src -= src_stride * 3;
  1610. do {
  1611. int x;
  1612. for (x = 0; x < w; x++)
  1613. tmp_ptr[x] = FILTER_8TAP(src, x, filterx, 1);
  1614. tmp_ptr += 64;
  1615. src += src_stride;
  1616. } while (--tmp_h);
  1617. tmp_ptr = tmp + 64 * 3;
  1618. do {
  1619. int x;
  1620. for (x = 0; x < w; x++)
  1621. if (avg) {
  1622. dst[x] = (dst[x] + FILTER_8TAP(tmp_ptr, x, filtery, 64) + 1) >> 1;
  1623. } else {
  1624. dst[x] = FILTER_8TAP(tmp_ptr, x, filtery, 64);
  1625. }
  1626. tmp_ptr += 64;
  1627. dst += dst_stride;
  1628. } while (--h);
  1629. }
  1630. #define filter_8tap_2d_fn(opn, opa) \
  1631. static av_noinline void opn##_8tap_2d_hv_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1632. const uint8_t *src, ptrdiff_t src_stride, \
  1633. int w, int h, const int16_t *filterx, \
  1634. const int16_t *filtery) \
  1635. { \
  1636. do_8tap_2d_c(dst, dst_stride, src, src_stride, w, h, filterx, filtery, opa); \
  1637. }
  1638. filter_8tap_2d_fn(put, 0)
  1639. filter_8tap_2d_fn(avg, 1)
  1640. #undef filter_8tap_2d_fn
  1641. #define filter_fn_1d(sz, dir, dir_m, type, type_idx, avg) \
  1642. static void avg##_8tap_##type##_##sz##dir##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1643. const uint8_t *src, ptrdiff_t src_stride, \
  1644. int h, int mx, int my) \
  1645. { \
  1646. avg##_8tap_1d_##dir##_c(dst, dst_stride, src, src_stride, sz, h, \
  1647. vp9_subpel_filters[type_idx][dir_m]); \
  1648. }
  1649. #define filter_fn_2d(sz, type, type_idx, avg) \
  1650. static void avg##_8tap_##type##_##sz##hv_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1651. const uint8_t *src, ptrdiff_t src_stride, \
  1652. int h, int mx, int my) \
  1653. { \
  1654. avg##_8tap_2d_hv_c(dst, dst_stride, src, src_stride, sz, h, \
  1655. vp9_subpel_filters[type_idx][mx], \
  1656. vp9_subpel_filters[type_idx][my]); \
  1657. }
  1658. #define FILTER_BILIN(src, x, mxy, stride) \
  1659. (src[x] + ((mxy * (src[x + stride] - src[x]) + 8) >> 4))
  1660. static av_always_inline void do_bilin_1d_c(uint8_t *dst, ptrdiff_t dst_stride,
  1661. const uint8_t *src, ptrdiff_t src_stride,
  1662. int w, int h, ptrdiff_t ds, int mxy, int avg)
  1663. {
  1664. do {
  1665. int x;
  1666. for (x = 0; x < w; x++)
  1667. if (avg) {
  1668. dst[x] = (dst[x] + FILTER_BILIN(src, x, mxy, ds) + 1) >> 1;
  1669. } else {
  1670. dst[x] = FILTER_BILIN(src, x, mxy, ds);
  1671. }
  1672. dst += dst_stride;
  1673. src += src_stride;
  1674. } while (--h);
  1675. }
  1676. #define bilin_1d_fn(opn, opa, dir, ds) \
  1677. static av_noinline void opn##_bilin_1d_##dir##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1678. const uint8_t *src, ptrdiff_t src_stride, \
  1679. int w, int h, int mxy) \
  1680. { \
  1681. do_bilin_1d_c(dst, dst_stride, src, src_stride, w, h, ds, mxy, opa); \
  1682. }
  1683. bilin_1d_fn(put, 0, v, src_stride)
  1684. bilin_1d_fn(put, 0, h, 1)
  1685. bilin_1d_fn(avg, 1, v, src_stride)
  1686. bilin_1d_fn(avg, 1, h, 1)
  1687. #undef bilin_1d_fn
  1688. static av_always_inline void do_bilin_2d_c(uint8_t *dst, ptrdiff_t dst_stride,
  1689. const uint8_t *src, ptrdiff_t src_stride,
  1690. int w, int h, int mx, int my, int avg)
  1691. {
  1692. uint8_t tmp[64 * 65], *tmp_ptr = tmp;
  1693. int tmp_h = h + 1;
  1694. do {
  1695. int x;
  1696. for (x = 0; x < w; x++)
  1697. tmp_ptr[x] = FILTER_BILIN(src, x, mx, 1);
  1698. tmp_ptr += 64;
  1699. src += src_stride;
  1700. } while (--tmp_h);
  1701. tmp_ptr = tmp;
  1702. do {
  1703. int x;
  1704. for (x = 0; x < w; x++)
  1705. if (avg) {
  1706. dst[x] = (dst[x] + FILTER_BILIN(tmp_ptr, x, my, 64) + 1) >> 1;
  1707. } else {
  1708. dst[x] = FILTER_BILIN(tmp_ptr, x, my, 64);
  1709. }
  1710. tmp_ptr += 64;
  1711. dst += dst_stride;
  1712. } while (--h);
  1713. }
  1714. #define bilin_2d_fn(opn, opa) \
  1715. static av_noinline void opn##_bilin_2d_hv_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1716. const uint8_t *src, ptrdiff_t src_stride, \
  1717. int w, int h, int mx, int my) \
  1718. { \
  1719. do_bilin_2d_c(dst, dst_stride, src, src_stride, w, h, mx, my, opa); \
  1720. }
  1721. bilin_2d_fn(put, 0)
  1722. bilin_2d_fn(avg, 1)
  1723. #undef bilin_2d_fn
  1724. #define bilinf_fn_1d(sz, dir, dir_m, avg) \
  1725. static void avg##_bilin_##sz##dir##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1726. const uint8_t *src, ptrdiff_t src_stride, \
  1727. int h, int mx, int my) \
  1728. { \
  1729. avg##_bilin_1d_##dir##_c(dst, dst_stride, src, src_stride, sz, h, dir_m); \
  1730. }
  1731. #define bilinf_fn_2d(sz, avg) \
  1732. static void avg##_bilin_##sz##hv_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1733. const uint8_t *src, ptrdiff_t src_stride, \
  1734. int h, int mx, int my) \
  1735. { \
  1736. avg##_bilin_2d_hv_c(dst, dst_stride, src, src_stride, sz, h, mx, my); \
  1737. }
  1738. #define filter_fn(sz, avg) \
  1739. filter_fn_1d(sz, h, mx, regular, FILTER_8TAP_REGULAR, avg) \
  1740. filter_fn_1d(sz, v, my, regular, FILTER_8TAP_REGULAR, avg) \
  1741. filter_fn_2d(sz, regular, FILTER_8TAP_REGULAR, avg) \
  1742. filter_fn_1d(sz, h, mx, smooth, FILTER_8TAP_SMOOTH, avg) \
  1743. filter_fn_1d(sz, v, my, smooth, FILTER_8TAP_SMOOTH, avg) \
  1744. filter_fn_2d(sz, smooth, FILTER_8TAP_SMOOTH, avg) \
  1745. filter_fn_1d(sz, h, mx, sharp, FILTER_8TAP_SHARP, avg) \
  1746. filter_fn_1d(sz, v, my, sharp, FILTER_8TAP_SHARP, avg) \
  1747. filter_fn_2d(sz, sharp, FILTER_8TAP_SHARP, avg) \
  1748. bilinf_fn_1d(sz, h, mx, avg) \
  1749. bilinf_fn_1d(sz, v, my, avg) \
  1750. bilinf_fn_2d(sz, avg)
  1751. #define filter_fn_set(avg) \
  1752. filter_fn(64, avg) \
  1753. filter_fn(32, avg) \
  1754. filter_fn(16, avg) \
  1755. filter_fn(8, avg) \
  1756. filter_fn(4, avg)
  1757. filter_fn_set(put)
  1758. filter_fn_set(avg)
  1759. #undef filter_fn
  1760. #undef filter_fn_set
  1761. #undef filter_fn_1d
  1762. #undef filter_fn_2d
  1763. #undef bilinf_fn_1d
  1764. #undef bilinf_fn_2d
  1765. static av_cold void vp9dsp_mc_init(VP9DSPContext *dsp)
  1766. {
  1767. #define init_fpel(idx1, idx2, sz, type) \
  1768. dsp->mc[idx1][FILTER_8TAP_SMOOTH ][idx2][0][0] = type##sz##_c; \
  1769. dsp->mc[idx1][FILTER_8TAP_REGULAR][idx2][0][0] = type##sz##_c; \
  1770. dsp->mc[idx1][FILTER_8TAP_SHARP ][idx2][0][0] = type##sz##_c; \
  1771. dsp->mc[idx1][FILTER_BILINEAR ][idx2][0][0] = type##sz##_c
  1772. #define init_copy_avg(idx, sz) \
  1773. init_fpel(idx, 0, sz, copy); \
  1774. init_fpel(idx, 1, sz, avg)
  1775. init_copy_avg(0, 64);
  1776. init_copy_avg(1, 32);
  1777. init_copy_avg(2, 16);
  1778. init_copy_avg(3, 8);
  1779. init_copy_avg(4, 4);
  1780. #undef init_copy_avg
  1781. #undef init_fpel
  1782. #define init_subpel1(idx1, idx2, idxh, idxv, sz, dir, type) \
  1783. dsp->mc[idx1][FILTER_8TAP_SMOOTH ][idx2][idxh][idxv] = type##_8tap_smooth_##sz##dir##_c; \
  1784. dsp->mc[idx1][FILTER_8TAP_REGULAR][idx2][idxh][idxv] = type##_8tap_regular_##sz##dir##_c; \
  1785. dsp->mc[idx1][FILTER_8TAP_SHARP ][idx2][idxh][idxv] = type##_8tap_sharp_##sz##dir##_c; \
  1786. dsp->mc[idx1][FILTER_BILINEAR ][idx2][idxh][idxv] = type##_bilin_##sz##dir##_c
  1787. #define init_subpel2(idx, idxh, idxv, dir, type) \
  1788. init_subpel1(0, idx, idxh, idxv, 64, dir, type); \
  1789. init_subpel1(1, idx, idxh, idxv, 32, dir, type); \
  1790. init_subpel1(2, idx, idxh, idxv, 16, dir, type); \
  1791. init_subpel1(3, idx, idxh, idxv, 8, dir, type); \
  1792. init_subpel1(4, idx, idxh, idxv, 4, dir, type)
  1793. #define init_subpel3(idx, type) \
  1794. init_subpel2(idx, 1, 1, hv, type); \
  1795. init_subpel2(idx, 0, 1, v, type); \
  1796. init_subpel2(idx, 1, 0, h, type)
  1797. init_subpel3(0, put);
  1798. init_subpel3(1, avg);
  1799. #undef init_subpel1
  1800. #undef init_subpel2
  1801. #undef init_subpel3
  1802. }
  1803. static av_always_inline void do_scaled_8tap_c(uint8_t *dst, ptrdiff_t dst_stride,
  1804. const uint8_t *src, ptrdiff_t src_stride,
  1805. int w, int h, int mx, int my,
  1806. int dx, int dy, int avg,
  1807. const int16_t (*filters)[8])
  1808. {
  1809. int tmp_h = (((h - 1) * dy + my) >> 4) + 8;
  1810. uint8_t tmp[64 * 135], *tmp_ptr = tmp;
  1811. src -= src_stride * 3;
  1812. do {
  1813. int x;
  1814. int imx = mx, ioff = 0;
  1815. for (x = 0; x < w; x++) {
  1816. tmp_ptr[x] = FILTER_8TAP(src, ioff, filters[imx], 1);
  1817. imx += dx;
  1818. ioff += imx >> 4;
  1819. imx &= 0xf;
  1820. }
  1821. tmp_ptr += 64;
  1822. src += src_stride;
  1823. } while (--tmp_h);
  1824. tmp_ptr = tmp + 64 * 3;
  1825. do {
  1826. int x;
  1827. const int16_t *filter = filters[my];
  1828. for (x = 0; x < w; x++)
  1829. if (avg) {
  1830. dst[x] = (dst[x] + FILTER_8TAP(tmp_ptr, x, filter, 64) + 1) >> 1;
  1831. } else {
  1832. dst[x] = FILTER_8TAP(tmp_ptr, x, filter, 64);
  1833. }
  1834. my += dy;
  1835. tmp_ptr += (my >> 4) * 64;
  1836. my &= 0xf;
  1837. dst += dst_stride;
  1838. } while (--h);
  1839. }
  1840. #define scaled_filter_8tap_fn(opn, opa) \
  1841. static av_noinline void opn##_scaled_8tap_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1842. const uint8_t *src, ptrdiff_t src_stride, \
  1843. int w, int h, int mx, int my, int dx, int dy, \
  1844. const int16_t (*filters)[8]) \
  1845. { \
  1846. do_scaled_8tap_c(dst, dst_stride, src, src_stride, w, h, mx, my, dx, dy, \
  1847. opa, filters); \
  1848. }
  1849. scaled_filter_8tap_fn(put, 0)
  1850. scaled_filter_8tap_fn(avg, 1)
  1851. #undef scaled_filter_8tap_fn
  1852. #undef FILTER_8TAP
  1853. #define scaled_filter_fn(sz, type, type_idx, avg) \
  1854. static void avg##_scaled_##type##_##sz##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1855. const uint8_t *src, ptrdiff_t src_stride, \
  1856. int h, int mx, int my, int dx, int dy) \
  1857. { \
  1858. avg##_scaled_8tap_c(dst, dst_stride, src, src_stride, sz, h, mx, my, dx, dy, \
  1859. vp9_subpel_filters[type_idx]); \
  1860. }
  1861. static av_always_inline void do_scaled_bilin_c(uint8_t *dst, ptrdiff_t dst_stride,
  1862. const uint8_t *src, ptrdiff_t src_stride,
  1863. int w, int h, int mx, int my,
  1864. int dx, int dy, int avg)
  1865. {
  1866. uint8_t tmp[64 * 129], *tmp_ptr = tmp;
  1867. int tmp_h = (((h - 1) * dy + my) >> 4) + 2;
  1868. do {
  1869. int x;
  1870. int imx = mx, ioff = 0;
  1871. for (x = 0; x < w; x++) {
  1872. tmp_ptr[x] = FILTER_BILIN(src, ioff, imx, 1);
  1873. imx += dx;
  1874. ioff += imx >> 4;
  1875. imx &= 0xf;
  1876. }
  1877. tmp_ptr += 64;
  1878. src += src_stride;
  1879. } while (--tmp_h);
  1880. tmp_ptr = tmp;
  1881. do {
  1882. int x;
  1883. for (x = 0; x < w; x++)
  1884. if (avg) {
  1885. dst[x] = (dst[x] + FILTER_BILIN(tmp_ptr, x, my, 64) + 1) >> 1;
  1886. } else {
  1887. dst[x] = FILTER_BILIN(tmp_ptr, x, my, 64);
  1888. }
  1889. my += dy;
  1890. tmp_ptr += (my >> 4) * 64;
  1891. my &= 0xf;
  1892. dst += dst_stride;
  1893. } while (--h);
  1894. }
  1895. #define scaled_bilin_fn(opn, opa) \
  1896. static av_noinline void opn##_scaled_bilin_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1897. const uint8_t *src, ptrdiff_t src_stride, \
  1898. int w, int h, int mx, int my, int dx, int dy) \
  1899. { \
  1900. do_scaled_bilin_c(dst, dst_stride, src, src_stride, w, h, mx, my, dx, dy, opa); \
  1901. }
  1902. scaled_bilin_fn(put, 0)
  1903. scaled_bilin_fn(avg, 1)
  1904. #undef scaled_bilin_fn
  1905. #undef FILTER_BILIN
  1906. #define scaled_bilinf_fn(sz, avg) \
  1907. static void avg##_scaled_bilin_##sz##_c(uint8_t *dst, ptrdiff_t dst_stride, \
  1908. const uint8_t *src, ptrdiff_t src_stride, \
  1909. int h, int mx, int my, int dx, int dy) \
  1910. { \
  1911. avg##_scaled_bilin_c(dst, dst_stride, src, src_stride, sz, h, mx, my, dx, dy); \
  1912. }
  1913. #define scaled_filter_fns(sz, avg) \
  1914. scaled_filter_fn(sz, regular, FILTER_8TAP_REGULAR, avg) \
  1915. scaled_filter_fn(sz, smooth, FILTER_8TAP_SMOOTH, avg) \
  1916. scaled_filter_fn(sz, sharp, FILTER_8TAP_SHARP, avg) \
  1917. scaled_bilinf_fn(sz, avg)
  1918. #define scaled_filter_fn_set(avg) \
  1919. scaled_filter_fns(64, avg) \
  1920. scaled_filter_fns(32, avg) \
  1921. scaled_filter_fns(16, avg) \
  1922. scaled_filter_fns(8, avg) \
  1923. scaled_filter_fns(4, avg)
  1924. scaled_filter_fn_set(put)
  1925. scaled_filter_fn_set(avg)
  1926. #undef scaled_filter_fns
  1927. #undef scaled_filter_fn_set
  1928. #undef scaled_filter_fn
  1929. #undef scaled_bilinf_fn
  1930. static av_cold void vp9dsp_scaled_mc_init(VP9DSPContext *dsp)
  1931. {
  1932. #define init_scaled(idx1, idx2, sz, type) \
  1933. dsp->smc[idx1][FILTER_8TAP_SMOOTH ][idx2] = type##_scaled_smooth_##sz##_c; \
  1934. dsp->smc[idx1][FILTER_8TAP_REGULAR][idx2] = type##_scaled_regular_##sz##_c; \
  1935. dsp->smc[idx1][FILTER_8TAP_SHARP ][idx2] = type##_scaled_sharp_##sz##_c; \
  1936. dsp->smc[idx1][FILTER_BILINEAR ][idx2] = type##_scaled_bilin_##sz##_c
  1937. #define init_scaled_put_avg(idx, sz) \
  1938. init_scaled(idx, 0, sz, put); \
  1939. init_scaled(idx, 1, sz, avg)
  1940. init_scaled_put_avg(0, 64);
  1941. init_scaled_put_avg(1, 32);
  1942. init_scaled_put_avg(2, 16);
  1943. init_scaled_put_avg(3, 8);
  1944. init_scaled_put_avg(4, 4);
  1945. #undef init_scaled_put_avg
  1946. #undef init_scaled
  1947. }
  1948. av_cold void ff_vp9dsp_init(VP9DSPContext *dsp)
  1949. {
  1950. vp9dsp_intrapred_init(dsp);
  1951. vp9dsp_itxfm_init(dsp);
  1952. vp9dsp_loopfilter_init(dsp);
  1953. vp9dsp_mc_init(dsp);
  1954. vp9dsp_scaled_mc_init(dsp);
  1955. if (ARCH_X86) ff_vp9dsp_init_x86(dsp);
  1956. }