You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

2329 lines
100KB

  1. /**
  2. * VP8 ARMv6 optimisations
  3. *
  4. * Copyright (c) 2011 The WebM project authors. All Rights Reserved.
  5. * Copyright (c) 2010 Rob Clark <rob@ti.com>
  6. * Copyright (c) 2011 Mans Rullgard <mans@mansr.com>
  7. *
  8. * This file is part of Libav.
  9. *
  10. * Libav is free software; you can redistribute it and/or
  11. * modify it under the terms of the GNU Lesser General Public
  12. * License as published by the Free Software Foundation; either
  13. * version 2.1 of the License, or (at your option) any later version.
  14. *
  15. * Libav is distributed in the hope that it will be useful,
  16. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  17. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  18. * Lesser General Public License for more details.
  19. *
  20. * You should have received a copy of the GNU Lesser General Public
  21. * License along with Libav; if not, write to the Free Software
  22. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  23. *
  24. * This code was partially ported from libvpx, which uses this license:
  25. *
  26. * Use of this source code is governed by a BSD-style license
  27. * that can be found in the LICENSE file in the root of the source
  28. * tree. An additional intellectual property rights grant can be found
  29. * in the file PATENTS. All contributing project authors may
  30. * be found in the AUTHORS file in the root of the source tree.
  31. *
  32. * (Note that the "LICENSE", "AUTHORS" and "PATENTS" files can be
  33. * found in the libvpx source tree.)
  34. */
  35. #include "asm.S"
  36. @ idct
  37. @ void vp8_luma_dc_wht(DCTELEM block[4][4][16], DCTELEM dc[16])
  38. function ff_vp8_luma_dc_wht_armv6, export=1
  39. push {r4 - r10, lr}
  40. @ load dc[] and zero memory
  41. mov r12, #0
  42. ldr r2, [r1] @ dc0[0,1]
  43. ldr r3, [r1, #4] @ dc0[2,3]
  44. ldr r4, [r1, #8] @ dc1[0,1]
  45. ldr r5, [r1, #12] @ dc1[2,3]
  46. ldr r6, [r1, #16] @ dc2[0,1]
  47. ldr r7, [r1, #20] @ dc2[2,3]
  48. ldr r8, [r1, #24] @ dc3[0,1]
  49. ldr r9, [r1, #28] @ dc3[2,3]
  50. str r12,[r1]
  51. str r12,[r1, #4]
  52. str r12,[r1, #8]
  53. str r12,[r1, #12]
  54. str r12,[r1, #16]
  55. str r12,[r1, #20]
  56. str r12,[r1, #24]
  57. str r12,[r1, #28]
  58. @ loop1
  59. uadd16 r12, r2, r8 @ t0[0,1]
  60. uadd16 r14, r3, r9 @ t0[2,3]
  61. usub16 r2, r2, r8 @ t3[0,1]
  62. usub16 r3, r3, r9 @ t3[2,3]
  63. uadd16 r8, r4, r6 @ t1[0,1]
  64. uadd16 r9, r5, r7 @ t1[2,3]
  65. usub16 r4, r4, r6 @ t2[0,1]
  66. usub16 r5, r5, r7 @ t2[2,3]
  67. uadd16 r6, r12, r8 @ dc0[0,1]
  68. uadd16 r7, r14, r9 @ dc0[2,3]
  69. usub16 r12, r12, r8 @ dc2[0,1]
  70. usub16 r14, r14, r9 @ dc2[2,3]
  71. uadd16 r8, r2, r4 @ dc1[0,1]
  72. uadd16 r9, r3, r5 @ dc1[2,3]
  73. usub16 r2, r2, r4 @ dc3[0,1]
  74. usub16 r3, r3, r5 @ dc3[2,3]
  75. mov r1, #3
  76. orr r1, r1, #0x30000 @ 3 | 3 (round)
  77. @ "transpose"
  78. pkhbt r4, r6, r8, lsl #16 @ dc{0,1}[0]
  79. pkhtb r6, r8, r6, asr #16 @ dc{0,1}[1]
  80. pkhbt r5, r12, r2, lsl #16 @ dc{2,3}[0]
  81. pkhtb r12, r2, r12, asr #16 @ dc{2,3}[1]
  82. pkhbt r8, r7, r9, lsl #16 @ dc{0,1}[2]
  83. uadd16 r4, r4, r1
  84. uadd16 r5, r5, r1
  85. pkhtb r7, r9, r7, asr #16 @ dc{0,1}[3]
  86. pkhbt r2, r14, r3, lsl #16 @ dc{2,3}[2]
  87. pkhtb r14, r3, r14, asr #16 @ dc{2,3}[3]
  88. @ loop2
  89. uadd16 r9, r4, r7 @ t0[0,1]
  90. uadd16 r3, r5, r14 @ t0[2,3]
  91. usub16 r4, r4, r7 @ t3[0,1]
  92. usub16 r5, r5, r14 @ t3[2,3]
  93. uadd16 r7, r6, r8 @ t1[0,1]
  94. uadd16 r14, r12, r2 @ t1[2,3]
  95. usub16 r6, r6, r8 @ t2[0,1]
  96. usub16 r12, r12, r2 @ t2[2,3]
  97. uadd16 r8, r9, r7 @ block[0,1][0]
  98. uadd16 r2, r3, r14 @ block[2,3][0]
  99. usub16 r9, r9, r7 @ block[0,1][2]
  100. usub16 r3, r3, r14 @ block[2,3][2]
  101. uadd16 r7, r4, r6 @ block[0,1][1]
  102. uadd16 r14, r5, r12 @ block[2,3][1]
  103. usub16 r4, r4, r6 @ block[0,1][3]
  104. usub16 r5, r5, r12 @ block[2,3][3]
  105. @ store
  106. mov r6, r8, asr #19 @ block[1][0]
  107. mov r12, r7, asr #19 @ block[1][1]
  108. mov r1, r9, asr #19 @ block[1][2]
  109. mov r10, r4, asr #19 @ block[1][3]
  110. sxth r8, r8
  111. sxth r7, r7
  112. sxth r9, r9
  113. sxth r4, r4
  114. asr r8, #3 @ block[0][0]
  115. asr r7, #3 @ block[0][1]
  116. asr r9, #3 @ block[0][2]
  117. asr r4, #3 @ block[0][3]
  118. strh r8, [r0], #32
  119. strh r7, [r0], #32
  120. strh r9, [r0], #32
  121. strh r4, [r0], #32
  122. strh r6, [r0], #32
  123. strh r12,[r0], #32
  124. strh r1, [r0], #32
  125. strh r10,[r0], #32
  126. mov r6, r2, asr #19 @ block[3][0]
  127. mov r12, r14, asr #19 @ block[3][1]
  128. mov r1, r3, asr #19 @ block[3][2]
  129. mov r10, r5, asr #19 @ block[3][3]
  130. sxth r2, r2
  131. sxth r14, r14
  132. sxth r3, r3
  133. sxth r5, r5
  134. asr r2, #3 @ block[2][0]
  135. asr r14, #3 @ block[2][1]
  136. asr r3, #3 @ block[2][2]
  137. asr r5, #3 @ block[2][3]
  138. strh r2, [r0], #32
  139. strh r14,[r0], #32
  140. strh r3, [r0], #32
  141. strh r5, [r0], #32
  142. strh r6, [r0], #32
  143. strh r12,[r0], #32
  144. strh r1, [r0], #32
  145. strh r10,[r0], #32
  146. pop {r4 - r10, pc}
  147. endfunc
  148. @ void vp8_luma_dc_wht_dc(DCTELEM block[4][4][16], DCTELEM dc[16])
  149. function ff_vp8_luma_dc_wht_dc_armv6, export=1
  150. ldrsh r2, [r1]
  151. mov r3, #0
  152. add r2, r2, #3
  153. strh r3, [r1]
  154. asr r2, r2, #3
  155. .rept 16
  156. strh r2, [r0], #32
  157. .endr
  158. bx lr
  159. endfunc
  160. @ void vp8_idct_add(uint8_t *dst, DCTELEM block[16], int stride)
  161. function ff_vp8_idct_add_armv6, export=1
  162. push {r4 - r11, lr}
  163. sub sp, sp, #32
  164. mov r3, #0x00004E00 @ cos
  165. orr r3, r3, #0x0000007B @ cospi8sqrt2minus1 = 20091
  166. mov r4, #0x00008A00 @ sin
  167. orr r4, r4, #0x0000008C @ sinpi8sqrt2 = 35468
  168. mov r5, #0x2 @ i=2
  169. 1:
  170. ldr r6, [r1, #8] @ i5 | i4 = block1[1] | block1[0]
  171. ldr r12,[r1, #24] @ i13 | i12 = block3[1] | block3[0]
  172. ldr r14,[r1, #16] @ i9 | i8 = block2[1] | block2[0]
  173. smulwt r9, r3, r6 @ (ip[5] * cospi8sqrt2minus1) >> 16
  174. smulwb r7, r3, r6 @ (ip[4] * cospi8sqrt2minus1) >> 16
  175. smulwt r10, r4, r6 @ (ip[5] * sinpi8sqrt2) >> 16
  176. smulwb r8, r4, r6 @ (ip[4] * sinpi8sqrt2) >> 16
  177. pkhbt r7, r7, r9, lsl #16 @ 5c | 4c
  178. smulwt r11, r3, r12 @ (ip[13] * cospi8sqrt2minus1) >> 16
  179. pkhbt r8, r8, r10, lsl #16 @ 5s | 4s = t2 first half
  180. uadd16 r6, r6, r7 @ 5c+5 | 4c+4 = t3 first half
  181. smulwt r7, r4, r12 @ (ip[13] * sinpi8sqrt2) >> 16
  182. smulwb r9, r3, r12 @ (ip[12] * cospi8sqrt2minus1) >> 16
  183. smulwb r10, r4, r12 @ (ip[12] * sinpi8sqrt2) >> 16
  184. subs r5, r5, #1 @ i--
  185. pkhbt r9, r9, r11, lsl #16 @ 13c | 12c
  186. ldr r11,[r1] @ i1 | i0
  187. pkhbt r10, r10, r7, lsl #16 @ 13s | 12s = t3 second half
  188. uadd16 r7, r12, r9 @ 13c+13 | 12c+12 = t2 second half
  189. usub16 r7, r8, r7 @ c = t2
  190. uadd16 r6, r6, r10 @ d = t3
  191. uadd16 r10, r11, r14 @ a = t0
  192. usub16 r8, r11, r14 @ b = t1
  193. uadd16 r9, r10, r6 @ a+d = tmp{0,1}[0]
  194. usub16 r10, r10, r6 @ a-d = tmp{0,1}[3]
  195. uadd16 r6, r8, r7 @ b+c = tmp{0,1}[1]
  196. usub16 r7, r8, r7 @ b-c = tmp{0,1}[2]
  197. mov r8, #0
  198. str r6, [sp, #8] @ o5 | o4
  199. str r7, [sp, #16] @ o9 | o8
  200. str r10,[sp, #24] @ o13 | o12
  201. str r9, [sp], #4 @ o1 | o0
  202. str r8, [r1, #24]
  203. str r8, [r1, #16]
  204. str r8, [r1, #8]
  205. str r8, [r1], #4
  206. bne 1b
  207. mov r5, #0x2 @ i=2
  208. sub sp, sp, #8
  209. 2:
  210. ldr r6, [sp, #8] @ i5 | i4 = tmp{0,1}[1]
  211. ldr r14,[sp, #4] @ i3 | i2 = tmp{2,3}[0]
  212. ldr r12,[sp, #12] @ i7 | i6 = tmp{2,3}[1]
  213. ldr r1, [sp], #16 @ i1 | i0 = tmp{0,1}[0]
  214. smulwt r9, r3, r6 @ (ip[5] * cospi8sqrt2minus1) >> 16
  215. smulwt r7, r3, r1 @ (ip[1] * cospi8sqrt2minus1) >> 16
  216. smulwt r10, r4, r6 @ (ip[5] * sinpi8sqrt2) >> 16
  217. smulwt r8, r4, r1 @ (ip[1] * sinpi8sqrt2) >> 16
  218. pkhbt r11, r1, r6, lsl #16 @ i4 | i0 = t0/t1 first half
  219. pkhbt r7, r7, r9, lsl #16 @ 5c | 1c
  220. pkhbt r8, r8, r10, lsl #16 @ 5s | 1s = temp1 = t2 first half
  221. pkhtb r1, r6, r1, asr #16 @ i5 | i1
  222. uadd16 r1, r7, r1 @ 5c+5 | 1c+1 = temp2 (d) = t3 first half
  223. pkhbt r9, r14, r12, lsl #16 @ i6 | i2 = t0/t1 second half
  224. uadd16 r10, r11, r9 @ a = t0
  225. usub16 r9, r11, r9 @ b = t1
  226. pkhtb r6, r12, r14, asr #16 @ i7 | i3
  227. subs r5, r5, #0x1 @ i--
  228. smulwt r7, r3, r6 @ (ip[7] * cospi8sqrt2minus1) >> 16
  229. smulwt r11, r4, r6 @ (ip[7] * sinpi8sqrt2) >> 16
  230. smulwb r12, r3, r6 @ (ip[3] * cospi8sqrt2minus1) >> 16
  231. smulwb r14, r4, r6 @ (ip[3] * sinpi8sqrt2) >> 16
  232. pkhbt r7, r12, r7, lsl #16 @ 7c | 3c
  233. pkhbt r11, r14, r11, lsl #16 @ 7s | 3s = temp1 (d) = t3 second half
  234. mov r14, #0x4 @ set up 4's
  235. orr r14, r14, #0x40000 @ 4|4
  236. uadd16 r6, r7, r6 @ 7c+7 | 3c+3 = temp2 (c) = t2 second half
  237. usub16 r12, r8, r6 @ c (o5 | o1) = t2
  238. uadd16 r6, r11, r1 @ d (o7 | o3) = t3
  239. uadd16 r10, r10, r14 @ t0 + 4
  240. uadd16 r9, r9, r14 @ t1 + 4
  241. uadd16 r7, r10, r6 @ a+d = dst{0,1}[0]
  242. usub16 r6, r10, r6 @ a-d = dst{0,1}[3]
  243. uadd16 r10, r9, r12 @ b+c = dst{0,1}[1]
  244. usub16 r1, r9, r12 @ b-c = dst{0,1}[2]
  245. mov r9, r6, asr #3 @ o[1][3]
  246. mov r12, r1, asr #3 @ o[1][2]
  247. pkhtb r8, r12, r7, asr #19 @ o[1][0,2]
  248. pkhtb r11, r9, r10, asr #19 @ o[1][1,3]
  249. ldr r12,[r0]
  250. ldr r9, [r0, r2]
  251. sxth r7, r7
  252. sxth r6, r6
  253. sxth r10, r10
  254. sxth r1, r1
  255. asr r7, #3 @ o[0][0]
  256. asr r10, #3 @ o[0][1]
  257. pkhbt r7, r7, r1, lsl #13 @ o[0][0,2]
  258. pkhbt r10, r10, r6, lsl #13 @ o[0][1,3]
  259. uxtab16 r7, r7, r12
  260. uxtab16 r10, r10, r12, ror #8
  261. uxtab16 r8, r8, r9
  262. uxtab16 r11, r11, r9, ror #8
  263. usat16 r7, #8, r7
  264. usat16 r10, #8, r10
  265. usat16 r8, #8, r8
  266. usat16 r11, #8, r11
  267. orr r7, r7, r10, lsl #8
  268. orr r8, r8, r11, lsl #8
  269. str r8, [r0, r2]
  270. str_post r7, r0, r2, lsl #1
  271. bne 2b
  272. pop {r4 - r11, pc}
  273. endfunc
  274. @ void vp8_idct_dc_add(uint8_t *dst, DCTELEM block[16], int stride)
  275. function ff_vp8_idct_dc_add_armv6, export=1
  276. push {r4 - r5, lr}
  277. ldrsh r3, [r1]
  278. mov r4, #0
  279. add r3, r3, #4
  280. asr r3, #3
  281. strh r4, [r1], #32
  282. ldr r4, [r0, r2]
  283. ldr_post r5, r0, r2, lsl #1
  284. pkhbt r3, r3, r3, lsl #16
  285. uxtab16 lr, r3, r5 @ a1+2 | a1+0
  286. uxtab16 r5, r3, r5, ror #8 @ a1+3 | a1+1
  287. uxtab16 r12, r3, r4
  288. uxtab16 r4, r3, r4, ror #8
  289. usat16 lr, #8, lr
  290. usat16 r5, #8, r5
  291. usat16 r12, #8, r12
  292. usat16 r4, #8, r4
  293. orr lr, lr, r5, lsl #8
  294. orr r12, r12, r4, lsl #8
  295. ldr r5, [r0]
  296. ldr r4, [r0, r2]
  297. sub r0, r0, r2, lsl #1
  298. str r12,[r0, r2]
  299. str_post lr, r0, r2, lsl #1
  300. uxtab16 lr, r3, r5
  301. uxtab16 r5, r3, r5, ror #8
  302. uxtab16 r12, r3, r4
  303. uxtab16 r4, r3, r4, ror #8
  304. usat16 lr, #8, lr
  305. usat16 r5, #8, r5
  306. usat16 r12, #8, r12
  307. usat16 r4, #8, r4
  308. orr lr, lr, r5, lsl #8
  309. orr r12, r12, r4, lsl #8
  310. str r12,[r0, r2]
  311. str_post lr, r0, r2, lsl #1
  312. pop {r4 - r5, pc}
  313. endfunc
  314. @ void vp8_idct_dc_add4uv(uint8_t *dst, DCTELEM block[4][16], int stride)
  315. function ff_vp8_idct_dc_add4uv_armv6, export=1
  316. push {lr}
  317. bl ff_vp8_idct_dc_add_armv6
  318. sub r0, r0, r2, lsl #2
  319. add r0, r0, #4
  320. bl ff_vp8_idct_dc_add_armv6
  321. sub r0, r0, #4
  322. bl ff_vp8_idct_dc_add_armv6
  323. sub r0, r0, r2, lsl #2
  324. add r0, r0, #4
  325. bl ff_vp8_idct_dc_add_armv6
  326. pop {pc}
  327. endfunc
  328. @ void vp8_idct_dc_add4y(uint8_t *dst, DCTELEM block[4][16], int stride)
  329. function ff_vp8_idct_dc_add4y_armv6, export=1
  330. push {lr}
  331. bl ff_vp8_idct_dc_add_armv6
  332. sub r0, r0, r2, lsl #2
  333. add r0, r0, #4
  334. bl ff_vp8_idct_dc_add_armv6
  335. sub r0, r0, r2, lsl #2
  336. add r0, r0, #4
  337. bl ff_vp8_idct_dc_add_armv6
  338. sub r0, r0, r2, lsl #2
  339. add r0, r0, #4
  340. bl ff_vp8_idct_dc_add_armv6
  341. pop {pc}
  342. endfunc
  343. @ loopfilter
  344. @ void vp8_v_loop_filter16_simple(uint8_t *dst, int stride, int flim)
  345. function ff_vp8_v_loop_filter16_simple_armv6, export=1
  346. push {r4 - r11, lr}
  347. ldr_dpren r3, r0, r1, lsl #1 @ p1
  348. ldr_dpren r4, r0, r1 @ p0
  349. ldr r5, [r0] @ q0
  350. ldr r6, [r0, r1] @ q1
  351. orr r2, r2, r2, lsl #16
  352. mov r9, #4 @ count
  353. mov lr, #0 @ need 0 in a couple places
  354. orr r12, r2, r2, lsl #8 @ splat int -> byte
  355. ldr r2, c0x80808080
  356. 1:
  357. @ vp8_simple_filter_mask()
  358. uqsub8 r7, r3, r6 @ p1 - q1
  359. uqsub8 r8, r6, r3 @ q1 - p1
  360. uqsub8 r10, r4, r5 @ p0 - q0
  361. uqsub8 r11, r5, r4 @ q0 - p0
  362. orr r8, r8, r7 @ abs(p1 - q1)
  363. orr r10, r10, r11 @ abs(p0 - q0)
  364. uqadd8 r10, r10, r10 @ abs(p0 - q0) * 2
  365. uhadd8 r8, r8, lr @ abs(p1 - q2) >> 1
  366. uqadd8 r10, r10, r8 @ abs(p0 - q0)*2 + abs(p1 - q1)/2
  367. mvn r8, #0
  368. usub8 r10, r12, r10 @ compare to flimit. usub8 sets GE flags
  369. sel r10, r8, lr @ filter mask: F or 0
  370. cmp r10, #0
  371. beq 2f @ skip filtering if all masks are 0x00
  372. @ vp8_simple_filter()
  373. eor r3, r3, r2 @ p1 offset to convert to a signed value
  374. eor r6, r6, r2 @ q1 offset to convert to a signed value
  375. eor r4, r4, r2 @ p0 offset to convert to a signed value
  376. eor r5, r5, r2 @ q0 offset to convert to a signed value
  377. qsub8 r3, r3, r6 @ vp8_filter = p1 - q1
  378. qsub8 r6, r5, r4 @ q0 - p0
  379. qadd8 r3, r3, r6 @ += q0 - p0
  380. ldr r7, c0x04040404
  381. qadd8 r3, r3, r6 @ += q0 - p0
  382. ldr r8, c0x03030303
  383. qadd8 r3, r3, r6 @ vp8_filter = p1-q1 + 3*(q0-p0))
  384. @STALL
  385. and r3, r3, r10 @ vp8_filter &= mask
  386. qadd8 r7, r3, r7 @ Filter1 = vp8_filter + 4
  387. qadd8 r8, r3, r8 @ Filter2 = vp8_filter + 3
  388. shadd8 r7, r7, lr
  389. shadd8 r8, r8, lr
  390. shadd8 r7, r7, lr
  391. shadd8 r8, r8, lr
  392. shadd8 r7, r7, lr @ Filter1 >>= 3
  393. shadd8 r8, r8, lr @ Filter2 >>= 3
  394. qsub8 r5, r5, r7 @ u = q0 - Filter1
  395. qadd8 r4, r4, r8 @ u = p0 + Filter2
  396. eor r5, r5, r2 @ *oq0 = u^0x80
  397. eor r4, r4, r2 @ *op0 = u^0x80
  398. T sub r7, r0, r1
  399. str r5, [r0] @ store oq0 result
  400. A str r4, [r0, -r1] @ store op0 result
  401. T str r4, [r7]
  402. 2:
  403. subs r9, r9, #1 @ counter--
  404. add r0, r0, #4 @ next row
  405. T itttt ne
  406. A ldrne r3, [r0, -r1, lsl #1] @ p1
  407. T subne r3, r0, r1, lsl #1
  408. T ldrne r3, [r3] @ p1
  409. A ldrne r4, [r0, -r1] @ p0
  410. T subne r4, r0, r1
  411. T ldrne r4, [r4] @ p0
  412. T itt ne
  413. ldrne r5, [r0] @ q0
  414. ldrne r6, [r0, r1] @ q1
  415. bne 1b
  416. pop {r4 - r11, pc}
  417. endfunc
  418. c0x01010101: .long 0x01010101
  419. c0x03030303: .long 0x03030303
  420. c0x04040404: .long 0x04040404
  421. c0x7F7F7F7F: .long 0x7F7F7F7F
  422. c0x80808080: .long 0x80808080
  423. @ void vp8_v_loop_filter16_inner(uint8_t *dst, int stride,
  424. @ int fE, int fI, int hev_thresh)
  425. @ and
  426. @ void vp8_v_loop_filter8uv_inner(uint8_t *dstU, uint8_t *dstV, int stride,
  427. @ int fE, int fI, int hev_thresh)
  428. @ call:
  429. @ void vp8_v_loop_filter_inner(uint8_t *dst, int stride,
  430. @ int fE, int fI, int hev_thresh, int count)
  431. function ff_vp8_v_loop_filter_inner_armv6, export=1
  432. push {r4 - r11, lr}
  433. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  434. ldr r5, [sp, #40] @ counter
  435. ldr r6, [sp, #36] @ load thresh address
  436. sub sp, sp, #16 @ create temp buffer
  437. ldr r10,[r0, r1] @ p2
  438. ldr_post r9, r0, r1, lsl #1 @ p3
  439. ldr r12,[r0, r1] @ p0
  440. ldr_post r11, r0, r1, lsl #1 @ p1
  441. orr r2, r2, r2, lsl #16
  442. orr r3, r3, r3, lsl #16
  443. orr r6, r6, r6, lsl #16
  444. orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
  445. orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
  446. orr r3, r6, r6, lsl #8 @ thresh splat int -> byte
  447. 1:
  448. @ vp8_filter_mask() function
  449. @ calculate breakout conditions
  450. uqsub8 r6, r9, r10 @ p3 - p2
  451. uqsub8 r7, r10, r9 @ p2 - p3
  452. uqsub8 r8, r10, r11 @ p2 - p1
  453. uqsub8 r10, r11, r10 @ p1 - p2
  454. orr r6, r6, r7 @ abs (p3-p2)
  455. orr r8, r8, r10 @ abs (p2-p1)
  456. uqsub8 lr, r6, r2 @ compare to limit. lr: vp8_filter_mask
  457. uqsub8 r8, r8, r2 @ compare to limit
  458. uqsub8 r6, r11, r12 @ p1 - p0
  459. orr lr, lr, r8
  460. uqsub8 r7, r12, r11 @ p0 - p1
  461. ldr r10,[r0, r1] @ q1
  462. ldr_post r9, r0, r1, lsl #1 @ q0
  463. orr r6, r6, r7 @ abs (p1-p0)
  464. uqsub8 r7, r6, r2 @ compare to limit
  465. uqsub8 r8, r6, r3 @ compare to thresh -- save r8 for later
  466. orr lr, lr, r7
  467. uqsub8 r6, r11, r10 @ p1 - q1
  468. uqsub8 r7, r10, r11 @ q1 - p1
  469. uqsub8 r11, r12, r9 @ p0 - q0
  470. uqsub8 r12, r9, r12 @ q0 - p0
  471. orr r6, r6, r7 @ abs (p1-q1)
  472. ldr r7, c0x7F7F7F7F
  473. orr r12, r11, r12 @ abs (p0-q0)
  474. ldr_post r11, r0, r1 @ q2
  475. uqadd8 r12, r12, r12 @ abs (p0-q0) * 2
  476. and r6, r7, r6, lsr #1 @ abs (p1-q1) / 2
  477. uqsub8 r7, r9, r10 @ q0 - q1
  478. uqadd8 r12, r12, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
  479. uqsub8 r6, r10, r9 @ q1 - q0
  480. uqsub8 r12, r12, r4 @ compare to flimit
  481. uqsub8 r9, r11, r10 @ q2 - q1
  482. orr lr, lr, r12
  483. ldr_post r12, r0, r1 @ q3
  484. uqsub8 r10, r10, r11 @ q1 - q2
  485. orr r6, r7, r6 @ abs (q1-q0)
  486. orr r10, r9, r10 @ abs (q2-q1)
  487. uqsub8 r7, r6, r2 @ compare to limit
  488. uqsub8 r10, r10, r2 @ compare to limit
  489. uqsub8 r6, r6, r3 @ compare to thresh -- save r6 for later
  490. orr lr, lr, r7
  491. orr lr, lr, r10
  492. uqsub8 r10, r12, r11 @ q3 - q2
  493. uqsub8 r9, r11, r12 @ q2 - q3
  494. mvn r11, #0 @ r11 == -1
  495. orr r10, r10, r9 @ abs (q3-q2)
  496. uqsub8 r10, r10, r2 @ compare to limit
  497. mov r12, #0
  498. orr lr, lr, r10
  499. sub r0, r0, r1, lsl #2
  500. usub8 lr, r12, lr @ use usub8 instead of ssub8
  501. sel lr, r11, r12 @ filter mask: lr
  502. cmp lr, #0
  503. beq 2f @ skip filtering
  504. sub r0, r0, r1, lsl #1 @ move r0 pointer down by 6 lines
  505. @vp8_hevmask() function
  506. @calculate high edge variance
  507. orr r10, r6, r8 @ calculate vp8_hevmask
  508. usub8 r10, r12, r10 @ use usub8 instead of ssub8
  509. sel r6, r12, r11 @ obtain vp8_hevmask: r6
  510. @vp8_filter() function
  511. ldr r8, [r0, r1] @ p0
  512. ldr_post r7, r0, r1, lsl #1 @ p1
  513. ldr r12, c0x80808080
  514. ldr r10,[r0, r1] @ q1
  515. ldr_post r9, r0, r1, lsl #1 @ q0
  516. eor r7, r7, r12 @ p1 offset to convert to a signed value
  517. eor r8, r8, r12 @ p0 offset to convert to a signed value
  518. eor r9, r9, r12 @ q0 offset to convert to a signed value
  519. eor r10, r10, r12 @ q1 offset to convert to a signed value
  520. str r9, [sp] @ store qs0 temporarily
  521. str r8, [sp, #4] @ store ps0 temporarily
  522. str r10,[sp, #8] @ store qs1 temporarily
  523. str r7, [sp, #12] @ store ps1 temporarily
  524. qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
  525. qsub8 r8, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
  526. and r7, r7, r6 @ vp8_filter (r7) &= hev
  527. qadd8 r7, r7, r8
  528. ldr r9, c0x03030303 @ r9 = 3 --modified for vp8
  529. qadd8 r7, r7, r8
  530. ldr r10, c0x04040404
  531. qadd8 r7, r7, r8
  532. and r7, r7, lr @ vp8_filter &= mask@
  533. qadd8 r8, r7, r9 @ Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
  534. qadd8 r7, r7, r10 @ vp8_filter = vp8_signed_char_clamp(vp8_filter+4)
  535. mov r9, #0
  536. shadd8 r8, r8, r9 @ Filter2 >>= 3
  537. shadd8 r7, r7, r9 @ vp8_filter >>= 3
  538. shadd8 r8, r8, r9
  539. shadd8 r7, r7, r9
  540. shadd8 lr, r8, r9 @ lr: Filter2
  541. shadd8 r7, r7, r9 @ r7: filter
  542. @calculate output
  543. ldr r8, [sp] @ load qs0
  544. ldr r9, [sp, #4] @ load ps0
  545. ldr r10, c0x01010101
  546. qsub8 r8, r8, r7 @ u = vp8_signed_char_clamp(qs0 - vp8_filter)
  547. qadd8 r9, r9, lr @ u = vp8_signed_char_clamp(ps0 + Filter2)
  548. mov lr, #0
  549. sadd8 r7, r7, r10 @ vp8_filter += 1
  550. shadd8 r7, r7, lr @ vp8_filter >>= 1
  551. ldr r11,[sp, #12] @ load ps1
  552. ldr r10,[sp, #8] @ load qs1
  553. bic r7, r7, r6 @ vp8_filter &= ~hev
  554. sub r0, r0, r1, lsl #2
  555. qadd8 r11, r11, r7 @ u = vp8_signed_char_clamp(ps1 + vp8_filter)
  556. qsub8 r10, r10, r7 @ u = vp8_signed_char_clamp(qs1 - vp8_filter)
  557. eor r11, r11, r12 @ *op1 = u^0x80
  558. eor r9, r9, r12 @ *op0 = u^0x80
  559. eor r8, r8, r12 @ *oq0 = u^0x80
  560. eor r10, r10, r12 @ *oq1 = u^0x80
  561. str r9, [r0, r1] @ store op0 result
  562. str_post r11, r0, r1, lsl #1 @ store op1
  563. str r10,[r0, r1] @ store oq1
  564. str_post r8, r0, r1, lsl #1 @ store oq0 result
  565. sub r0, r0, r1, lsl #1
  566. 2:
  567. add r0, r0, #4
  568. sub r0, r0, r1, lsl #2
  569. subs r5, r5, #1
  570. T ittt ne
  571. ldrne r10,[r0, r1] @ p2
  572. A ldrne r9, [r0], r1, lsl #1 @ p3
  573. T ldrne r9, [r0] @ p3
  574. T addne r0, r0, r1, lsl #1
  575. T ittt ne
  576. ldrne r12,[r0, r1] @ p0
  577. A ldrne r11,[r0], r1, lsl #1 @ p1
  578. T ldrne r11,[r0] @ p3
  579. T addne r0, r0, r1, lsl #1
  580. bne 1b
  581. add sp, sp, #16
  582. pop {r4 - r11, pc}
  583. endfunc
  584. @ void vp8_v_loop_filter16(uint8_t *dst, int stride,
  585. @ int fE, int fI, int hev_thresh)
  586. @ and
  587. @ void vp8_v_loop_filter8uv(uint8_t *dstU, uint8_t *dstV, int stride,
  588. @ int fE, int fI, int hev_thresh)
  589. @ call:
  590. @ void vp8_v_loop_filter(uint8_t *dst, int stride,
  591. @ int fE, int fI, int hev_thresh, int count)
  592. function ff_vp8_v_loop_filter_armv6, export=1
  593. push {r4 - r11, lr}
  594. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  595. ldr r5, [sp, #40] @ counter
  596. ldr r6, [sp, #36] @ load thresh address
  597. sub sp, sp, #16 @ create temp buffer
  598. ldr r10,[r0, r1] @ p2
  599. ldr_post r9, r0, r1, lsl #1 @ p3
  600. ldr r12,[r0, r1] @ p0
  601. ldr_post r11, r0, r1, lsl #1 @ p1
  602. orr r2, r2, r2, lsl #16
  603. orr r3, r3, r3, lsl #16
  604. orr r6, r6, r6, lsl #16
  605. orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
  606. orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
  607. orr r3, r6, r6, lsl #8 @ thresh splat int -> byte
  608. 1:
  609. @ vp8_filter_mask() function
  610. @ calculate breakout conditions
  611. uqsub8 r6, r9, r10 @ p3 - p2
  612. uqsub8 r7, r10, r9 @ p2 - p3
  613. uqsub8 r8, r10, r11 @ p2 - p1
  614. uqsub8 r10, r11, r10 @ p1 - p2
  615. orr r6, r6, r7 @ abs (p3-p2)
  616. orr r8, r8, r10 @ abs (p2-p1)
  617. uqsub8 lr, r6, r2 @ compare to limit. lr: vp8_filter_mask
  618. uqsub8 r8, r8, r2 @ compare to limit
  619. uqsub8 r6, r11, r12 @ p1 - p0
  620. orr lr, lr, r8
  621. uqsub8 r7, r12, r11 @ p0 - p1
  622. ldr r10,[r0, r1] @ q1
  623. ldr_post r9, r0, r1, lsl #1 @ q0
  624. orr r6, r6, r7 @ abs (p1-p0)
  625. uqsub8 r7, r6, r2 @ compare to limit
  626. uqsub8 r8, r6, r3 @ compare to thresh -- save r8 for later
  627. orr lr, lr, r7
  628. uqsub8 r6, r11, r10 @ p1 - q1
  629. uqsub8 r7, r10, r11 @ q1 - p1
  630. uqsub8 r11, r12, r9 @ p0 - q0
  631. uqsub8 r12, r9, r12 @ q0 - p0
  632. orr r6, r6, r7 @ abs (p1-q1)
  633. ldr r7, c0x7F7F7F7F
  634. orr r12, r11, r12 @ abs (p0-q0)
  635. ldr_post r11, r0, r1 @ q2
  636. uqadd8 r12, r12, r12 @ abs (p0-q0) * 2
  637. and r6, r7, r6, lsr #1 @ abs (p1-q1) / 2
  638. uqsub8 r7, r9, r10 @ q0 - q1
  639. uqadd8 r12, r12, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
  640. uqsub8 r6, r10, r9 @ q1 - q0
  641. uqsub8 r12, r12, r4 @ compare to flimit
  642. uqsub8 r9, r11, r10 @ q2 - q1
  643. orr lr, lr, r12
  644. ldr_post r12, r0, r1 @ q3
  645. uqsub8 r10, r10, r11 @ q1 - q2
  646. orr r6, r7, r6 @ abs (q1-q0)
  647. orr r10, r9, r10 @ abs (q2-q1)
  648. uqsub8 r7, r6, r2 @ compare to limit
  649. uqsub8 r10, r10, r2 @ compare to limit
  650. uqsub8 r6, r6, r3 @ compare to thresh -- save r6 for later
  651. orr lr, lr, r7
  652. orr lr, lr, r10
  653. uqsub8 r10, r12, r11 @ q3 - q2
  654. uqsub8 r9, r11, r12 @ q2 - q3
  655. mvn r11, #0 @ r11 == -1
  656. orr r10, r10, r9 @ abs (q3-q2)
  657. uqsub8 r10, r10, r2 @ compare to limit
  658. mov r12, #0
  659. orr lr, lr, r10
  660. usub8 lr, r12, lr @ use usub8 instead of ssub8
  661. sel lr, r11, r12 @ filter mask: lr
  662. cmp lr, #0
  663. beq 2f @ skip filtering
  664. @vp8_hevmask() function
  665. @calculate high edge variance
  666. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 6 lines
  667. sub r0, r0, r1, lsl #1
  668. orr r10, r6, r8
  669. usub8 r10, r12, r10
  670. sel r6, r12, r11 @ hev mask: r6
  671. @vp8_mbfilter() function
  672. @p2, q2 are only needed at the end. Do not need to load them in now.
  673. ldr r8, [r0, r1] @ p0
  674. ldr_post r7, r0, r1, lsl #1 @ p1
  675. ldr r12, c0x80808080
  676. ldr_post r9, r0, r1 @ q0
  677. ldr r10,[r0] @ q1
  678. eor r7, r7, r12 @ ps1
  679. eor r8, r8, r12 @ ps0
  680. eor r9, r9, r12 @ qs0
  681. eor r10, r10, r12 @ qs1
  682. qsub8 r12, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
  683. str r7, [sp, #12] @ store ps1 temporarily
  684. qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
  685. str r10,[sp, #8] @ store qs1 temporarily
  686. qadd8 r7, r7, r12
  687. str r9, [sp] @ store qs0 temporarily
  688. qadd8 r7, r7, r12
  689. str r8, [sp, #4] @ store ps0 temporarily
  690. qadd8 r7, r7, r12 @ vp8_filter: r7
  691. ldr r10, c0x03030303 @ r10 = 3 --modified for vp8
  692. ldr r9, c0x04040404
  693. and r7, r7, lr @ vp8_filter &= mask (lr is free)
  694. mov r12, r7 @ Filter2: r12
  695. and r12, r12, r6 @ Filter2 &= hev
  696. @save bottom 3 bits so that we round one side +4 and the other +3
  697. qadd8 r8, r12, r9 @ Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
  698. qadd8 r12, r12, r10 @ Filter2 (r12) = vp8_signed_char_clamp(Filter2+3)
  699. mov r10, #0
  700. shadd8 r8, r8, r10 @ Filter1 >>= 3
  701. shadd8 r12, r12, r10 @ Filter2 >>= 3
  702. shadd8 r8, r8, r10
  703. shadd8 r12, r12, r10
  704. shadd8 r8, r8, r10 @ r8: Filter1
  705. shadd8 r12, r12, r10 @ r12: Filter2
  706. ldr r9, [sp] @ load qs0
  707. ldr r11,[sp, #4] @ load ps0
  708. qsub8 r9, r9, r8 @ qs0 = vp8_signed_char_clamp(qs0 - Filter1)
  709. qadd8 r11, r11, r12 @ ps0 = vp8_signed_char_clamp(ps0 + Filter2)
  710. bic r12, r7, r6 @ vp8_filter &= ~hev ( r6 is free)
  711. @roughly 3/7th difference across boundary
  712. mov lr, #0x1b @ 27
  713. mov r7, #0x3f @ 63
  714. sxtb16 r6, r12
  715. sxtb16 r10, r12, ror #8
  716. smlabb r8, r6, lr, r7
  717. smlatb r6, r6, lr, r7
  718. smlabb r7, r10, lr, r7
  719. smultb r10, r10, lr
  720. ssat r8, #8, r8, asr #7
  721. ssat r6, #8, r6, asr #7
  722. add r10, r10, #63
  723. ssat r7, #8, r7, asr #7
  724. ssat r10, #8, r10, asr #7
  725. ldr lr, c0x80808080
  726. pkhbt r6, r8, r6, lsl #16
  727. pkhbt r10, r7, r10, lsl #16
  728. uxtb16 r6, r6
  729. uxtb16 r10, r10
  730. sub r0, r0, r1
  731. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7)
  732. qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs0 - u)
  733. qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps0 + u)
  734. eor r8, r8, lr @ *oq0 = s^0x80
  735. str r8, [r0] @ store *oq0
  736. sub r0, r0, r1
  737. eor r10, r10, lr @ *op0 = s^0x80
  738. str r10,[r0] @ store *op0
  739. @roughly 2/7th difference across boundary
  740. mov lr, #0x12 @ 18
  741. mov r7, #0x3f @ 63
  742. sxtb16 r6, r12
  743. sxtb16 r10, r12, ror #8
  744. smlabb r8, r6, lr, r7
  745. smlatb r6, r6, lr, r7
  746. smlabb r9, r10, lr, r7
  747. smlatb r10, r10, lr, r7
  748. ssat r8, #8, r8, asr #7
  749. ssat r6, #8, r6, asr #7
  750. ssat r9, #8, r9, asr #7
  751. ssat r10, #8, r10, asr #7
  752. ldr lr, c0x80808080
  753. pkhbt r6, r8, r6, lsl #16
  754. pkhbt r10, r9, r10, lsl #16
  755. ldr r9, [sp, #8] @ load qs1
  756. ldr r11, [sp, #12] @ load ps1
  757. uxtb16 r6, r6
  758. uxtb16 r10, r10
  759. sub r0, r0, r1
  760. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7)
  761. qadd8 r11, r11, r10 @ s = vp8_signed_char_clamp(ps1 + u)
  762. qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs1 - u)
  763. eor r11, r11, lr @ *op1 = s^0x80
  764. str_post r11, r0, r1 @ store *op1
  765. eor r8, r8, lr @ *oq1 = s^0x80
  766. add r0, r0, r1, lsl #1
  767. mov r7, #0x3f @ 63
  768. str_post r8, r0, r1 @ store *oq1
  769. @roughly 1/7th difference across boundary
  770. mov lr, #0x9 @ 9
  771. ldr r9, [r0] @ load q2
  772. sxtb16 r6, r12
  773. sxtb16 r10, r12, ror #8
  774. smlabb r8, r6, lr, r7
  775. smlatb r6, r6, lr, r7
  776. smlabb r12, r10, lr, r7
  777. smlatb r10, r10, lr, r7
  778. ssat r8, #8, r8, asr #7
  779. ssat r6, #8, r6, asr #7
  780. ssat r12, #8, r12, asr #7
  781. ssat r10, #8, r10, asr #7
  782. sub r0, r0, r1, lsl #2
  783. pkhbt r6, r8, r6, lsl #16
  784. pkhbt r10, r12, r10, lsl #16
  785. sub r0, r0, r1
  786. ldr lr, c0x80808080
  787. ldr r11, [r0] @ load p2
  788. uxtb16 r6, r6
  789. uxtb16 r10, r10
  790. eor r9, r9, lr
  791. eor r11, r11, lr
  792. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7)
  793. qadd8 r8, r11, r10 @ s = vp8_signed_char_clamp(ps2 + u)
  794. qsub8 r10, r9, r10 @ s = vp8_signed_char_clamp(qs2 - u)
  795. eor r8, r8, lr @ *op2 = s^0x80
  796. str_post r8, r0, r1, lsl #2 @ store *op2
  797. add r0, r0, r1
  798. eor r10, r10, lr @ *oq2 = s^0x80
  799. str_post r10, r0, r1, lsl #1 @ store *oq2
  800. 2:
  801. add r0, r0, #4
  802. sub r0, r0, r1, lsl #3
  803. subs r5, r5, #1
  804. T ittt ne
  805. ldrne r10,[r0, r1] @ p2
  806. A ldrne r9, [r0], r1, lsl #1 @ p3
  807. T ldrne r9, [r0] @ p3
  808. T addne r0, r0, r1, lsl #1
  809. T ittt ne
  810. ldrne r12,[r0, r1] @ p0
  811. A ldrne r11,[r0], r1, lsl #1 @ p1
  812. T ldrne r11,[r0] @ p3
  813. T addne r0, r0, r1, lsl #1
  814. bne 1b
  815. add sp, sp, #16
  816. pop {r4 - r11, pc}
  817. endfunc
  818. .macro TRANSPOSE_MATRIX i0, i1, i2, i3, o3, o2, o1, o0
  819. @ input: $0, $1, $2, $3
  820. @ output: $4, $5, $6, $7
  821. @ i0: 03 02 01 00
  822. @ i1: 13 12 11 10
  823. @ i2: 23 22 21 20
  824. @ i3: 33 32 31 30
  825. @ o3 o2 o1 o0
  826. uxtb16 \o1, \i1 @ xx 12 xx 10
  827. uxtb16 \o0, \i0 @ xx 02 xx 00
  828. uxtb16 \o3, \i3 @ xx 32 xx 30
  829. uxtb16 \o2, \i2 @ xx 22 xx 20
  830. orr \o1, \o0, \o1, lsl #8 @ 12 02 10 00
  831. orr \o3, \o2, \o3, lsl #8 @ 32 22 30 20
  832. uxtb16 \i1, \i1, ror #8 @ xx 13 xx 11
  833. uxtb16 \i3, \i3, ror #8 @ xx 33 xx 31
  834. uxtb16 \i0, \i0, ror #8 @ xx 03 xx 01
  835. uxtb16 \i2, \i2, ror #8 @ xx 23 xx 21
  836. orr \i0, \i0, \i1, lsl #8 @ 13 03 11 01
  837. orr \i2, \i2, \i3, lsl #8 @ 33 23 31 21
  838. pkhtb \o2, \o3, \o1, asr #16 @ 32 22 12 02 -- p1
  839. pkhbt \o0, \o1, \o3, lsl #16 @ 30 20 10 00 -- p3
  840. pkhtb \o3, \i2, \i0, asr #16 @ 33 23 13 03 -- p0
  841. pkhbt \o1, \i0, \i2, lsl #16 @ 31 21 11 01 -- p2
  842. .endm
  843. @ void vp8_h_loop_filter16_simple(uint8_t *dst, int stride, int flim)
  844. function ff_vp8_h_loop_filter16_simple_armv6, export=1
  845. push {r4 - r11, lr}
  846. orr r12, r2, r2, lsl #16
  847. ldr r2, c0x80808080
  848. orr r12, r12, r12, lsl #8
  849. @ load soure data to r7, r8, r9, r10
  850. sub r0, r0, #2
  851. ldr r8, [r0, r1]
  852. ldr_post r7, r0, r1, lsl #1
  853. ldr r10,[r0, r1]
  854. ldr_post r9, r0, r1, lsl #1
  855. add r0, r0, #2
  856. mov r11, #4 @ count (r11) for 4-in-parallel
  857. 1:
  858. @transpose r7, r8, r9, r10 to r3, r4, r5, r6
  859. TRANSPOSE_MATRIX r7, r8, r9, r10, r6, r5, r4, r3
  860. @ vp8_simple_filter_mask() function
  861. uqsub8 r7, r3, r6 @ p1 - q1
  862. uqsub8 r8, r6, r3 @ q1 - p1
  863. uqsub8 r9, r4, r5 @ p0 - q0
  864. uqsub8 r10, r5, r4 @ q0 - p0
  865. orr r7, r7, r8 @ abs(p1 - q1)
  866. orr r9, r9, r10 @ abs(p0 - q0)
  867. mov r8, #0
  868. uqadd8 r9, r9, r9 @ abs(p0 - q0) * 2
  869. uhadd8 r7, r7, r8 @ abs(p1 - q1) / 2
  870. uqadd8 r7, r7, r9 @ abs(p0 - q0)*2 + abs(p1 - q1)/2
  871. mvn r10, #0 @ r10 == -1
  872. usub8 r7, r12, r7 @ compare to flimit
  873. sel lr, r10, r8 @ filter mask
  874. cmp lr, #0
  875. beq 2f @ skip filtering
  876. @vp8_simple_filter() function
  877. eor r3, r3, r2 @ p1 offset to convert to a signed value
  878. eor r6, r6, r2 @ q1 offset to convert to a signed value
  879. eor r4, r4, r2 @ p0 offset to convert to a signed value
  880. eor r5, r5, r2 @ q0 offset to convert to a signed value
  881. qsub8 r3, r3, r6 @ vp8_filter = p1 - q1
  882. qsub8 r6, r5, r4 @ q0 - p0
  883. qadd8 r3, r3, r6 @ vp8_filter += q0 - p0
  884. ldr r9, c0x03030303 @ r9 = 3
  885. qadd8 r3, r3, r6 @ vp8_filter += q0 - p0
  886. ldr r7, c0x04040404
  887. qadd8 r3, r3, r6 @ vp8_filter = p1-q1 + 3*(q0-p0))
  888. @STALL
  889. and r3, r3, lr @ vp8_filter &= mask
  890. qadd8 r9, r3, r9 @ Filter2 = vp8_filter + 3
  891. qadd8 r3, r3, r7 @ Filter1 = vp8_filter + 4
  892. shadd8 r9, r9, r8
  893. shadd8 r3, r3, r8
  894. shadd8 r9, r9, r8
  895. shadd8 r3, r3, r8
  896. shadd8 r9, r9, r8 @ Filter2 >>= 3
  897. shadd8 r3, r3, r8 @ Filter1 >>= 3
  898. @calculate output
  899. sub r0, r0, r1, lsl #2
  900. qadd8 r4, r4, r9 @ u = p0 + Filter2
  901. qsub8 r5, r5, r3 @ u = q0 - Filter1
  902. eor r4, r4, r2 @ *op0 = u^0x80
  903. eor r5, r5, r2 @ *oq0 = u^0x80
  904. strb r4, [r0, #-1] @ store the result
  905. mov r4, r4, lsr #8
  906. strb_post r5, r0, r1
  907. mov r5, r5, lsr #8
  908. strb r4, [r0, #-1]
  909. mov r4, r4, lsr #8
  910. strb_post r5, r0, r1
  911. mov r5, r5, lsr #8
  912. strb r4, [r0, #-1]
  913. mov r4, r4, lsr #8
  914. strb_post r5, r0, r1
  915. mov r5, r5, lsr #8
  916. strb r4, [r0, #-1]
  917. strb_post r5, r0, r1
  918. 2:
  919. subs r11, r11, #1
  920. @ load soure data to r7, r8, r9, r10
  921. sub r0, r0, #2
  922. T ittt ne
  923. ldrne r8, [r0, r1]
  924. A ldrne r7, [r0], r1, lsl #1
  925. T ldrne r7, [r0]
  926. T addne r0, r0, r1, lsl #1
  927. T ittt ne
  928. ldrne r10,[r0, r1]
  929. A ldrne r9, [r0], r1, lsl #1
  930. T ldrne r9, [r0]
  931. T addne r0, r0, r1, lsl #1
  932. add r0, r0, #2
  933. bne 1b
  934. pop {r4 - r11, pc}
  935. endfunc
  936. @ void vp8_h_loop_filter16_inner(uint8_t *dst, int stride,
  937. @ int fE, int fI, int hev_thresh)
  938. @ and
  939. @ void vp8_h_loop_filter8uv_inner(uint8_t *dstU, uint8_t *dstV, int stride,
  940. @ int fE, int fI, int hev_thresh)
  941. @ call:
  942. @ void vp8_h_loop_filter_inner(uint8_t *dst, int stride,
  943. @ int fE, int fI, int hev_thresh, int count)
  944. function ff_vp8_h_loop_filter_inner_armv6, export=1
  945. push {r4 - r11, lr}
  946. sub r0, r0, #4 @ move r0 pointer down by 4
  947. ldr r5, [sp, #40] @ counter
  948. ldr r9, [sp, #36] @ load thresh address
  949. sub sp, sp, #16 @ create temp buffer
  950. ldr r7, [r0, r1] @ transpose will make it into p3-p0
  951. ldr_post r6, r0, r1, lsl #1 @ load source data
  952. ldr lr, [r0, r1]
  953. ldr_post r8, r0, r1, lsl #1
  954. orr r2, r2, r2, lsl #16
  955. orr r3, r3, r3, lsl #16
  956. orr r9, r9, r9, lsl #16
  957. orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
  958. orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
  959. orr r3, r9, r9, lsl #8 @ thresh splat int -> byte
  960. 1:
  961. @ vp8_filter_mask() function
  962. @ calculate breakout conditions
  963. @ transpose the source data for 4-in-parallel operation
  964. TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
  965. uqsub8 r7, r9, r10 @ p3 - p2
  966. uqsub8 r8, r10, r9 @ p2 - p3
  967. uqsub8 r9, r10, r11 @ p2 - p1
  968. uqsub8 r10, r11, r10 @ p1 - p2
  969. orr r7, r7, r8 @ abs (p3-p2)
  970. orr r10, r9, r10 @ abs (p2-p1)
  971. uqsub8 lr, r7, r2 @ compare to limit. lr: vp8_filter_mask
  972. uqsub8 r10, r10, r2 @ compare to limit
  973. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  974. orr lr, lr, r10
  975. uqsub8 r6, r11, r12 @ p1 - p0
  976. uqsub8 r7, r12, r11 @ p0 - p1
  977. add r0, r0, #4 @ move r0 pointer up by 4
  978. orr r6, r6, r7 @ abs (p1-p0)
  979. str r11,[sp, #12] @ save p1
  980. uqsub8 r10, r6, r2 @ compare to limit
  981. uqsub8 r11, r6, r3 @ compare to thresh
  982. orr lr, lr, r10
  983. @ transpose uses 8 regs(r6 - r12 and lr). Need to save reg value now
  984. @ transpose the source data for 4-in-parallel operation
  985. str r11,[sp] @ push r11 to stack
  986. ldr r7, [r0, r1]
  987. ldr_post r6, r0, r1, lsl #1 @ load source data
  988. str r12,[sp, #4] @ save current reg before load q0 - q3 data
  989. str lr, [sp, #8]
  990. ldr lr, [r0, r1]
  991. ldr_post r8, r0, r1, lsl #1
  992. TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
  993. ldr lr, [sp, #8] @ load back (f)limit accumulator
  994. uqsub8 r6, r12, r11 @ q3 - q2
  995. uqsub8 r7, r11, r12 @ q2 - q3
  996. uqsub8 r12, r11, r10 @ q2 - q1
  997. uqsub8 r11, r10, r11 @ q1 - q2
  998. orr r6, r6, r7 @ abs (q3-q2)
  999. orr r7, r12, r11 @ abs (q2-q1)
  1000. uqsub8 r6, r6, r2 @ compare to limit
  1001. uqsub8 r7, r7, r2 @ compare to limit
  1002. ldr r11,[sp, #4] @ load back p0
  1003. ldr r12,[sp, #12] @ load back p1
  1004. orr lr, lr, r6
  1005. orr lr, lr, r7
  1006. uqsub8 r6, r11, r9 @ p0 - q0
  1007. uqsub8 r7, r9, r11 @ q0 - p0
  1008. uqsub8 r8, r12, r10 @ p1 - q1
  1009. uqsub8 r11, r10, r12 @ q1 - p1
  1010. orr r6, r6, r7 @ abs (p0-q0)
  1011. ldr r7, c0x7F7F7F7F
  1012. orr r8, r8, r11 @ abs (p1-q1)
  1013. uqadd8 r6, r6, r6 @ abs (p0-q0) * 2
  1014. and r8, r7, r8, lsr #1 @ abs (p1-q1) / 2
  1015. uqsub8 r11, r10, r9 @ q1 - q0
  1016. uqadd8 r6, r8, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
  1017. uqsub8 r12, r9, r10 @ q0 - q1
  1018. uqsub8 r6, r6, r4 @ compare to flimit
  1019. orr r9, r11, r12 @ abs (q1-q0)
  1020. uqsub8 r8, r9, r2 @ compare to limit
  1021. uqsub8 r10, r9, r3 @ compare to thresh
  1022. orr lr, lr, r6
  1023. orr lr, lr, r8
  1024. mvn r11, #0 @ r11 == -1
  1025. mov r12, #0
  1026. usub8 lr, r12, lr
  1027. ldr r9, [sp] @ load the compared result
  1028. sel lr, r11, r12 @ filter mask: lr
  1029. cmp lr, #0
  1030. beq 2f @ skip filtering
  1031. @vp8_hevmask() function
  1032. @calculate high edge variance
  1033. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  1034. orr r9, r9, r10
  1035. ldrh r7, [r0, #-2]
  1036. ldrh_post r8, r0, r1
  1037. usub8 r9, r12, r9
  1038. sel r6, r12, r11 @ hev mask: r6
  1039. @vp8_filter() function
  1040. @ load soure data to r6, r11, r12, lr
  1041. ldrh r9, [r0, #-2]
  1042. ldrh_post r10, r0, r1
  1043. pkhbt r12, r7, r8, lsl #16
  1044. ldrh r7, [r0, #-2]
  1045. ldrh_post r8, r0, r1
  1046. pkhbt r11, r9, r10, lsl #16
  1047. ldrh r9, [r0, #-2]
  1048. ldrh_post r10, r0, r1
  1049. @ Transpose needs 8 regs(r6 - r12, and lr). Save r6 and lr first
  1050. str r6, [sp]
  1051. str lr, [sp, #4]
  1052. pkhbt r6, r7, r8, lsl #16
  1053. pkhbt lr, r9, r10, lsl #16
  1054. @transpose r12, r11, r6, lr to r7, r8, r9, r10
  1055. TRANSPOSE_MATRIX r12, r11, r6, lr, r10, r9, r8, r7
  1056. @load back hev_mask r6 and filter_mask lr
  1057. ldr r12, c0x80808080
  1058. ldr r6, [sp]
  1059. ldr lr, [sp, #4]
  1060. eor r7, r7, r12 @ p1 offset to convert to a signed value
  1061. eor r8, r8, r12 @ p0 offset to convert to a signed value
  1062. eor r9, r9, r12 @ q0 offset to convert to a signed value
  1063. eor r10, r10, r12 @ q1 offset to convert to a signed value
  1064. str r9, [sp] @ store qs0 temporarily
  1065. str r8, [sp, #4] @ store ps0 temporarily
  1066. str r10,[sp, #8] @ store qs1 temporarily
  1067. str r7, [sp, #12] @ store ps1 temporarily
  1068. qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
  1069. qsub8 r8, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
  1070. and r7, r7, r6 @ vp8_filter (r7) &= hev (r7 : filter)
  1071. qadd8 r7, r7, r8
  1072. ldr r9, c0x03030303 @ r9 = 3 --modified for vp8
  1073. qadd8 r7, r7, r8
  1074. ldr r10, c0x04040404
  1075. qadd8 r7, r7, r8
  1076. and r7, r7, lr @ vp8_filter &= mask
  1077. qadd8 r8, r7, r9 @ Filter2 (r8) = vp8_signed_char_clamp(vp8_filter+3)
  1078. qadd8 r7, r7, r10 @ vp8_filter = vp8_signed_char_clamp(vp8_filter+4)
  1079. mov r9, #0
  1080. shadd8 r8, r8, r9 @ Filter2 >>= 3
  1081. shadd8 r7, r7, r9 @ vp8_filter >>= 3
  1082. shadd8 r8, r8, r9
  1083. shadd8 r7, r7, r9
  1084. shadd8 lr, r8, r9 @ lr: filter2
  1085. shadd8 r7, r7, r9 @ r7: filter
  1086. @calculate output
  1087. ldr r8, [sp] @ load qs0
  1088. ldr r9, [sp, #4] @ load ps0
  1089. ldr r10, c0x01010101
  1090. qsub8 r8, r8, r7 @ u = vp8_signed_char_clamp(qs0 - vp8_filter)
  1091. qadd8 r9, r9, lr @ u = vp8_signed_char_clamp(ps0 + Filter2)
  1092. eor r8, r8, r12
  1093. eor r9, r9, r12
  1094. mov lr, #0
  1095. sadd8 r7, r7, r10
  1096. shadd8 r7, r7, lr
  1097. ldr r10,[sp, #8] @ load qs1
  1098. ldr r11,[sp, #12] @ load ps1
  1099. bic r7, r7, r6 @ r7: vp8_filter
  1100. qsub8 r10, r10, r7 @ u = vp8_signed_char_clamp(qs1 - vp8_filter)
  1101. qadd8 r11, r11, r7 @ u = vp8_signed_char_clamp(ps1 + vp8_filter)
  1102. eor r10, r10, r12
  1103. eor r11, r11, r12
  1104. sub r0, r0, r1, lsl #2
  1105. @we can use TRANSPOSE_MATRIX macro to transpose output - input: q1, q0, p0, p1
  1106. TRANSPOSE_MATRIX r11, r9, r8, r10, lr, r12, r7, r6
  1107. strh r6, [r0, #-2] @ store the result
  1108. mov r6, r6, lsr #16
  1109. strh_post r6, r0, r1
  1110. strh r7, [r0, #-2]
  1111. mov r7, r7, lsr #16
  1112. strh_post r7, r0, r1
  1113. strh r12, [r0, #-2]
  1114. mov r12, r12, lsr #16
  1115. strh_post r12, r0, r1
  1116. strh lr, [r0, #-2]
  1117. mov lr, lr, lsr #16
  1118. strh_post lr, r0, r1
  1119. 2:
  1120. sub r0, r0, #4
  1121. subs r5, r5, #1
  1122. T ittt ne
  1123. ldrne r7, [r0, r1]
  1124. A ldrne r6, [r0], r1, lsl #1 @ load source data
  1125. T ldrne r6, [r0] @ load source data
  1126. T addne r0, r0, r1, lsl #1
  1127. T ittt ne
  1128. ldrne lr, [r0, r1]
  1129. A ldrne r8, [r0], r1, lsl #1
  1130. T ldrne r8, [r0]
  1131. T addne r0, r0, r1, lsl #1
  1132. bne 1b
  1133. add sp, sp, #16
  1134. pop {r4 - r11, pc}
  1135. endfunc
  1136. @ void vp8_h_loop_filter16(uint8_t *dst, int stride,
  1137. @ int fE, int fI, int hev_thresh)
  1138. @ and
  1139. @ void vp8_h_loop_filter8uv(uint8_t *dstU, uint8_t *dstV, int stride,
  1140. @ int fE, int fI, int hev_thresh)
  1141. @ call:
  1142. @ void vp8_h_loop_filter(uint8_t *dst, int stride,
  1143. @ int fE, int fI, int hev_thresh, int count)
  1144. function ff_vp8_h_loop_filter_armv6, export=1
  1145. push {r4 - r11, lr}
  1146. sub r0, r0, #4 @ move r0 pointer down by 4
  1147. ldr r5, [sp, #40] @ counter
  1148. ldr r9, [sp, #36] @ load thresh address
  1149. sub sp, sp, #16 @ create temp buffer
  1150. ldr r7, [r0, r1] @ transpose will make it into p3-p0
  1151. ldr_post r6, r0, r1, lsl #1 @ load source data
  1152. ldr lr, [r0, r1]
  1153. ldr_post r8, r0, r1, lsl #1
  1154. orr r2, r2, r2, lsl #16
  1155. orr r3, r3, r3, lsl #16
  1156. orr r9, r9, r9, lsl #16
  1157. orr r4, r2, r2, lsl #8 @ flimE splat int -> byte
  1158. orr r2, r3, r3, lsl #8 @ flimI splat int -> byte
  1159. orr r3, r9, r9, lsl #8 @ thresh splat int -> byte
  1160. 1:
  1161. @ vp8_filter_mask() function
  1162. @ calculate breakout conditions
  1163. @ transpose the source data for 4-in-parallel operation
  1164. TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
  1165. uqsub8 r7, r9, r10 @ p3 - p2
  1166. uqsub8 r8, r10, r9 @ p2 - p3
  1167. uqsub8 r9, r10, r11 @ p2 - p1
  1168. uqsub8 r10, r11, r10 @ p1 - p2
  1169. orr r7, r7, r8 @ abs (p3-p2)
  1170. orr r10, r9, r10 @ abs (p2-p1)
  1171. uqsub8 lr, r7, r2 @ compare to limit. lr: vp8_filter_mask
  1172. uqsub8 r10, r10, r2 @ compare to limit
  1173. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  1174. orr lr, lr, r10
  1175. uqsub8 r6, r11, r12 @ p1 - p0
  1176. uqsub8 r7, r12, r11 @ p0 - p1
  1177. add r0, r0, #4 @ move r0 pointer up by 4
  1178. orr r6, r6, r7 @ abs (p1-p0)
  1179. str r11,[sp, #12] @ save p1
  1180. uqsub8 r10, r6, r2 @ compare to limit
  1181. uqsub8 r11, r6, r3 @ compare to thresh
  1182. orr lr, lr, r10
  1183. @ transpose uses 8 regs(r6 - r12 and lr). Need to save reg value now
  1184. @ transpose the source data for 4-in-parallel operation
  1185. str r11,[sp] @ push r11 to stack
  1186. ldr r7, [r0, r1]
  1187. ldr_post r6, r0, r1, lsl #1 @ load source data
  1188. str r12,[sp, #4] @ save current reg before load q0 - q3 data
  1189. str lr, [sp, #8]
  1190. ldr lr, [r0, r1]
  1191. ldr_post r8, r0, r1, lsl #1
  1192. TRANSPOSE_MATRIX r6, r7, r8, lr, r12, r11, r10, r9
  1193. ldr lr, [sp, #8] @ load back (f)limit accumulator
  1194. uqsub8 r6, r12, r11 @ q3 - q2
  1195. uqsub8 r7, r11, r12 @ q2 - q3
  1196. uqsub8 r12, r11, r10 @ q2 - q1
  1197. uqsub8 r11, r10, r11 @ q1 - q2
  1198. orr r6, r6, r7 @ abs (q3-q2)
  1199. orr r7, r12, r11 @ abs (q2-q1)
  1200. uqsub8 r6, r6, r2 @ compare to limit
  1201. uqsub8 r7, r7, r2 @ compare to limit
  1202. ldr r11,[sp, #4] @ load back p0
  1203. ldr r12,[sp, #12] @ load back p1
  1204. orr lr, lr, r6
  1205. orr lr, lr, r7
  1206. uqsub8 r6, r11, r9 @ p0 - q0
  1207. uqsub8 r7, r9, r11 @ q0 - p0
  1208. uqsub8 r8, r12, r10 @ p1 - q1
  1209. uqsub8 r11, r10, r12 @ q1 - p1
  1210. orr r6, r6, r7 @ abs (p0-q0)
  1211. ldr r7, c0x7F7F7F7F
  1212. orr r8, r8, r11 @ abs (p1-q1)
  1213. uqadd8 r6, r6, r6 @ abs (p0-q0) * 2
  1214. and r8, r7, r8, lsr #1 @ abs (p1-q1) / 2
  1215. uqsub8 r11, r10, r9 @ q1 - q0
  1216. uqadd8 r6, r8, r6 @ abs (p0-q0)*2 + abs (p1-q1)/2
  1217. uqsub8 r12, r9, r10 @ q0 - q1
  1218. uqsub8 r6, r6, r4 @ compare to flimit
  1219. orr r9, r11, r12 @ abs (q1-q0)
  1220. uqsub8 r8, r9, r2 @ compare to limit
  1221. uqsub8 r10, r9, r3 @ compare to thresh
  1222. orr lr, lr, r6
  1223. orr lr, lr, r8
  1224. mvn r11, #0 @ r11 == -1
  1225. mov r12, #0
  1226. usub8 lr, r12, lr
  1227. ldr r9, [sp] @ load the compared result
  1228. sel lr, r11, r12 @ filter mask: lr
  1229. cmp lr, #0
  1230. beq 2f @ skip filtering
  1231. @vp8_hevmask() function
  1232. @calculate high edge variance
  1233. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  1234. orr r9, r9, r10
  1235. ldrh r7, [r0, #-2]
  1236. ldrh_post r8, r0, r1
  1237. usub8 r9, r12, r9
  1238. sel r6, r12, r11 @ hev mask: r6
  1239. @ vp8_mbfilter() function
  1240. @ p2, q2 are only needed at the end. do not need to load them in now.
  1241. @ Transpose needs 8 regs(r6 - r12, and lr). Save r6 and lr first
  1242. @ load soure data to r6, r11, r12, lr
  1243. ldrh r9, [r0, #-2]
  1244. ldrh_post r10, r0, r1
  1245. pkhbt r12, r7, r8, lsl #16
  1246. ldrh r7, [r0, #-2]
  1247. ldrh_post r8, r0, r1
  1248. pkhbt r11, r9, r10, lsl #16
  1249. ldrh r9, [r0, #-2]
  1250. ldrh_post r10, r0, r1
  1251. str r6, [sp] @ save r6
  1252. str lr, [sp, #4] @ save lr
  1253. pkhbt r6, r7, r8, lsl #16
  1254. pkhbt lr, r9, r10, lsl #16
  1255. @transpose r12, r11, r6, lr to p1, p0, q0, q1
  1256. TRANSPOSE_MATRIX r12, r11, r6, lr, r10, r9, r8, r7
  1257. @load back hev_mask r6 and filter_mask lr
  1258. ldr r12, c0x80808080
  1259. ldr r6, [sp]
  1260. ldr lr, [sp, #4]
  1261. eor r7, r7, r12 @ ps1
  1262. eor r8, r8, r12 @ ps0
  1263. eor r9, r9, r12 @ qs0
  1264. eor r10, r10, r12 @ qs1
  1265. qsub8 r12, r9, r8 @ vp8_signed_char_clamp(vp8_filter + 3 * ( qs0 - ps0))
  1266. str r7, [sp, #12] @ store ps1 temporarily
  1267. qsub8 r7, r7, r10 @ vp8_signed_char_clamp(ps1-qs1)
  1268. str r10,[sp, #8] @ store qs1 temporarily
  1269. qadd8 r7, r7, r12
  1270. str r9, [sp] @ store qs0 temporarily
  1271. qadd8 r7, r7, r12
  1272. str r8, [sp, #4] @ store ps0 temporarily
  1273. qadd8 r7, r7, r12 @ vp8_filter: r7
  1274. ldr r10, c0x03030303 @ r10 = 3 --modified for vp8
  1275. ldr r9, c0x04040404
  1276. and r7, r7, lr @ vp8_filter &= mask (lr is free)
  1277. mov r12, r7 @ Filter2: r12
  1278. and r12, r12, r6 @ Filter2 &= hev
  1279. @save bottom 3 bits so that we round one side +4 and the other +3
  1280. qadd8 r8, r12, r9 @ Filter1 (r8) = vp8_signed_char_clamp(Filter2+4)
  1281. qadd8 r12, r12, r10 @ Filter2 (r12) = vp8_signed_char_clamp(Filter2+3)
  1282. mov r10, #0
  1283. shadd8 r8, r8, r10 @ Filter1 >>= 3
  1284. shadd8 r12, r12, r10 @ Filter2 >>= 3
  1285. shadd8 r8, r8, r10
  1286. shadd8 r12, r12, r10
  1287. shadd8 r8, r8, r10 @ r8: Filter1
  1288. shadd8 r12, r12, r10 @ r12: Filter2
  1289. ldr r9, [sp] @ load qs0
  1290. ldr r11,[sp, #4] @ load ps0
  1291. qsub8 r9, r9, r8 @ qs0 = vp8_signed_char_clamp(qs0 - Filter1)
  1292. qadd8 r11, r11, r12 @ ps0 = vp8_signed_char_clamp(ps0 + Filter2)
  1293. bic r12, r7, r6 @vp8_filter &= ~hev ( r6 is free)
  1294. @roughly 3/7th difference across boundary
  1295. mov lr, #0x1b @ 27
  1296. mov r7, #0x3f @ 63
  1297. sxtb16 r6, r12
  1298. sxtb16 r10, r12, ror #8
  1299. smlabb r8, r6, lr, r7
  1300. smlatb r6, r6, lr, r7
  1301. smlabb r7, r10, lr, r7
  1302. smultb r10, r10, lr
  1303. ssat r8, #8, r8, asr #7
  1304. ssat r6, #8, r6, asr #7
  1305. add r10, r10, #63
  1306. ssat r7, #8, r7, asr #7
  1307. ssat r10, #8, r10, asr #7
  1308. ldr lr, c0x80808080
  1309. pkhbt r6, r8, r6, lsl #16
  1310. pkhbt r10, r7, r10, lsl #16
  1311. uxtb16 r6, r6
  1312. uxtb16 r10, r10
  1313. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  1314. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 27)>>7)
  1315. qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs0 - u)
  1316. qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps0 + u)
  1317. eor r8, r8, lr @ *oq0 = s^0x80
  1318. eor r10, r10, lr @ *op0 = s^0x80
  1319. strb r10,[r0, #-1] @ store op0 result
  1320. strb_post r8, r0, r1 @ store oq0 result
  1321. mov r10, r10, lsr #8
  1322. mov r8, r8, lsr #8
  1323. strb r10,[r0, #-1]
  1324. strb_post r8, r0, r1
  1325. mov r10, r10, lsr #8
  1326. mov r8, r8, lsr #8
  1327. strb r10,[r0, #-1]
  1328. strb_post r8, r0, r1
  1329. mov r10, r10, lsr #8
  1330. mov r8, r8, lsr #8
  1331. strb r10,[r0, #-1]
  1332. strb_post r8, r0, r1
  1333. @roughly 2/7th difference across boundary
  1334. mov lr, #0x12 @ 18
  1335. mov r7, #0x3f @ 63
  1336. sxtb16 r6, r12
  1337. sxtb16 r10, r12, ror #8
  1338. smlabb r8, r6, lr, r7
  1339. smlatb r6, r6, lr, r7
  1340. smlabb r9, r10, lr, r7
  1341. smlatb r10, r10, lr, r7
  1342. ssat r8, #8, r8, asr #7
  1343. ssat r6, #8, r6, asr #7
  1344. ssat r9, #8, r9, asr #7
  1345. ssat r10, #8, r10, asr #7
  1346. sub r0, r0, r1, lsl #2 @ move r0 pointer down by 4 lines
  1347. pkhbt r6, r8, r6, lsl #16
  1348. pkhbt r10, r9, r10, lsl #16
  1349. ldr r9, [sp, #8] @ load qs1
  1350. ldr r11,[sp, #12] @ load ps1
  1351. ldr lr, c0x80808080
  1352. uxtb16 r6, r6
  1353. uxtb16 r10, r10
  1354. add r0, r0, #2
  1355. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 18)>>7)
  1356. qsub8 r8, r9, r10 @ s = vp8_signed_char_clamp(qs1 - u)
  1357. qadd8 r10, r11, r10 @ s = vp8_signed_char_clamp(ps1 + u)
  1358. eor r8, r8, lr @ *oq1 = s^0x80
  1359. eor r10, r10, lr @ *op1 = s^0x80
  1360. ldrb r11,[r0, #-5] @ load p2 for 1/7th difference across boundary
  1361. strb r10,[r0, #-4] @ store op1
  1362. strb r8, [r0, #-1] @ store oq1
  1363. ldrb_post r9, r0, r1 @ load q2 for 1/7th difference across boundary
  1364. mov r10, r10, lsr #8
  1365. mov r8, r8, lsr #8
  1366. ldrb r6, [r0, #-5]
  1367. strb r10,[r0, #-4]
  1368. strb r8, [r0, #-1]
  1369. ldrb_post r7, r0, r1
  1370. mov r10, r10, lsr #8
  1371. mov r8, r8, lsr #8
  1372. orr r11, r11, r6, lsl #8
  1373. orr r9, r9, r7, lsl #8
  1374. ldrb r6, [r0, #-5]
  1375. strb r10,[r0, #-4]
  1376. strb r8, [r0, #-1]
  1377. ldrb_post r7, r0, r1
  1378. mov r10, r10, lsr #8
  1379. mov r8, r8, lsr #8
  1380. orr r11, r11, r6, lsl #16
  1381. orr r9, r9, r7, lsl #16
  1382. ldrb r6, [r0, #-5]
  1383. strb r10,[r0, #-4]
  1384. strb r8, [r0, #-1]
  1385. ldrb_post r7, r0, r1
  1386. orr r11, r11, r6, lsl #24
  1387. orr r9, r9, r7, lsl #24
  1388. @roughly 1/7th difference across boundary
  1389. eor r9, r9, lr
  1390. eor r11, r11, lr
  1391. mov lr, #0x9 @ 9
  1392. mov r7, #0x3f @ 63
  1393. sxtb16 r6, r12
  1394. sxtb16 r10, r12, ror #8
  1395. smlabb r8, r6, lr, r7
  1396. smlatb r6, r6, lr, r7
  1397. smlabb r12, r10, lr, r7
  1398. smlatb r10, r10, lr, r7
  1399. ssat r8, #8, r8, asr #7
  1400. ssat r6, #8, r6, asr #7
  1401. ssat r12, #8, r12, asr #7
  1402. ssat r10, #8, r10, asr #7
  1403. sub r0, r0, r1, lsl #2
  1404. pkhbt r6, r8, r6, lsl #16
  1405. pkhbt r10, r12, r10, lsl #16
  1406. uxtb16 r6, r6
  1407. uxtb16 r10, r10
  1408. ldr lr, c0x80808080
  1409. orr r10, r6, r10, lsl #8 @ u = vp8_signed_char_clamp((63 + Filter2 * 9)>>7)
  1410. qadd8 r8, r11, r10 @ s = vp8_signed_char_clamp(ps2 + u)
  1411. qsub8 r10, r9, r10 @ s = vp8_signed_char_clamp(qs2 - u)
  1412. eor r8, r8, lr @ *op2 = s^0x80
  1413. eor r10, r10, lr @ *oq2 = s^0x80
  1414. strb r8, [r0, #-5] @ store *op2
  1415. strb_post r10, r0, r1 @ store *oq2
  1416. mov r8, r8, lsr #8
  1417. mov r10, r10, lsr #8
  1418. strb r8, [r0, #-5]
  1419. strb_post r10, r0, r1
  1420. mov r8, r8, lsr #8
  1421. mov r10, r10, lsr #8
  1422. strb r8, [r0, #-5]
  1423. strb_post r10, r0, r1
  1424. mov r8, r8, lsr #8
  1425. mov r10, r10, lsr #8
  1426. strb r8, [r0, #-5]
  1427. strb_post r10, r0, r1
  1428. @adjust r0 pointer for next loop
  1429. sub r0, r0, #2
  1430. 2:
  1431. sub r0, r0, #4
  1432. subs r5, r5, #1
  1433. T ittt ne
  1434. ldrne r7, [r0, r1]
  1435. A ldrne r6, [r0], r1, lsl #1 @ load source data
  1436. T ldrne r6, [r0]
  1437. T addne r0, r0, r1, lsl #1
  1438. T ittt ne
  1439. ldrne lr, [r0, r1]
  1440. A ldrne r8, [r0], r1, lsl #1
  1441. T ldrne r8, [r0]
  1442. T addne r0, r0, r1, lsl #1
  1443. bne 1b
  1444. add sp, sp, #16
  1445. pop {r4 - r11, pc}
  1446. endfunc
  1447. @ MC
  1448. @ void put_vp8_pixels16(uint8_t *dst, int dststride, uint8_t *src,
  1449. @ int srcstride, int h, int mx, int my)
  1450. function ff_put_vp8_pixels16_armv6, export=1
  1451. push {r4 - r11}
  1452. ldr r12,[sp, #32] @ h
  1453. 1:
  1454. subs r12, r12, #2
  1455. ldr r5, [r2, #4]
  1456. ldr r6, [r2, #8]
  1457. ldr r7, [r2, #12]
  1458. ldr_post r4, r2, r3
  1459. ldr r9, [r2, #4]
  1460. ldr r10,[r2, #8]
  1461. ldr r11,[r2, #12]
  1462. ldr_post r8, r2, r3
  1463. strd r6, r7, [r0, #8]
  1464. strd_post r4, r5, r0, r1
  1465. strd r10, r11,[r0, #8]
  1466. strd_post r8, r9, r0, r1
  1467. bgt 1b
  1468. pop {r4 - r11}
  1469. bx lr
  1470. endfunc
  1471. @ void put_vp8_pixels8(uint8_t *dst, int dststride, uint8_t *src,
  1472. @ int srcstride, int h, int mx, int my)
  1473. function ff_put_vp8_pixels8_armv6, export=1
  1474. push {r4 - r11}
  1475. ldr r12,[sp, #32] @ h
  1476. 1:
  1477. subs r12, r12, #4
  1478. ldr r5, [r2, #4]
  1479. ldr_post r4, r2, r3
  1480. ldr r7, [r2, #4]
  1481. ldr_post r6, r2, r3
  1482. ldr r9, [r2, #4]
  1483. ldr_post r8, r2, r3
  1484. ldr r11,[r2, #4]
  1485. ldr_post r10, r2, r3
  1486. strd_post r4, r5, r0, r1
  1487. strd_post r6, r7, r0, r1
  1488. strd_post r8, r9, r0, r1
  1489. strd_post r10, r11, r0, r1
  1490. bgt 1b
  1491. pop {r4 - r11}
  1492. bx lr
  1493. endfunc
  1494. @ void put_vp8_pixels4(uint8_t *dst, int dststride, uint8_t *src,
  1495. @ int srcstride, int h, int mx, int my)
  1496. function ff_put_vp8_pixels4_armv6, export=1
  1497. ldr r12, [sp, #0] @ h
  1498. push {r4 - r6, lr}
  1499. 1:
  1500. subs r12, r12, #4
  1501. ldr r5, [r2, r3]
  1502. ldr_post r4, r2, r3, lsl #1
  1503. ldr lr, [r2, r3]
  1504. ldr_post r6, r2, r3, lsl #1
  1505. str r5, [r0, r1]
  1506. str_post r4, r0, r1, lsl #1
  1507. str lr, [r0, r1]
  1508. str_post r6, r0, r1, lsl #1
  1509. bgt 1b
  1510. pop {r4 - r6, pc}
  1511. endfunc
  1512. @ note: worst case sum of all 6-tap filter values * 255 is 0x7f80 so 16 bit
  1513. @ arithmatic can be used to apply filters
  1514. const sixtap_filters_13245600, align=4
  1515. .short 2, 108, -11, 36, -8, 1, 0, 0
  1516. .short 3, 77, -16, 77, -16, 3, 0, 0
  1517. .short 1, 36, -8, 108, -11, 2, 0, 0
  1518. endconst
  1519. const fourtap_filters_1324, align=4
  1520. .short -6, 12, 123, -1
  1521. .short -9, 50, 93, -6
  1522. .short -6, 93, 50, -9
  1523. .short -1, 123, 12, -6
  1524. endconst
  1525. @ void put_vp8_epel_h6(uint8_t *dst, int dststride, uint8_t *src,
  1526. @ int srcstride, int w, int h, int mx)
  1527. function ff_put_vp8_epel_h6_armv6, export=1
  1528. push {r4 - r11, lr}
  1529. sub r2, r2, #2
  1530. movrel lr, sixtap_filters_13245600 - 16
  1531. ldr r12,[sp, #44] @ vp8_filter index
  1532. ldr r4, [sp, #36] @ width
  1533. add lr, lr, r12, lsl #3
  1534. sub r3, r3, r4 @ src_stride - block_width
  1535. sub r1, r1, r4 @ dst_stride - block_width
  1536. lsr r4, #2
  1537. str r4, [sp, #36] @ "4-in-parallel" loop counter @40
  1538. str r3, [sp, #44] @ src_stride - block_width @48
  1539. push {r1} @ dst_stride - block_width @0
  1540. @ height @44
  1541. ldr r1, [lr], #4 @ coefficients
  1542. ldr r3, [lr], #4
  1543. ldr lr, [lr]
  1544. 1:
  1545. @ 3 loads, 10 shuffles and then mul/acc/add/shr
  1546. @ o0: i0/i1/i2/i3/i4/i5 -> i0/i2 (ld1) | i1/i3 (ld1) | i4/i5 (ld2)
  1547. @ o1: i1/i2/i3/i4/i5/i6 -> i1/i3 (ld1) | i2/i4 (ld2) | i5/i6 (ld2/3)
  1548. @ o2: i2/i3/i4/i5/i6/i7 -> i2/i4 (ld2) | i3/i5 (ld2) | i6/i7 (ld3)
  1549. @ o3: i3/i4/i5/i6/i7/i8 -> i3/i5 (ld2) | i4/i6 (ld2/3) | i7/i8 (ld3)
  1550. ldr r7, [r2, #5] @ ld3 -> src[5-8]
  1551. ldr r6, [r2, #2] @ ld2 -> src[2-5]
  1552. ldr r5, [r2], #4 @ ld1 -> src[0-3]
  1553. pkhtb r7, r7, r7, asr #8 @ src[8,7,7,6]
  1554. uxtb16 r9, r6, ror #8 @ src[5] | src[3]
  1555. uxtb16 r6, r6 @ src[4] | src[2]
  1556. uxtb16 r8, r5, ror #8 @ src[3] | src[1]
  1557. uxtb16 r11, r7, ror #8 @ src[8] | src[7]
  1558. uxtb16 r7, r7 @ src[7] | src[6]
  1559. pkhtb r10, r9, r6, asr #16 @ src[5] | src[4]
  1560. uxtb16 r5, r5 @ src[2] | src[0]
  1561. smuad r11, r11, lr @ filter[3][2] -> r11
  1562. subs r4, r4, #1
  1563. pkhbt r12, r10, r7, lsl #16 @ src[6] | src[4]
  1564. smuad r7, r7, lr @ filter[2][2] -> r7
  1565. smuad r5, r5, r1 @ filter[0][0] -> r5
  1566. smlad r11, r9, r1, r11 @ filter[3][0] -> r11
  1567. smlad r7, r9, r3, r7 @ filter[2][1] -> r7
  1568. smuad r9, r8, r1 @ filter[1][0] -> r9
  1569. smlad r5, r8, r3, r5 @ filter[0][1] -> r5
  1570. pkhtb r8, r12, r10, asr #16 @ src[6] | src[5]
  1571. smlad r11, r12, r3, r11 @ filter[3][1] -> r11
  1572. smlad r9, r6, r3, r9 @ filter[1][1] -> r9
  1573. smlad r5, r10, lr, r5 @ filter[0][2] -> r5
  1574. smlad r7, r6, r1, r7 @ filter[2][0] -> r7
  1575. smlad r9, r8, lr, r9 @ filter[1][2] -> r9
  1576. add r5, r5, #0x40 @ round_shift_and_clamp[0]
  1577. add r9, r9, #0x40 @ round_shift_and_clamp[1]
  1578. add r7, r7, #0x40 @ round_shift_and_clamp[2]
  1579. add r11, r11, #0x40 @ round_shift_and_clamp[3]
  1580. usat r5, #8, r5, asr #7
  1581. usat r9, #8, r9, asr #7
  1582. usat r7, #8, r7, asr #7
  1583. usat r11, #8, r11, asr #7
  1584. strb r5, [r0], #1 @ store res[0]
  1585. strb r9, [r0], #1 @ store res[1]
  1586. strb r7, [r0], #1 @ store res[2]
  1587. strb r11,[r0], #1 @ store res[3]
  1588. bne 1b
  1589. ldr r12,[sp, #44] @ height = outer-loop counter
  1590. subs r12, r12, #1
  1591. T itttt ne
  1592. ldrne r4, [sp, #40] @ 4-in-parallel loop counter
  1593. ldrne r5, [sp, #48]
  1594. ldrne r6, [sp]
  1595. strne r12,[sp, #44]
  1596. add r2, r2, r5 @ move to next input/output lines
  1597. add r0, r0, r6
  1598. bne 1b
  1599. add sp, sp, #4 @ restore stack after push{r1} above
  1600. pop {r4 - r11, pc}
  1601. endfunc
  1602. @ void put_vp8_epel_v6(uint8_t *dst, int dststride, uint8_t *src,
  1603. @ int srcstride, int w, int h, int my)
  1604. function ff_put_vp8_epel_v6_armv6, export=1
  1605. push {r4 - r11, lr}
  1606. movrel lr, sixtap_filters_13245600 - 16
  1607. ldr r12,[sp, #44] @ vp8_filter index
  1608. ldr r4, [sp, #36] @ width
  1609. add lr, lr, r12, lsl #3
  1610. sub r1, r1, r4 @ dst_stride - block_width
  1611. lsr r4, #2
  1612. str r4, [sp, #36] @ "4-in-parallel" loop counter @40
  1613. str r3, [sp, #44] @ src_stride - block_width @48
  1614. push {r1} @ dst_stride - block_width @0
  1615. @ height @44
  1616. 1:
  1617. add r1, r3, r3, lsl #1 @ stride * 3
  1618. ldr_dpren r5, r2, r3 @ src[0,1,2,3 + stride * 1]
  1619. ldr r6, [r2, r3] @ src[0,1,2,3 + stride * 3]
  1620. ldr r7, [r2, r3, lsl #1] @ src[0,1,2,3 + stride * 4]
  1621. ldr r8, [r2, r1] @ src[0,1,2,3 + stride * 5]
  1622. @ byte -> word and "transpose"
  1623. uxtb16 r9, r5, ror #8 @ src[3 + stride*1] | src[1 + stride*1]
  1624. uxtb16 r10, r6, ror #8 @ src[3 + stride*3] | src[1 + stride*3]
  1625. uxtb16 r11, r7, ror #8 @ src[3 + stride*4] | src[1 + stride*4]
  1626. uxtb16 r12, r8, ror #8 @ src[3 + stride*5] | src[1 + stride*5]
  1627. uxtb16 r5, r5 @ src[2 + stride*1] | src[0 + stride*1]
  1628. uxtb16 r6, r6 @ src[2 + stride*3] | src[0 + stride*3]
  1629. uxtb16 r7, r7 @ src[2 + stride*4] | src[0 + stride*4]
  1630. uxtb16 r8, r8 @ src[2 + stride*5] | src[0 + stride*5]
  1631. pkhbt r1, r9, r10, lsl #16 @ src[1 + stride*3] | src[1 + stride*1]
  1632. pkhtb r9, r10, r9, asr #16 @ src[3 + stride*3] | src[3 + stride*1]
  1633. pkhbt r10, r11, r12, lsl #16 @ src[1 + stride*5] | src[1 + stride*4]
  1634. pkhtb r11, r12, r11, asr #16 @ src[3 + stride*5] | src[3 + stride*4]
  1635. pkhbt r12, r5, r6, lsl #16 @ src[0 + stride*3] | src[0 + stride*1]
  1636. pkhtb r5, r6, r5, asr #16 @ src[2 + stride*3] | src[2 + stride*1]
  1637. pkhbt r6, r7, r8, lsl #16 @ src[0 + stride*5] | src[0 + stride*4]
  1638. pkhtb r7, r8, r7, asr #16 @ src[2 + stride*5] | src[2 + stride*4]
  1639. ldr r8, [lr, #4] @ stall - if only I had more registers...
  1640. smuad r12, r12, r8 @ filter[0][1]
  1641. smuad r1, r1, r8 @ filter[1][1]
  1642. smuad r5, r5, r8 @ filter[2][1]
  1643. smuad r9, r9, r8 @ filter[3][1]
  1644. ldr r8, [lr, #8] @ stall - if only I had more registers...
  1645. smlad r12, r6, r8, r12 @ filter[0][2]
  1646. smlad r1, r10, r8, r1 @ filter[1][2]
  1647. ldr_dpren r6, r2, r3, lsl #1 @ src[0,1,2,3 + stride * 0]
  1648. ldr r10,[r2], #4 @ src[0,1,2,3 + stride * 2]
  1649. smlad r5, r7, r8, r5 @ filter[2][2]
  1650. smlad r9, r11, r8, r9 @ filter[3][2]
  1651. uxtb16 r7, r6, ror #8 @ src[3 + stride*0] | src[1 + stride*0]
  1652. uxtb16 r11, r10, ror #8 @ src[3 + stride*2] | src[1 + stride*2]
  1653. uxtb16 r6, r6 @ src[2 + stride*0] | src[0 + stride*0]
  1654. uxtb16 r10, r10 @ src[2 + stride*2] | src[0 + stride*2]
  1655. pkhbt r8, r7, r11, lsl #16 @ src[1 + stride*2] | src[1 + stride*0]
  1656. pkhtb r7, r11, r7, asr #16 @ src[3 + stride*2] | src[3 + stride*0]
  1657. pkhbt r11, r6, r10, lsl #16 @ src[0 + stride*2] | src[0 + stride*0]
  1658. pkhtb r6, r10, r6, asr #16 @ src[2 + stride*2] | src[2 + stride*0]
  1659. ldr r10,[lr] @ stall - if only I had more registers...
  1660. subs r4, r4, #1 @ counter--
  1661. smlad r12, r11, r10, r12 @ filter[0][0]
  1662. smlad r1, r8, r10, r1 @ filter[1][0]
  1663. smlad r5, r6, r10, r5 @ filter[2][0]
  1664. smlad r9, r7, r10, r9 @ filter[3][0]
  1665. add r12, r12, #0x40 @ round_shift_and_clamp[0]
  1666. add r1, r1, #0x40 @ round_shift_and_clamp[1]
  1667. add r5, r5, #0x40 @ round_shift_and_clamp[2]
  1668. add r9, r9, #0x40 @ round_shift_and_clamp[3]
  1669. usat r12, #8, r12, asr #7
  1670. usat r1, #8, r1, asr #7
  1671. usat r5, #8, r5, asr #7
  1672. usat r9, #8, r9, asr #7
  1673. strb r12,[r0], #1 @ store res[0]
  1674. strb r1, [r0], #1 @ store res[1]
  1675. strb r5, [r0], #1 @ store res[2]
  1676. strb r9, [r0], #1 @ store res[3]
  1677. bne 1b
  1678. ldr r12,[sp, #44] @ height = outer-loop counter
  1679. subs r12, r12, #1
  1680. T itttt ne
  1681. ldrne r4, [sp, #40] @ 4-in-parallel loop counter
  1682. ldrne r6, [sp, #0]
  1683. subne r2, r2, r4, lsl #2
  1684. strne r12,[sp, #44]
  1685. add r0, r0, r6
  1686. add r2, r2, r3 @ move to next input/output lines
  1687. bne 1b
  1688. add sp, sp, #4 @ restore stack after push{r1} above
  1689. pop {r4 - r11, pc}
  1690. endfunc
  1691. @ void put_vp8_epel_h4(uint8_t *dst, int dststride, uint8_t *src,
  1692. @ int srcstride, int w, int h, int mx)
  1693. function ff_put_vp8_epel_h4_armv6, export=1
  1694. push {r4 - r11, lr}
  1695. subs r2, r2, #1
  1696. movrel lr, fourtap_filters_1324 - 4
  1697. ldr r4, [sp, #36] @ width
  1698. ldr r12,[sp, #44] @ vp8_filter index
  1699. add lr, lr, r12, lsl #2
  1700. sub r3, r3, r4 @ src_stride - block_width
  1701. sub r1, r1, r4 @ dst_stride - block_width
  1702. ldr r5, [lr]
  1703. ldr r6, [lr, #4]
  1704. asr r4, #2
  1705. ldr lr, [sp, #40] @ height = outer-loop counter
  1706. str r4, [sp, #36] @ "4-in-parallel" inner loop counter
  1707. 1:
  1708. @ 3 loads, 5 uxtb16s and then mul/acc/add/shr
  1709. @ o0: i0/i1/i2/i3 -> i0/i2(ld1) + i1/i3(ld1)
  1710. @ o1: i1/i2/i3/i4 -> i1/i3(ld1) + i2/i4(ld2)
  1711. @ o2: i2/i3/i4/i5 -> i2/i4(ld2) + i3/i5(ld2)
  1712. @ o3: i3/i4/i5/i6 -> i3/i5(ld2) + i4/i6(ld3)
  1713. ldr r9, [r2, #3] @ load source data
  1714. ldr r8, [r2, #2]
  1715. ldr r7, [r2], #4
  1716. uxtb16 r9, r9, ror #8 @ src[6] | src[4]
  1717. uxtb16 r10, r8, ror #8 @ src[5] | src[3]
  1718. uxtb16 r8, r8 @ src[4] | src[2]
  1719. uxtb16 r11, r7, ror #8 @ src[3] | src[1]
  1720. uxtb16 r7, r7 @ src[2] | src[0]
  1721. smuad r9, r9, r6 @ filter[3][1] -> r9
  1722. smuad r12, r10, r6 @ filter[2][1] -> r12
  1723. smuad r7, r7, r5 @ filter[0][0] -> r7
  1724. smlad r9, r10, r5, r9 @ filter[3][0] -> r9
  1725. smuad r10, r11, r5 @ filter[1][0] -> r10
  1726. smlad r12, r8, r5, r12 @ filter[2][0] -> r12
  1727. smlad r7, r11, r6, r7 @ filter[0][1] -> r7
  1728. smlad r10, r8, r6, r10 @ filter[1][1] -> r10
  1729. subs r4, r4, #1 @ counter--
  1730. add r7, r7, #0x40 @ round_shift_and_clamp[0]
  1731. add r10, r10, #0x40 @ round_shift_and_clamp[1]
  1732. add r12, r12, #0x40 @ round_shift_and_clamp[2]
  1733. add r9, r9, #0x40 @ round_shift_and_clamp[3]
  1734. usat r7, #8, r7, asr #7
  1735. usat r10, #8, r10, asr #7
  1736. usat r12, #8, r12, asr #7
  1737. usat r9, #8, r9, asr #7
  1738. strb r7, [r0], #1 @ store res[0]
  1739. strb r10,[r0], #1 @ store res[1]
  1740. strb r12,[r0], #1 @ store res[2]
  1741. strb r9, [r0], #1 @ store res[3]
  1742. bne 1b
  1743. subs lr, lr, #1
  1744. T it ne
  1745. ldrne r4, [sp, #36] @ 4-in-parallel loop counter
  1746. add r2, r2, r3 @ move to next input/output lines
  1747. add r0, r0, r1
  1748. bne 1b
  1749. pop {r4 - r11, pc}
  1750. endfunc
  1751. @ void put_vp8_epel_v4(uint8_t *dst, int dststride, uint8_t *src,
  1752. @ int srcstride, int w, int h, int my)
  1753. function ff_put_vp8_epel_v4_armv6, export=1
  1754. push {r4 - r11, lr}
  1755. movrel lr, fourtap_filters_1324 - 4
  1756. ldr r12,[sp, #44] @ vp8_filter index
  1757. ldr r4, [sp, #36] @ width
  1758. add lr, lr, r12, lsl #2
  1759. sub r1, r1, r4 @ dst_stride - block_width
  1760. asr r4, #2
  1761. ldr r5, [lr]
  1762. ldr r6, [lr, #4]
  1763. str r4, [sp, #36] @ "4-in-parallel" loop counter @40
  1764. str r3, [sp, #44] @ src_stride @48
  1765. push {r1} @ dst_stride - block_width @36
  1766. @ height @44
  1767. 1:
  1768. ldr lr, [r2, r3, lsl #1] @ load source pixels
  1769. ldr r12,[r2, r3]
  1770. ldr_dpren r7, r2, r3
  1771. ldr r11,[r2], #4
  1772. @ byte -> word and "transpose"
  1773. uxtb16 r8, lr, ror #8 @ src[3 + stride*3] | src[1 + stride*3]
  1774. uxtb16 r9, r12, ror #8 @ src[3 + stride*2] | src[1 + stride*2]
  1775. uxtb16 r3, r7, ror #8 @ src[3 + stride*0] | src[1 + stride*0]
  1776. uxtb16 r1, r11, ror #8 @ src[3 + stride*1] | src[1 + stride*1]
  1777. uxtb16 lr, lr @ src[2 + stride*3] | src[0 + stride*3]
  1778. uxtb16 r12, r12 @ src[2 + stride*2] | src[0 + stride*2]
  1779. uxtb16 r7, r7 @ src[2 + stride*0] | src[0 + stride*0]
  1780. uxtb16 r11, r11 @ src[2 + stride*1] | src[0 + stride*1]
  1781. pkhbt r10, r1, r8, lsl #16 @ src[1 + stride*3] | src[1 + stride*1]
  1782. pkhtb r1, r8, r1, asr #16 @ src[3 + stride*3] | src[3 + stride*1]
  1783. pkhbt r8, r3, r9, lsl #16 @ src[1 + stride*2] | src[1 + stride*0]
  1784. pkhtb r3, r9, r3, asr #16 @ src[3 + stride*2] | src[3 + stride*0]
  1785. pkhbt r9, r11, lr, lsl #16 @ src[0 + stride*3] | src[0 + stride*1]
  1786. pkhtb r11, lr, r11, asr #16 @ src[2 + stride*3] | src[2 + stride*1]
  1787. pkhbt lr, r7, r12, lsl #16 @ src[0 + stride*2] | src[0 + stride*0]
  1788. pkhtb r7, r12, r7, asr #16 @ src[2 + stride*2] | src[2 + stride*0]
  1789. smuad r9, r9, r6 @ filter[0][1]
  1790. smuad r10, r10, r6 @ filter[1][1]
  1791. smuad r11, r11, r6 @ filter[2][1]
  1792. smuad r1, r1, r6 @ filter[3][1]
  1793. smlad r9, lr, r5, r9 @ filter[0][0]
  1794. smlad r10, r8, r5, r10 @ filter[1][0]
  1795. smlad r11, r7, r5, r11 @ filter[2][0]
  1796. smlad r1, r3, r5, r1 @ filter[3][0]
  1797. subs r4, r4, #1 @ counter--
  1798. ldr r3, [sp, #48] @ FIXME prevent clobber of r3 above?
  1799. add r9, r9, #0x40 @ round_shift_and_clamp[0]
  1800. add r10, r10, #0x40 @ round_shift_and_clamp[1]
  1801. add r11, r11, #0x40 @ round_shift_and_clamp[2]
  1802. add r1, r1, #0x40 @ round_shift_and_clamp[3]
  1803. usat r9, #8, r9, asr #7
  1804. usat r10, #8, r10, asr #7
  1805. usat r11, #8, r11, asr #7
  1806. usat r1, #8, r1, asr #7
  1807. strb r9, [r0], #1 @ store result
  1808. strb r10,[r0], #1
  1809. strb r11,[r0], #1
  1810. strb r1, [r0], #1
  1811. bne 1b
  1812. ldr r12,[sp, #44] @ height = outer-loop counter
  1813. subs r12, r12, #1
  1814. T ittt ne
  1815. ldrne r4, [sp, #40] @ 4-in-parallel loop counter
  1816. ldrne r9, [sp, #0]
  1817. strne r12,[sp, #44]
  1818. sub r2, r2, r4, lsl #2
  1819. add r0, r0, r9
  1820. add r2, r2, r3 @ move to next input/output lines
  1821. bne 1b
  1822. add sp, sp, #4 @ restore stack after push{r1} above
  1823. pop {r4 - r11, pc}
  1824. endfunc
  1825. @ void put_vp8_bilin_h(uint8_t *dst, int dststride, uint8_t *src,
  1826. @ int srcstride, int w, int h, int mx)
  1827. function ff_put_vp8_bilin_h_armv6, export=1
  1828. push {r4 - r9, lr}
  1829. ldr r8, [sp, #36] @ vp8_filter index
  1830. ldr r12,[sp, #32] @ height = outer-loop counter
  1831. ldr r4, [sp, #28] @ width
  1832. lsl r5, r8, #16 @ mx << 16
  1833. sub r3, r3, r4 @ src_stride - block_width
  1834. sub r1, r1, r4 @ dst_stride - block_width
  1835. asr r4, #2
  1836. sub r5, r5, r8 @ (mx << 16) | (-mx)
  1837. str r4, [sp, #28] @ "4-in-parallel" loop counter
  1838. add r5, r5, #8 @ (8 - mx) | (mx << 16) = filter coefficients
  1839. 1:
  1840. ldrb r6, [r2], #1 @ load source data
  1841. ldrb r7, [r2], #1
  1842. ldrb r8, [r2], #1
  1843. ldrb r9, [r2], #1
  1844. ldrb lr, [r2]
  1845. pkhbt r6, r6, r7, lsl #16 @ src[1] | src[0]
  1846. pkhbt r7, r7, r8, lsl #16 @ src[2] | src[1]
  1847. pkhbt r8, r8, r9, lsl #16 @ src[3] | src[2]
  1848. pkhbt r9, r9, lr, lsl #16 @ src[4] | src[3]
  1849. smuad r6, r6, r5 @ apply the filter
  1850. smuad r7, r7, r5
  1851. smuad r8, r8, r5
  1852. smuad r9, r9, r5
  1853. subs r4, r4, #1 @ counter--
  1854. add r6, r6, #0x4 @ round_shift_and_clamp
  1855. add r7, r7, #0x4
  1856. add r8, r8, #0x4
  1857. add r9, r9, #0x4
  1858. asr r6, #3
  1859. asr r7, #3
  1860. pkhbt r6, r6, r8, lsl #13
  1861. pkhbt r7, r7, r9, lsl #13
  1862. orr r6, r6, r7, lsl #8
  1863. str r6, [r0], #4 @ store result
  1864. bne 1b
  1865. ldr r4, [sp, #28] @ 4-in-parallel loop counter
  1866. subs r12, r12, #1
  1867. add r2, r2, r3 @ move to next input/output lines
  1868. add r0, r0, r1
  1869. bne 1b
  1870. pop {r4 - r9, pc}
  1871. endfunc
  1872. @ void put_vp8_bilin_v(uint8_t *dst, int dststride, uint8_t *src,
  1873. @ int srcstride, int w, int h, int my)
  1874. function ff_put_vp8_bilin_v_armv6, export=1
  1875. push {r4 - r11, lr}
  1876. ldr r11,[sp, #44] @ vp8_filter index
  1877. ldr r4, [sp, #36] @ width
  1878. mov r5, r11, lsl #16 @ mx << 16
  1879. ldr r12,[sp, #40] @ height = outer-loop counter
  1880. sub r1, r1, r4
  1881. sub r5, r5, r11 @ (mx << 16) | (-mx)
  1882. asr r4, #2
  1883. add r5, r5, #8 @ (8 - mx) | (mx << 16) = filter coefficients
  1884. str r4, [sp, #36] @ "4-in-parallel" loop counter
  1885. 1:
  1886. ldrb r10,[r2, r3] @ load the data
  1887. ldrb r6, [r2], #1
  1888. ldrb r11,[r2, r3]
  1889. ldrb r7, [r2], #1
  1890. ldrb lr, [r2, r3]
  1891. ldrb r8, [r2], #1
  1892. ldrb r9, [r2, r3]
  1893. pkhbt r6, r6, r10, lsl #16
  1894. ldrb r10,[r2], #1
  1895. pkhbt r7, r7, r11, lsl #16
  1896. pkhbt r8, r8, lr, lsl #16
  1897. pkhbt r9, r10, r9, lsl #16
  1898. smuad r6, r6, r5 @ apply the filter
  1899. smuad r7, r7, r5
  1900. smuad r8, r8, r5
  1901. smuad r9, r9, r5
  1902. subs r4, r4, #1 @ counter--
  1903. add r6, r6, #0x4 @ round_shift_and_clamp
  1904. add r7, r7, #0x4
  1905. add r8, r8, #0x4
  1906. add r9, r9, #0x4
  1907. asr r6, #3
  1908. asr r7, #3
  1909. pkhbt r6, r6, r8, lsl #13
  1910. pkhbt r7, r7, r9, lsl #13
  1911. orr r6, r6, r7, lsl #8
  1912. str r6, [r0], #4 @ store result
  1913. bne 1b
  1914. ldr r4, [sp, #36] @ 4-in-parallel loop counter
  1915. subs r12, r12, #1
  1916. add r2, r2, r3 @ move to next input/output lines
  1917. add r0, r0, r1
  1918. sub r2, r2, r4, lsl #2
  1919. bne 1b
  1920. pop {r4 - r11, pc}
  1921. endfunc