vpx_convolve8_neon_asm.asm 7.2 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273
  1. ;
  2. ; Copyright (c) 2013 The WebM project authors. All Rights Reserved.
  3. ;
  4. ; Use of this source code is governed by a BSD-style license
  5. ; that can be found in the LICENSE file in the root of the source
  6. ; tree. An additional intellectual property rights grant can be found
  7. ; in the file PATENTS. All contributing project authors may
  8. ; be found in the AUTHORS file in the root of the source tree.
  9. ;
  10. ; These functions are only valid when:
  11. ; x_step_q4 == 16
  12. ; w%4 == 0
  13. ; h%4 == 0
  14. ; taps == 8
  15. ; VP9_FILTER_WEIGHT == 128
  16. ; VP9_FILTER_SHIFT == 7
  17. EXPORT |vpx_convolve8_horiz_neon|
  18. EXPORT |vpx_convolve8_vert_neon|
  19. ARM
  20. REQUIRE8
  21. PRESERVE8
  22. AREA ||.text||, CODE, READONLY, ALIGN=2
  23. ; Multiply and accumulate by q0
  24. MACRO
  25. MULTIPLY_BY_Q0 $dst, $src0, $src1, $src2, $src3, $src4, $src5, $src6, $src7
  26. vmull.s16 $dst, $src0, d0[0]
  27. vmlal.s16 $dst, $src1, d0[1]
  28. vmlal.s16 $dst, $src2, d0[2]
  29. vmlal.s16 $dst, $src3, d0[3]
  30. vmlal.s16 $dst, $src4, d1[0]
  31. vmlal.s16 $dst, $src5, d1[1]
  32. vmlal.s16 $dst, $src6, d1[2]
  33. vmlal.s16 $dst, $src7, d1[3]
  34. MEND
  35. ; r0 const uint8_t *src
  36. ; r1 int src_stride
  37. ; r2 uint8_t *dst
  38. ; r3 int dst_stride
  39. ; sp[]const int16_t *filter
  40. ; sp[]int x0_q4
  41. ; sp[]int x_step_q4 ; unused
  42. ; sp[]int y0_q4
  43. ; sp[]int y_step_q4 ; unused
  44. ; sp[]int w
  45. ; sp[]int h
  46. |vpx_convolve8_horiz_neon| PROC
  47. push {r4-r10, lr}
  48. sub r0, r0, #3 ; adjust for taps
  49. ldrd r4, r5, [sp, #32] ; filter, x0_q4
  50. add r4, r5, lsl #4
  51. ldrd r6, r7, [sp, #52] ; w, h
  52. vld1.s16 {q0}, [r4] ; filter
  53. sub r8, r1, r1, lsl #2 ; -src_stride * 3
  54. add r8, r8, #4 ; -src_stride * 3 + 4
  55. sub r4, r3, r3, lsl #2 ; -dst_stride * 3
  56. add r4, r4, #4 ; -dst_stride * 3 + 4
  57. rsb r9, r6, r1, lsl #2 ; reset src for outer loop
  58. sub r9, r9, #7
  59. rsb r12, r6, r3, lsl #2 ; reset dst for outer loop
  60. mov r10, r6 ; w loop counter
  61. vpx_convolve8_loop_horiz_v
  62. vld1.8 {d24}, [r0], r1
  63. vld1.8 {d25}, [r0], r1
  64. vld1.8 {d26}, [r0], r1
  65. vld1.8 {d27}, [r0], r8
  66. vtrn.16 q12, q13
  67. vtrn.8 d24, d25
  68. vtrn.8 d26, d27
  69. pld [r0, r1, lsl #2]
  70. vmovl.u8 q8, d24
  71. vmovl.u8 q9, d25
  72. vmovl.u8 q10, d26
  73. vmovl.u8 q11, d27
  74. ; save a few instructions in the inner loop
  75. vswp d17, d18
  76. vmov d23, d21
  77. add r0, r0, #3
  78. vpx_convolve8_loop_horiz
  79. add r5, r0, #64
  80. vld1.32 {d28[]}, [r0], r1
  81. vld1.32 {d29[]}, [r0], r1
  82. vld1.32 {d31[]}, [r0], r1
  83. vld1.32 {d30[]}, [r0], r8
  84. pld [r5]
  85. vtrn.16 d28, d31
  86. vtrn.16 d29, d30
  87. vtrn.8 d28, d29
  88. vtrn.8 d31, d30
  89. pld [r5, r1]
  90. ; extract to s16
  91. vtrn.32 q14, q15
  92. vmovl.u8 q12, d28
  93. vmovl.u8 q13, d29
  94. pld [r5, r1, lsl #1]
  95. ; src[] * filter
  96. MULTIPLY_BY_Q0 q1, d16, d17, d20, d22, d18, d19, d23, d24
  97. MULTIPLY_BY_Q0 q2, d17, d20, d22, d18, d19, d23, d24, d26
  98. MULTIPLY_BY_Q0 q14, d20, d22, d18, d19, d23, d24, d26, d27
  99. MULTIPLY_BY_Q0 q15, d22, d18, d19, d23, d24, d26, d27, d25
  100. pld [r5, -r8]
  101. ; += 64 >> 7
  102. vqrshrun.s32 d2, q1, #7
  103. vqrshrun.s32 d3, q2, #7
  104. vqrshrun.s32 d4, q14, #7
  105. vqrshrun.s32 d5, q15, #7
  106. ; saturate
  107. vqmovn.u16 d2, q1
  108. vqmovn.u16 d3, q2
  109. ; transpose
  110. vtrn.16 d2, d3
  111. vtrn.32 d2, d3
  112. vtrn.8 d2, d3
  113. vst1.u32 {d2[0]}, [r2@32], r3
  114. vst1.u32 {d3[0]}, [r2@32], r3
  115. vst1.u32 {d2[1]}, [r2@32], r3
  116. vst1.u32 {d3[1]}, [r2@32], r4
  117. vmov q8, q9
  118. vmov d20, d23
  119. vmov q11, q12
  120. vmov q9, q13
  121. subs r6, r6, #4 ; w -= 4
  122. bgt vpx_convolve8_loop_horiz
  123. ; outer loop
  124. mov r6, r10 ; restore w counter
  125. add r0, r0, r9 ; src += src_stride * 4 - w
  126. add r2, r2, r12 ; dst += dst_stride * 4 - w
  127. subs r7, r7, #4 ; h -= 4
  128. bgt vpx_convolve8_loop_horiz_v
  129. pop {r4-r10, pc}
  130. ENDP
  131. |vpx_convolve8_vert_neon| PROC
  132. push {r4-r8, lr}
  133. ; adjust for taps
  134. sub r0, r0, r1
  135. sub r0, r0, r1, lsl #1
  136. ldr r4, [sp, #24] ; filter
  137. ldr r5, [sp, #36] ; y0_q4
  138. add r4, r5, lsl #4
  139. ldr r6, [sp, #44] ; w
  140. ldr lr, [sp, #48] ; h
  141. vld1.s16 {q0}, [r4] ; filter
  142. lsl r1, r1, #1
  143. lsl r3, r3, #1
  144. vpx_convolve8_loop_vert_h
  145. mov r4, r0
  146. add r7, r0, r1, asr #1
  147. mov r5, r2
  148. add r8, r2, r3, asr #1
  149. mov r12, lr ; h loop counter
  150. vld1.u32 {d16[0]}, [r4], r1
  151. vld1.u32 {d16[1]}, [r7], r1
  152. vld1.u32 {d18[0]}, [r4], r1
  153. vld1.u32 {d18[1]}, [r7], r1
  154. vld1.u32 {d20[0]}, [r4], r1
  155. vld1.u32 {d20[1]}, [r7], r1
  156. vld1.u32 {d22[0]}, [r4], r1
  157. vmovl.u8 q8, d16
  158. vmovl.u8 q9, d18
  159. vmovl.u8 q10, d20
  160. vmovl.u8 q11, d22
  161. vpx_convolve8_loop_vert
  162. ; always process a 4x4 block at a time
  163. vld1.u32 {d24[0]}, [r7], r1
  164. vld1.u32 {d26[0]}, [r4], r1
  165. vld1.u32 {d26[1]}, [r7], r1
  166. vld1.u32 {d24[1]}, [r4], r1
  167. ; extract to s16
  168. vmovl.u8 q12, d24
  169. vmovl.u8 q13, d26
  170. pld [r5]
  171. pld [r8]
  172. ; src[] * filter
  173. MULTIPLY_BY_Q0 q1, d16, d17, d18, d19, d20, d21, d22, d24
  174. pld [r5, r3]
  175. pld [r8, r3]
  176. MULTIPLY_BY_Q0 q2, d17, d18, d19, d20, d21, d22, d24, d26
  177. pld [r7]
  178. pld [r4]
  179. MULTIPLY_BY_Q0 q14, d18, d19, d20, d21, d22, d24, d26, d27
  180. pld [r7, r1]
  181. pld [r4, r1]
  182. MULTIPLY_BY_Q0 q15, d19, d20, d21, d22, d24, d26, d27, d25
  183. ; += 64 >> 7
  184. vqrshrun.s32 d2, q1, #7
  185. vqrshrun.s32 d3, q2, #7
  186. vqrshrun.s32 d4, q14, #7
  187. vqrshrun.s32 d5, q15, #7
  188. ; saturate
  189. vqmovn.u16 d2, q1
  190. vqmovn.u16 d3, q2
  191. vst1.u32 {d2[0]}, [r5@32], r3
  192. vst1.u32 {d2[1]}, [r8@32], r3
  193. vst1.u32 {d3[0]}, [r5@32], r3
  194. vst1.u32 {d3[1]}, [r8@32], r3
  195. vmov q8, q10
  196. vmov d18, d22
  197. vmov d19, d24
  198. vmov q10, q13
  199. vmov d22, d25
  200. subs r12, r12, #4 ; h -= 4
  201. bgt vpx_convolve8_loop_vert
  202. ; outer loop
  203. add r0, r0, #4
  204. add r2, r2, #4
  205. subs r6, r6, #4 ; w -= 4
  206. bgt vpx_convolve8_loop_vert_h
  207. pop {r4-r8, pc}
  208. ENDP
  209. END