]> git.sesse.net Git - ffmpeg/blobdiff - libavcodec/arm/mdct_neon.S
arm: asm decode_block_coeffs_internal is vp8 specific
[ffmpeg] / libavcodec / arm / mdct_neon.S
index 4c65f8ff9e52fda572e7d1596965505959c97872..bfe259c396c4a7894d01e184fc24193fd51d8884 100644 (file)
@@ -2,48 +2,45 @@
  * ARM NEON optimised MDCT
  * Copyright (c) 2009 Mans Rullgard <mans@mansr.com>
  *
- * This file is part of FFmpeg.
+ * This file is part of Libav.
  *
- * FFmpeg is free software; you can redistribute it and/or
+ * Libav is free software; you can redistribute it and/or
  * modify it under the terms of the GNU Lesser General Public
  * License as published by the Free Software Foundation; either
  * version 2.1 of the License, or (at your option) any later version.
  *
- * FFmpeg is distributed in the hope that it will be useful,
+ * Libav is distributed in the hope that it will be useful,
  * but WITHOUT ANY WARRANTY; without even the implied warranty of
  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
  * Lesser General Public License for more details.
  *
  * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
+ * License along with Libav; if not, write to the Free Software
  * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  */
 
-#include "asm.S"
+#include "libavutil/arm/asm.S"
 
-        .fpu neon
-        .text
+#define ff_fft_calc_neon X(ff_fft_calc_neon)
 
 function ff_imdct_half_neon, export=1
         push            {r4-r8,lr}
 
         mov             r12, #1
-        ldr             lr,  [r0, #28]          @ mdct_bits
-        ldr             r4,  [r0, #32]          @ tcos
-        ldr             r5,  [r0, #36]          @ tsin
+        ldr             lr,  [r0, #20]          @ mdct_bits
+        ldr             r4,  [r0, #24]          @ tcos
         ldr             r3,  [r0, #8]           @ revtab
         lsl             r12, r12, lr            @ n  = 1 << nbits
         lsr             lr,  r12, #2            @ n4 = n >> 2
         add             r7,  r2,  r12,  lsl #1
-        mov             r12,  #-16
+        mov             r12, #-16
         sub             r7,  r7,  #16
 
         vld2.32         {d16-d17},[r7,:128],r12 @ d16=x,n1 d17=x,n0
         vld2.32         {d0-d1},  [r2,:128]!    @ d0 =m0,x d1 =m1,x
         vrev64.32       d17, d17
-        vld1.32         {d2},     [r4,:64]!     @ d2=c0,c1
+        vld2.32         {d2,d3},  [r4,:128]!    @ d2=c0,c1 d3=s0,s2
         vmul.f32        d6,  d17, d2
-        vld1.32         {d3},     [r5,:64]!     @ d3=s0,s1
         vmul.f32        d7,  d0,  d2
 1:
         subs            lr,  lr,  #2
@@ -60,9 +57,8 @@ function ff_imdct_half_neon, export=1
         vld2.32         {d16-d17},[r7,:128],r12
         vld2.32         {d0-d1},  [r2,:128]!
         vrev64.32       d17, d17
-        vld1.32         {d2},     [r4,:64]!
+        vld2.32         {d2,d3},  [r4,:128]!    @ d2=c0,c1 d3=s0,s2
         vmul.f32        d6,  d17, d2
-        vld1.32         {d3},     [r5,:64]!
         vmul.f32        d7,  d0,  d2
         vst2.32         {d4[0],d5[0]}, [r6,:64]
         vst2.32         {d4[1],d5[1]}, [r8,:64]
@@ -76,35 +72,29 @@ function ff_imdct_half_neon, export=1
         bl              ff_fft_calc_neon
 
         mov             r12, #1
-        ldr             lr,  [r4, #28]          @ mdct_bits
-        ldr             r5,  [r4, #36]          @ tsin
-        ldr             r4,  [r4, #32]          @ tcos
+        ldr             lr,  [r4, #20]          @ mdct_bits
+        ldr             r4,  [r4, #24]          @ tcos
         lsl             r12, r12, lr            @ n  = 1 << nbits
         lsr             lr,  r12, #3            @ n8 = n >> 3
 
-        add             r4,  r4,  lr,  lsl #2
-        add             r5,  r5,  lr,  lsl #2
+        add             r4,  r4,  lr,  lsl #3
         add             r6,  r6,  lr,  lsl #3
-        sub             r1,  r4,  #8
-        sub             r2,  r5,  #8
+        sub             r1,  r4,  #16
         sub             r3,  r6,  #16
 
         mov             r7,  #-16
-        mov             r12, #-8
         mov             r8,  r6
         mov             r0,  r3
 
         vld2.32         {d0-d1},  [r3,:128], r7 @ d0 =i1,r1 d1 =i0,r0
         vld2.32         {d20-d21},[r6,:128]!    @ d20=i2,r2 d21=i3,r3
-        vld1.32         {d18},    [r2,:64], r12 @ d18=s1,s0
+        vld2.32         {d16,d18},[r1,:128], r7 @ d16=c1,c0 d18=s1,s0
 1:
         subs            lr,  lr,  #2
         vmul.f32        d7,  d0,  d18
-        vld1.32         {d19},    [r5,:64]!     @ d19=s2,s3
+        vld2.32         {d17,d19},[r4,:128]!    @ d17=c2,c3 d19=s2,s3
         vmul.f32        d4,  d1,  d18
-        vld1.32         {d16},    [r1,:64], r12 @ d16=c1,c0
         vmul.f32        d5,  d21, d19
-        vld1.32         {d17},    [r4,:64]!     @ d17=c2,c3
         vmul.f32        d6,  d20, d19
         vmul.f32        d22, d1,  d16
         vmul.f32        d23, d21, d17
@@ -117,7 +107,7 @@ function ff_imdct_half_neon, export=1
         beq             1f
         vld2.32         {d0-d1},  [r3,:128], r7
         vld2.32         {d20-d21},[r6,:128]!
-        vld1.32         {d18},    [r2,:64], r12
+        vld2.32         {d16,d18},[r1,:128], r7 @ d16=c1,c0 d18=s1,s0
         vrev64.32       q3,  q3
         vst2.32         {d4,d6},  [r0,:128], r7
         vst2.32         {d5,d7},  [r8,:128]!
@@ -128,18 +118,18 @@ function ff_imdct_half_neon, export=1
         vst2.32         {d5,d7},  [r8,:128]
 
         pop             {r4-r8,pc}
-.endfunc
+endfunc
 
 function ff_imdct_calc_neon, export=1
         push            {r4-r6,lr}
 
-        ldr             r3,  [r0, #28]
+        ldr             r3,  [r0, #20]
         mov             r4,  #1
         mov             r5,  r1
         lsl             r4,  r4,  r3
         add             r1,  r1,  r4
 
-        bl              ff_imdct_half_neon
+        bl              X(ff_imdct_half_neon)
 
         add             r0,  r5,  r4,  lsl #2
         add             r1,  r5,  r4,  lsl #1
@@ -164,145 +154,126 @@ function ff_imdct_calc_neon, export=1
         bgt             1b
 
         pop             {r4-r6,pc}
-.endfunc
+endfunc
 
 function ff_mdct_calc_neon, export=1
         push            {r4-r10,lr}
 
         mov             r12, #1
-        ldr             lr,  [r0, #28]          @ mdct_bits
-        ldr             r4,  [r0, #32]          @ tcos
-        ldr             r5,  [r0, #36]          @ tsin
+        ldr             lr,  [r0, #20]          @ mdct_bits
+        ldr             r4,  [r0, #24]          @ tcos
         ldr             r3,  [r0, #8]           @ revtab
         lsl             lr,  r12, lr            @ n  = 1 << nbits
         add             r7,  r2,  lr            @ in4u
         sub             r9,  r7,  #16           @ in4d
         add             r2,  r7,  lr,  lsl #1   @ in3u
         add             r8,  r9,  lr,  lsl #1   @ in3d
+        add             r5,  r4,  lr,  lsl #1
+        sub             r5,  r5,  #16
+        sub             r3,  r3,  #4
         mov             r12, #-16
 
-        vld2.32         {d16,d18},[r9,:128],r12 @ x,x in4d1,in4d0
-        vld2.32         {d17,d19},[r8,:128],r12 @ x,x in3d1,in3d0
-        vld2.32         {d20,d21},[r7,:128]!    @ in4u0,in4u1 x,x
+        vld2.32         {d16,d18},[r9,:128],r12 @ in0u0,in0u1 in4d1,in4d0
+        vld2.32         {d17,d19},[r8,:128],r12 @ in2u0,in2u1 in3d1,in3d0
+        vld2.32         {d0, d2}, [r7,:128]!    @ in4u0,in4u1 in2d1,in2d0
         vrev64.32       q9,  q9                 @ in4d0,in4d1 in3d0,in3d1
-        vld2.32         {d0, d1}, [r2,:128]!    @ in3u0,in3u1 x,x
-        vsub.f32        d20, d18, d20           @ in4d-in4u      I
-        vld1.32         {d2},     [r4,:64]!     @ c0,c1
-        vadd.f32        d0,  d0,  d19           @ in3u+in3d     -R
-        vld1.32         {d3},     [r5,:64]!     @ s0,s1
+        vld2.32         {d1, d3}, [r2,:128]!    @ in3u0,in3u1 in1d1,in1d0
+        vsub.f32        d0,  d18, d0            @ in4d-in4u      I
+        vld2.32         {d20,d21},[r4,:128]!    @ c0,c1 s0,s1
+        vrev64.32       q1,  q1                 @ in2d0,in2d1 in1d0,in1d1
+        vld2.32         {d30,d31},[r5,:128],r12 @ c2,c3 s2,s3
+        vadd.f32        d1,  d1,  d19           @ in3u+in3d     -R
+        vsub.f32        d16, d16, d2            @ in0u-in2d      R
+        vadd.f32        d17, d17, d3            @ in2u+in1d     -I
 1:
-        vmul.f32        d7,  d20, d3            @  I*s
-        vmul.f32        d6,  d0,  d2            @ -R*c
-        ldr             r6,  [r3], #4
-        vmul.f32        d4,  d0,  d3            @ -R*s
-        vmul.f32        d5,  d20, d2            @  I*c
+        vmul.f32        d7,  d0,  d21           @  I*s
+A       ldr             r10, [r3, lr, lsr #1]
+T       lsr             r10, lr,  #1
+T       ldr             r10, [r3, r10]
+        vmul.f32        d6,  d1,  d20           @ -R*c
+        ldr             r6,  [r3, #4]!
+        vmul.f32        d4,  d1,  d21           @ -R*s
+        vmul.f32        d5,  d0,  d20           @  I*c
+        vmul.f32        d24, d16, d30           @  R*c
+        vmul.f32        d25, d17, d31           @ -I*s
+        vmul.f32        d22, d16, d31           @  R*s
+        vmul.f32        d23, d17, d30           @  I*c
         subs            lr,  lr,  #16
         vsub.f32        d6,  d6,  d7            @ -R*c-I*s
         vadd.f32        d7,  d4,  d5            @ -R*s+I*c
-        uxth            r10, r6,  ror #16
-        uxth            r6,  r6
-        add             r10, r1,  r10, lsl #3
-        add             r6,  r1,  r6,  lsl #3
+        vsub.f32        d24, d25, d24           @ I*s-R*c
+        vadd.f32        d25, d22, d23           @ R*s-I*c
         beq             1f
-        vld2.32         {d16,d18},[r9,:128],r12 @ x,x in4d1,in4d0
-        vld2.32         {d17,d19},[r8,:128],r12 @ x,x in3d1,in3d0
+        mov             r12, #-16
+        vld2.32         {d16,d18},[r9,:128],r12 @ in0u0,in0u1 in4d1,in4d0
+        vld2.32         {d17,d19},[r8,:128],r12 @ in2u0,in2u1 in3d1,in3d0
         vneg.f32        d7,  d7                 @  R*s-I*c
-        vld2.32         {d20,d21},[r7,:128]!    @ in4u0,in4u1 x,x
+        vld2.32         {d0, d2}, [r7,:128]!    @ in4u0,in4u1 in2d1,in2d0
         vrev64.32       q9,  q9                 @ in4d0,in4d1 in3d0,in3d1
-        vld2.32         {d0, d1}, [r2,:128]!    @ in3u0,in3u1 x,x
-        vsub.f32        d20, d18, d20           @ in4d-in4u      I
-        vld1.32         {d2},     [r4,:64]!     @ c0,c1
-        vadd.f32        d0,  d0,  d19           @ in3u+in3d     -R
-        vld1.32         {d3},     [r5,:64]!     @ s0,s1
+        vld2.32         {d1, d3}, [r2,:128]!    @ in3u0,in3u1 in1d1,in1d0
+        vsub.f32        d0,  d18, d0            @ in4d-in4u      I
+        vld2.32         {d20,d21},[r4,:128]!    @ c0,c1 s0,s1
+        vrev64.32       q1,  q1                 @ in2d0,in2d1 in1d0,in1d1
+        vld2.32         {d30,d31},[r5,:128],r12 @ c2,c3 s2,s3
+        vadd.f32        d1,  d1,  d19           @ in3u+in3d     -R
+        vsub.f32        d16, d16, d2            @ in0u-in2d      R
+        vadd.f32        d17, d17, d3            @ in2u+in1d     -I
+        uxth            r12, r6,  ror #16
+        uxth            r6,  r6
+        add             r12, r1,  r12, lsl #3
+        add             r6,  r1,  r6,  lsl #3
         vst2.32         {d6[0],d7[0]}, [r6,:64]
-        vst2.32         {d6[1],d7[1]}, [r10,:64]
+        vst2.32         {d6[1],d7[1]}, [r12,:64]
+        uxth            r6,  r10, ror #16
+        uxth            r10, r10
+        add             r6 , r1,  r6,  lsl #3
+        add             r10, r1,  r10, lsl #3
+        vst2.32         {d24[0],d25[0]},[r10,:64]
+        vst2.32         {d24[1],d25[1]},[r6,:64]
         b               1b
 1:
         vneg.f32        d7,  d7                 @  R*s-I*c
-        vst2.32         {d6[0],d7[0]}, [r6,:64]
-        vst2.32         {d6[1],d7[1]}, [r10,:64]
-
-        mov             r12, #1
-        ldr             lr,  [r0, #28]          @ mdct_bits
-        lsl             lr,  r12, lr            @ n  = 1 << nbits
-        sub             r8,  r2,  #16           @ in1d
-        add             r2,  r9,  #16           @ in0u
-        sub             r9,  r7,  #16           @ in2d
-        mov             r12, #-16
-
-        vld2.32         {d16,d18},[r9,:128],r12 @ x,x in2d1,in2d0
-        vld2.32         {d17,d19},[r8,:128],r12 @ x,x in1d1,in1d0
-        vld2.32         {d20,d21},[r7,:128]!    @ in2u0,in2u1 x,x
-        vrev64.32       q9,  q9                 @ in2d0,in2d1 in1d0,in1d1
-        vld2.32         {d0, d1}, [r2,:128]!    @ in0u0,in0u1 x,x
-        vsub.f32        d0,  d0,  d18           @ in0u-in2d      R
-        vld1.32         {d2},     [r4,:64]!     @ c0,c1
-        vadd.f32        d20, d20, d19           @ in2u+in1d     -I
-        vld1.32         {d3},     [r5,:64]!     @ s0,s1
-1:
-        vmul.f32        d6,  d0,  d2            @  R*c
-        vmul.f32        d7,  d20, d3            @ -I*s
-        ldr             r6,  [r3], #4
-        vmul.f32        d4,  d0,  d3            @ R*s
-        vmul.f32        d5,  d20, d2            @ I*c
-        subs            lr,  lr,  #16
-        vsub.f32        d6,  d7,  d6            @ I*s-R*c
-        vadd.f32        d7,  d4,  d5            @ R*s-I*c
-        uxth            r10, r6,  ror #16
+        uxth            r12, r6,  ror #16
         uxth            r6,  r6
-        add             r10, r1,  r10, lsl #3
+        add             r12, r1,  r12, lsl #3
         add             r6,  r1,  r6,  lsl #3
-        beq             1f
-        vld2.32         {d16,d18},[r9,:128],r12 @ x,x in2d1,in2d0
-        vld2.32         {d17,d19},[r8,:128],r12 @ x,x in1d1,in1d0
-        vld2.32         {d20,d21},[r7,:128]!    @ in2u0,in2u1 x,x
-        vrev64.32       q9,  q9                 @ in2d0,in2d1 in1d0,in1d1
-        vld2.32         {d0, d1}, [r2,:128]!    @ in0u0,in0u1 x,x
-        vsub.f32        d0,  d0,  d18           @ in0u-in2d      R
-        vld1.32         {d2},     [r4,:64]!     @ c0,c1
-        vadd.f32        d20, d20, d19           @ in2u+in1d     -I
-        vld1.32         {d3},     [r5,:64]!     @ s0,s1
-        vst2.32         {d6[0],d7[0]}, [r6,:64]
-        vst2.32         {d6[1],d7[1]}, [r10,:64]
-        b               1b
-1:
         vst2.32         {d6[0],d7[0]}, [r6,:64]
-        vst2.32         {d6[1],d7[1]}, [r10,:64]
+        vst2.32         {d6[1],d7[1]}, [r12,:64]
+        uxth            r6,  r10, ror #16
+        uxth            r10, r10
+        add             r6 , r1,  r6,  lsl #3
+        add             r10, r1,  r10, lsl #3
+        vst2.32         {d24[0],d25[0]},[r10,:64]
+        vst2.32         {d24[1],d25[1]},[r6,:64]
 
         mov             r4,  r0
         mov             r6,  r1
         bl              ff_fft_calc_neon
 
         mov             r12, #1
-        ldr             lr,  [r4, #28]          @ mdct_bits
-        ldr             r5,  [r4, #36]          @ tsin
-        ldr             r4,  [r4, #32]          @ tcos
+        ldr             lr,  [r4, #20]          @ mdct_bits
+        ldr             r4,  [r4, #24]          @ tcos
         lsl             r12, r12, lr            @ n  = 1 << nbits
         lsr             lr,  r12, #3            @ n8 = n >> 3
 
-        add             r4,  r4,  lr,  lsl #2
-        add             r5,  r5,  lr,  lsl #2
+        add             r4,  r4,  lr,  lsl #3
         add             r6,  r6,  lr,  lsl #3
-        sub             r1,  r4,  #8
-        sub             r2,  r5,  #8
+        sub             r1,  r4,  #16
         sub             r3,  r6,  #16
 
         mov             r7,  #-16
-        mov             r12, #-8
         mov             r8,  r6
         mov             r0,  r3
 
         vld2.32         {d0-d1},  [r3,:128], r7 @ d0 =r1,i1 d1 =r0,i0
         vld2.32         {d20-d21},[r6,:128]!    @ d20=r2,i2 d21=r3,i3
-        vld1.32         {d18},    [r2,:64], r12 @ d18=s1,s0
+        vld2.32         {d16,d18},[r1,:128], r7 @ c1,c0 s1,s0
 1:
         subs            lr,  lr,  #2
         vmul.f32        d7,  d0,  d18           @ r1*s1,r0*s0
-        vld1.32         {d19},    [r5,:64]!     @ s2,s3
+        vld2.32         {d17,d19},[r4,:128]!    @ c2,c3 s2,s3
         vmul.f32        d4,  d1,  d18           @ i1*s1,i0*s0
-        vld1.32         {d16},    [r1,:64], r12 @ c1,c0
         vmul.f32        d5,  d21, d19           @ i2*s2,i3*s3
-        vld1.32         {d17},    [r4,:64]!     @ c2,c3
         vmul.f32        d6,  d20, d19           @ r2*s2,r3*s3
         vmul.f32        d24, d0,  d16           @ r1*c1,r0*c0
         vmul.f32        d25, d20, d17           @ r2*c2,r3*c3
@@ -316,7 +287,7 @@ function ff_mdct_calc_neon, export=1
         beq             1f
         vld2.32         {d0-d1},  [r3,:128], r7
         vld2.32         {d20-d21},[r6,:128]!
-        vld1.32         {d18},    [r2,:64], r12
+        vld2.32         {d16,d18},[r1,:128], r7 @ c1,c0 s1,s0
         vrev64.32       q3,  q3
         vst2.32         {d4,d6},  [r0,:128], r7
         vst2.32         {d5,d7},  [r8,:128]!
@@ -327,4 +298,4 @@ function ff_mdct_calc_neon, export=1
         vst2.32         {d5,d7},  [r8,:128]
 
         pop             {r4-r10,pc}
-.endfunc
+endfunc