]> git.sesse.net Git - ffmpeg/blobdiff - libavcodec/sh4/dsputil_align.c
cosmetics: Fix typos in ADPCM codec long names.
[ffmpeg] / libavcodec / sh4 / dsputil_align.c
index a808dd1f821dbbcac4170d93257835455acc097e..7c49a06ea915b514b39d74ea1e3275507021ec88 100644 (file)
 /*
- * aligned/packed access motion 
+ * aligned/packed access motion
  *
  * Copyright (c) 2001-2003 BERO <bero@geocities.co.jp>
  *
- * This library is free software; you can redistribute it and/or
+ * This file is part of FFmpeg.
+ *
+ * FFmpeg is free software; you can redistribute it and/or
  * modify it under the terms of the GNU Lesser General Public
  * License as published by the Free Software Foundation; either
- * version 2 of the License, or (at your option) any later version.
+ * version 2.1 of the License, or (at your option) any later version.
  *
- * This library is distributed in the hope that it will be useful,
+ * FFmpeg is distributed in the hope that it will be useful,
  * but WITHOUT ANY WARRANTY; without even the implied warranty of
  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
  * Lesser General Public License for more details.
  *
  * You should have received a copy of the GNU Lesser General Public
- * License along with this library; if not, write to the Free Software
- * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
+ * License along with FFmpeg; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  */
 
 
-#include "../avcodec.h"
-#include "../dsputil.h"
+#include "libavcodec/avcodec.h"
+#include "libavcodec/dsputil.h"
 
 
-#define        LP(p)   *(uint32_t*)(p)
+#define         LP(p)           *(uint32_t*)(p)
 
 
-#define        UNPACK(ph,pl,tt0,tt1) do { \
-       uint32_t t0,t1; t0=tt0;t1=tt1; \
-       ph = ( (t0 & ~BYTE_VEC32(0x03))>>2) + ( (t1 & ~BYTE_VEC32(0x03))>>2); \
-       pl = (t0 & BYTE_VEC32(0x03)) + (t1 & BYTE_VEC32(0x03)); } while(0)
+#define         UNPACK(ph,pl,tt0,tt1) do { \
+        uint32_t t0,t1; t0=tt0;t1=tt1; \
+        ph = ( (t0 & ~BYTE_VEC32(0x03))>>2) + ( (t1 & ~BYTE_VEC32(0x03))>>2); \
+        pl = (t0 & BYTE_VEC32(0x03)) + (t1 & BYTE_VEC32(0x03)); } while(0)
 
-#define        rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x02))>>2) & BYTE_VEC32(0x03))
-#define        no_rnd_PACK(ph,pl,nph,npl)      ph + nph + (((pl + npl + BYTE_VEC32(0x01))>>2) & BYTE_VEC32(0x03))
+#define         rnd_PACK(ph,pl,nph,npl) ph + nph + (((pl + npl + BYTE_VEC32(0x02))>>2) & BYTE_VEC32(0x03))
+#define         no_rnd_PACK(ph,pl,nph,npl)      ph + nph + (((pl + npl + BYTE_VEC32(0x01))>>2) & BYTE_VEC32(0x03))
 
 /* little endian */
-#define        MERGE1(a,b,ofs) (ofs==0)?a:( ((a)>>(8*ofs))|((b)<<(32-8*ofs)) )
-#define        MERGE2(a,b,ofs) (ofs==3)?b:( ((a)>>(8*(ofs+1)))|((b)<<(32-8*(ofs+1))) )
+#define         MERGE1(a,b,ofs) (ofs==0)?a:( ((a)>>(8*ofs))|((b)<<(32-8*ofs)) )
+#define         MERGE2(a,b,ofs) (ofs==3)?b:( ((a)>>(8*(ofs+1)))|((b)<<(32-8*(ofs+1))) )
 /* big
-#define        MERGE1(a,b,ofs) (ofs==0)?a:( ((a)<<(8*ofs))|((b)>>(32-8*ofs)) )
-#define        MERGE2(a,b,ofs) (ofs==3)?b:( ((a)<<(8+8*ofs))|((b)>>(32-8-8*ofs)) )
+#define         MERGE1(a,b,ofs) (ofs==0)?a:( ((a)<<(8*ofs))|((b)>>(32-8*ofs)) )
+#define         MERGE2(a,b,ofs) (ofs==3)?b:( ((a)<<(8+8*ofs))|((b)>>(32-8-8*ofs)) )
 */
 
 
-#define        put(d,s)        d = s
-#define        avg(d,s)        d = rnd_avg32(s,d)
+#define         put(d,s)        d = s
+#define         avg(d,s)        d = rnd_avg32(s,d)
 
-#define        OP_C4(ofs) \
-       ref-=ofs; \
-       do { \
-               OP(LP(dest),MERGE1(LP(ref),LP(ref+4),ofs)); \
-               ref+=stride; \
-               dest+=stride; \
-       } while(--height)
+#define         OP_C4(ofs) \
+        ref-=ofs; \
+        do { \
+                OP(LP(dest),MERGE1(LP(ref),LP(ref+4),ofs)); \
+                ref+=stride; \
+                dest+=stride; \
+        } while(--height)
 
-#define        OP_C40() \
-       do { \
-               OP(LP(dest),LP(ref)); \
-               ref+=stride; \
-               dest+=stride; \
-       } while(--height)
+#define        OP_C40() \
+        do { \
+                OP(LP(dest),LP(ref)); \
+                ref+=stride; \
+                dest+=stride; \
+        } while(--height)
 
 
-#define        OP      put
+#define         OP      put
 
 static void put_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height)
 {
-       switch((int)ref&3){
-       case 0: OP_C40(); return;
-       case 1: OP_C4(1); return;
-       case 2: OP_C4(2); return;
-       case 3: OP_C4(3); return;
-       }
+        switch((int)ref&3){
+        case 0: OP_C40(); return;
+        case 1: OP_C4(1); return;
+        case 2: OP_C4(2); return;
+        case 3: OP_C4(3); return;
+        }
 }
 
-#undef OP
-#define        OP      avg
+#undef          OP
+#define         OP      avg
 
 static void avg_pixels4_c(uint8_t *dest,const uint8_t *ref, const int stride,int height)
 {
-       switch((int)ref&3){
-       case 0: OP_C40(); return;
-       case 1: OP_C4(1); return;
-       case 2: OP_C4(2); return;
-       case 3: OP_C4(3); return;
-       }
+        switch((int)ref&3){
+        case 0: OP_C40(); return;
+        case 1: OP_C4(1); return;
+        case 2: OP_C4(2); return;
+        case 3: OP_C4(3); return;
+        }
 }
 
-#undef OP
+#undef          OP
 
-#define        OP_C(ofs,sz,avg2) \
+#define         OP_C(ofs,sz,avg2) \
 { \
-       ref-=ofs; \
-       do { \
-               uint32_t        t0,t1; \
-               t0 = LP(ref+0); \
-               t1 = LP(ref+4); \
-               OP(LP(dest+0), MERGE1(t0,t1,ofs)); \
-               t0 = LP(ref+8); \
-               OP(LP(dest+4), MERGE1(t1,t0,ofs)); \
+        ref-=ofs; \
+        do { \
+                uint32_t        t0,t1; \
+                t0 = LP(ref+0); \
+                t1 = LP(ref+4); \
+                OP(LP(dest+0), MERGE1(t0,t1,ofs)); \
+                t0 = LP(ref+8); \
+                OP(LP(dest+4), MERGE1(t1,t0,ofs)); \
 if (sz==16) { \
-               t1 = LP(ref+12); \
-               OP(LP(dest+8), MERGE1(t0,t1,ofs)); \
-               t0 = LP(ref+16); \
-               OP(LP(dest+12), MERGE1(t1,t0,ofs)); \
+                t1 = LP(ref+12); \
+                OP(LP(dest+8), MERGE1(t0,t1,ofs)); \
+                t0 = LP(ref+16); \
+                OP(LP(dest+12), MERGE1(t1,t0,ofs)); \
 } \
-               ref+=stride; \
-               dest+= stride; \
-       } while(--height); \
+                ref+=stride; \
+                dest+= stride; \
+        } while(--height); \
 }
 
 /* aligned */
-#define        OP_C0(sz,avg2) \
+#define         OP_C0(sz,avg2) \
 { \
-       do { \
-               OP(LP(dest+0), LP(ref+0)); \
-               OP(LP(dest+4), LP(ref+4)); \
+        do { \
+                OP(LP(dest+0), LP(ref+0)); \
+                OP(LP(dest+4), LP(ref+4)); \
 if (sz==16) { \
-               OP(LP(dest+8), LP(ref+8)); \
-               OP(LP(dest+12), LP(ref+12)); \
+                OP(LP(dest+8), LP(ref+8)); \
+                OP(LP(dest+12), LP(ref+12)); \
 } \
-               ref+=stride; \
-               dest+= stride; \
-       } while(--height); \
+                ref+=stride; \
+                dest+= stride; \
+        } while(--height); \
 }
 
-#define        OP_X(ofs,sz,avg2) \
+#define         OP_X(ofs,sz,avg2) \
 { \
-       ref-=ofs; \
-       do { \
-               uint32_t        t0,t1; \
-               t0 = LP(ref+0); \
-               t1 = LP(ref+4); \
-               OP(LP(dest+0), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
-               t0 = LP(ref+8); \
-               OP(LP(dest+4), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
+        ref-=ofs; \
+        do { \
+                uint32_t        t0,t1; \
+                t0 = LP(ref+0); \
+                t1 = LP(ref+4); \
+                OP(LP(dest+0), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
+                t0 = LP(ref+8); \
+                OP(LP(dest+4), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
 if (sz==16) { \
-               t1 = LP(ref+12); \
-               OP(LP(dest+8), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
-               t0 = LP(ref+16); \
-               OP(LP(dest+12), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
+                t1 = LP(ref+12); \
+                OP(LP(dest+8), avg2(MERGE1(t0,t1,ofs),MERGE2(t0,t1,ofs))); \
+                t0 = LP(ref+16); \
+                OP(LP(dest+12), avg2(MERGE1(t1,t0,ofs),MERGE2(t1,t0,ofs))); \
 } \
-               ref+=stride; \
-               dest+= stride; \
-       } while(--height); \
+                ref+=stride; \
+                dest+= stride; \
+        } while(--height); \
 }
 
 /* aligned */
-#define        OP_Y0(sz,avg2) \
+#define         OP_Y0(sz,avg2) \
 { \
-       uint32_t t0,t1,t2,t3,t; \
+        uint32_t t0,t1,t2,t3,t; \
 \
-       t0 = LP(ref+0); \
-       t1 = LP(ref+4); \
+        t0 = LP(ref+0); \
+        t1 = LP(ref+4); \
 if (sz==16) { \
-       t2 = LP(ref+8); \
-       t3 = LP(ref+12); \
+        t2 = LP(ref+8); \
+        t3 = LP(ref+12); \
 } \
-       do { \
-               ref += stride; \
+        do { \
+                ref += stride; \
 \
-               t = LP(ref+0); \
-               OP(LP(dest+0), avg2(t0,t)); t0 = t; \
-               t = LP(ref+4); \
-               OP(LP(dest+4), avg2(t1,t)); t1 = t; \
+                t = LP(ref+0); \
+                OP(LP(dest+0), avg2(t0,t)); t0 = t; \
+                t = LP(ref+4); \
+                OP(LP(dest+4), avg2(t1,t)); t1 = t; \
 if (sz==16) { \
-               t = LP(ref+8); \
-               OP(LP(dest+8), avg2(t2,t)); t2 = t; \
-               t = LP(ref+12); \
-               OP(LP(dest+12), avg2(t3,t)); t3 = t; \
+                t = LP(ref+8); \
+                OP(LP(dest+8), avg2(t2,t)); t2 = t; \
+                t = LP(ref+12); \
+                OP(LP(dest+12), avg2(t3,t)); t3 = t; \
 } \
-               dest+= stride; \
-       } while(--height); \
+                dest+= stride; \
+        } while(--height); \
 }
 
-#define        OP_Y(ofs,sz,avg2) \
+#define         OP_Y(ofs,sz,avg2) \
 { \
-       uint32_t t0,t1,t2,t3,t,w0,w1; \
+        uint32_t t0,t1,t2,t3,t,w0,w1; \
 \
-       ref-=ofs; \
-       w0 = LP(ref+0); \
-       w1 = LP(ref+4); \
-       t0 = MERGE1(w0,w1,ofs); \
-       w0 = LP(ref+8); \
-       t1 = MERGE1(w1,w0,ofs); \
+        ref-=ofs; \
+        w0 = LP(ref+0); \
+        w1 = LP(ref+4); \
+        t0 = MERGE1(w0,w1,ofs); \
+        w0 = LP(ref+8); \
+        t1 = MERGE1(w1,w0,ofs); \
 if (sz==16) { \
-       w1 = LP(ref+12); \
-       t2 = MERGE1(w0,w1,ofs); \
-       w0 = LP(ref+16); \
-       t3 = MERGE1(w1,w0,ofs); \
+        w1 = LP(ref+12); \
+        t2 = MERGE1(w0,w1,ofs); \
+        w0 = LP(ref+16); \
+        t3 = MERGE1(w1,w0,ofs); \
 } \
-       do { \
-               ref += stride; \
+        do { \
+                ref += stride; \
 \
-               w0 = LP(ref+0); \
-               w1 = LP(ref+4); \
-               t = MERGE1(w0,w1,ofs); \
-               OP(LP(dest+0), avg2(t0,t)); t0 = t; \
-               w0 = LP(ref+8); \
-               t = MERGE1(w1,w0,ofs); \
-               OP(LP(dest+4), avg2(t1,t)); t1 = t; \
+                w0 = LP(ref+0); \
+                w1 = LP(ref+4); \
+                t = MERGE1(w0,w1,ofs); \
+                OP(LP(dest+0), avg2(t0,t)); t0 = t; \
+                w0 = LP(ref+8); \
+                t = MERGE1(w1,w0,ofs); \
+                OP(LP(dest+4), avg2(t1,t)); t1 = t; \
 if (sz==16) { \
-               w1 = LP(ref+12); \
-               t = MERGE1(w0,w1,ofs); \
-               OP(LP(dest+8), avg2(t2,t)); t2 = t; \
-               w0 = LP(ref+16); \
-               t = MERGE1(w1,w0,ofs); \
-               OP(LP(dest+12), avg2(t3,t)); t3 = t; \
+                w1 = LP(ref+12); \
+                t = MERGE1(w0,w1,ofs); \
+                OP(LP(dest+8), avg2(t2,t)); t2 = t; \
+                w0 = LP(ref+16); \
+                t = MERGE1(w1,w0,ofs); \
+                OP(LP(dest+12), avg2(t3,t)); t3 = t; \
 } \
-               dest+=stride; \
-       } while(--height); \
+                dest+=stride; \
+        } while(--height); \
 }
 
 #define OP_X0(sz,avg2) OP_X(0,sz,avg2)
 #define OP_XY0(sz,PACK) OP_XY(0,sz,PACK)
-#define        OP_XY(ofs,sz,PACK) \
+#define         OP_XY(ofs,sz,PACK) \
 { \
-       uint32_t        t2,t3,w0,w1; \
-       uint32_t        a0,a1,a2,a3,a4,a5,a6,a7; \
+        uint32_t        t2,t3,w0,w1; \
+        uint32_t        a0,a1,a2,a3,a4,a5,a6,a7; \
 \
-       ref -= ofs; \
-       w0 = LP(ref+0); \
-       w1 = LP(ref+4); \
-       UNPACK(a0,a1,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
-       w0 = LP(ref+8); \
-       UNPACK(a2,a3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
+        ref -= ofs; \
+        w0 = LP(ref+0); \
+        w1 = LP(ref+4); \
+        UNPACK(a0,a1,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
+        w0 = LP(ref+8); \
+        UNPACK(a2,a3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
 if (sz==16) { \
-       w1 = LP(ref+12); \
-       UNPACK(a4,a5,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
-       w0 = LP(ref+16); \
-       UNPACK(a6,a7,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
+        w1 = LP(ref+12); \
+        UNPACK(a4,a5,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
+        w0 = LP(ref+16); \
+        UNPACK(a6,a7,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
 } \
-       do { \
-               ref+=stride; \
-               w0 = LP(ref+0); \
-               w1 = LP(ref+4); \
-               UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
-               OP(LP(dest+0),PACK(a0,a1,t2,t3)); \
-               a0 = t2; a1 = t3; \
-               w0 = LP(ref+8); \
-               UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
-               OP(LP(dest+4),PACK(a2,a3,t2,t3)); \
-               a2 = t2; a3 = t3; \
+        do { \
+                ref+=stride; \
+                w0 = LP(ref+0); \
+                w1 = LP(ref+4); \
+                UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
+                OP(LP(dest+0),PACK(a0,a1,t2,t3)); \
+                a0 = t2; a1 = t3; \
+                w0 = LP(ref+8); \
+                UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
+                OP(LP(dest+4),PACK(a2,a3,t2,t3)); \
+                a2 = t2; a3 = t3; \
 if (sz==16) { \
-               w1 = LP(ref+12); \
-               UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
-               OP(LP(dest+8),PACK(a4,a5,t2,t3)); \
-               a4 = t2; a5 = t3; \
-               w0 = LP(ref+16); \
-               UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
-               OP(LP(dest+12),PACK(a6,a7,t2,t3)); \
-               a6 = t2; a7 = t3; \
+                w1 = LP(ref+12); \
+                UNPACK(t2,t3,MERGE1(w0,w1,ofs),MERGE2(w0,w1,ofs)); \
+                OP(LP(dest+8),PACK(a4,a5,t2,t3)); \
+                a4 = t2; a5 = t3; \
+                w0 = LP(ref+16); \
+                UNPACK(t2,t3,MERGE1(w1,w0,ofs),MERGE2(w1,w0,ofs)); \
+                OP(LP(dest+12),PACK(a6,a7,t2,t3)); \
+                a6 = t2; a7 = t3; \
 } \
-               dest+=stride; \
-       } while(--height); \
+                dest+=stride; \
+        } while(--height); \
 }
 
-#define        DEFFUNC(op,rnd,xy,sz,OP_N,avgfunc) \
-static void op##_##rnd##_pixels##sz##_##xy (uint8_t * dest, const uint8_t * ref,       \
-                                  const int stride, int height)        \
+#define         DEFFUNC(op,rnd,xy,sz,OP_N,avgfunc) \
+static void op##_##rnd##_pixels##sz##_##xy (uint8_t * dest, const uint8_t * ref, \
+                                const int stride, int height) \
 { \
-       switch((int)ref&3) { \
-       case 0:OP_N##0(sz,rnd##_##avgfunc); return; \
-       case 1:OP_N(1,sz,rnd##_##avgfunc); return; \
-       case 2:OP_N(2,sz,rnd##_##avgfunc); return; \
-       case 3:OP_N(3,sz,rnd##_##avgfunc); return; \
-       } \
+        switch((int)ref&3) { \
+        case 0:OP_N##0(sz,rnd##_##avgfunc); return; \
+        case 1:OP_N(1,sz,rnd##_##avgfunc); return; \
+        case 2:OP_N(2,sz,rnd##_##avgfunc); return; \
+        case 3:OP_N(3,sz,rnd##_##avgfunc); return; \
+        } \
 }
 
 #define OP put
@@ -305,21 +307,21 @@ DEFFUNC(avg,no_rnd,xy,16,OP_XY,PACK)
 
 #undef OP
 
-#define        put_no_rnd_pixels8_o    put_rnd_pixels8_o
-#define        put_no_rnd_pixels16_o   put_rnd_pixels16_o
-#define        avg_no_rnd_pixels8_o    avg_rnd_pixels8_o
-#define        avg_no_rnd_pixels16_o   avg_rnd_pixels16_o
+#define         put_no_rnd_pixels8_o     put_rnd_pixels8_o
+#define         put_no_rnd_pixels16_o    put_rnd_pixels16_o
+#define         avg_no_rnd_pixels8_o     avg_rnd_pixels8_o
+#define         avg_no_rnd_pixels16_o    avg_rnd_pixels16_o
 
-#define        put_pixels8_c   put_rnd_pixels8_o
-#define        put_pixels16_c  put_rnd_pixels16_o
-#define        avg_pixels8_c   avg_rnd_pixels8_o
-#define        avg_pixels16_c  avg_rnd_pixels16_o
-#define        put_no_rnd_pixels8_c    put_rnd_pixels8_o
-#define        put_no_rnd_pixels16_c   put_rnd_pixels16_o
-#define        avg_no_rnd_pixels8_c    avg_rnd_pixels8_o
-#define        avg_no_rnd_pixels16_c   avg_rnd_pixels16_o
+#define         put_pixels8_c            put_rnd_pixels8_o
+#define         put_pixels16_c           put_rnd_pixels16_o
+#define         avg_pixels8_c            avg_rnd_pixels8_o
+#define         avg_pixels16_c           avg_rnd_pixels16_o
+#define         put_no_rnd_pixels8_c     put_rnd_pixels8_o
+#define         put_no_rnd_pixels16_c    put_rnd_pixels16_o
+#define         avg_no_rnd_pixels8_c     avg_rnd_pixels8_o
+#define         avg_no_rnd_pixels16_c    avg_rnd_pixels16_o
 
-#define        QPEL
+#define         QPEL
 
 #ifdef QPEL
 
@@ -329,41 +331,41 @@ DEFFUNC(avg,no_rnd,xy,16,OP_XY,PACK)
 
 void dsputil_init_align(DSPContext* c, AVCodecContext *avctx)
 {
-       c->put_pixels_tab[0][0] = put_rnd_pixels16_o;
-       c->put_pixels_tab[0][1] = put_rnd_pixels16_x;
-       c->put_pixels_tab[0][2] = put_rnd_pixels16_y;
-       c->put_pixels_tab[0][3] = put_rnd_pixels16_xy;
-       c->put_pixels_tab[1][0] = put_rnd_pixels8_o;
-       c->put_pixels_tab[1][1] = put_rnd_pixels8_x;
-       c->put_pixels_tab[1][2] = put_rnd_pixels8_y;
-       c->put_pixels_tab[1][3] = put_rnd_pixels8_xy;
-
-       c->put_no_rnd_pixels_tab[0][0] = put_no_rnd_pixels16_o;
-       c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x;
-       c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y;
-       c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy;
-       c->put_no_rnd_pixels_tab[1][0] = put_no_rnd_pixels8_o;
-       c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x;
-       c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y;
-       c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy;
-
-       c->avg_pixels_tab[0][0] = avg_rnd_pixels16_o;
-       c->avg_pixels_tab[0][1] = avg_rnd_pixels16_x;
-       c->avg_pixels_tab[0][2] = avg_rnd_pixels16_y;
-       c->avg_pixels_tab[0][3] = avg_rnd_pixels16_xy;
-       c->avg_pixels_tab[1][0] = avg_rnd_pixels8_o;
-       c->avg_pixels_tab[1][1] = avg_rnd_pixels8_x;
-       c->avg_pixels_tab[1][2] = avg_rnd_pixels8_y;
-       c->avg_pixels_tab[1][3] = avg_rnd_pixels8_xy;
-
-       c->avg_no_rnd_pixels_tab[0][0] = avg_no_rnd_pixels16_o;
-       c->avg_no_rnd_pixels_tab[0][1] = avg_no_rnd_pixels16_x;
-       c->avg_no_rnd_pixels_tab[0][2] = avg_no_rnd_pixels16_y;
-       c->avg_no_rnd_pixels_tab[0][3] = avg_no_rnd_pixels16_xy;
-       c->avg_no_rnd_pixels_tab[1][0] = avg_no_rnd_pixels8_o;
-       c->avg_no_rnd_pixels_tab[1][1] = avg_no_rnd_pixels8_x;
-       c->avg_no_rnd_pixels_tab[1][2] = avg_no_rnd_pixels8_y;
-       c->avg_no_rnd_pixels_tab[1][3] = avg_no_rnd_pixels8_xy;
+        c->put_pixels_tab[0][0] = put_rnd_pixels16_o;
+        c->put_pixels_tab[0][1] = put_rnd_pixels16_x;
+        c->put_pixels_tab[0][2] = put_rnd_pixels16_y;
+        c->put_pixels_tab[0][3] = put_rnd_pixels16_xy;
+        c->put_pixels_tab[1][0] = put_rnd_pixels8_o;
+        c->put_pixels_tab[1][1] = put_rnd_pixels8_x;
+        c->put_pixels_tab[1][2] = put_rnd_pixels8_y;
+        c->put_pixels_tab[1][3] = put_rnd_pixels8_xy;
+
+        c->put_no_rnd_pixels_tab[0][0] = put_no_rnd_pixels16_o;
+        c->put_no_rnd_pixels_tab[0][1] = put_no_rnd_pixels16_x;
+        c->put_no_rnd_pixels_tab[0][2] = put_no_rnd_pixels16_y;
+        c->put_no_rnd_pixels_tab[0][3] = put_no_rnd_pixels16_xy;
+        c->put_no_rnd_pixels_tab[1][0] = put_no_rnd_pixels8_o;
+        c->put_no_rnd_pixels_tab[1][1] = put_no_rnd_pixels8_x;
+        c->put_no_rnd_pixels_tab[1][2] = put_no_rnd_pixels8_y;
+        c->put_no_rnd_pixels_tab[1][3] = put_no_rnd_pixels8_xy;
+
+        c->avg_pixels_tab[0][0] = avg_rnd_pixels16_o;
+        c->avg_pixels_tab[0][1] = avg_rnd_pixels16_x;
+        c->avg_pixels_tab[0][2] = avg_rnd_pixels16_y;
+        c->avg_pixels_tab[0][3] = avg_rnd_pixels16_xy;
+        c->avg_pixels_tab[1][0] = avg_rnd_pixels8_o;
+        c->avg_pixels_tab[1][1] = avg_rnd_pixels8_x;
+        c->avg_pixels_tab[1][2] = avg_rnd_pixels8_y;
+        c->avg_pixels_tab[1][3] = avg_rnd_pixels8_xy;
+
+        c->avg_no_rnd_pixels_tab[0][0] = avg_no_rnd_pixels16_o;
+        c->avg_no_rnd_pixels_tab[0][1] = avg_no_rnd_pixels16_x;
+        c->avg_no_rnd_pixels_tab[0][2] = avg_no_rnd_pixels16_y;
+        c->avg_no_rnd_pixels_tab[0][3] = avg_no_rnd_pixels16_xy;
+        c->avg_no_rnd_pixels_tab[1][0] = avg_no_rnd_pixels8_o;
+        c->avg_no_rnd_pixels_tab[1][1] = avg_no_rnd_pixels8_x;
+        c->avg_no_rnd_pixels_tab[1][2] = avg_no_rnd_pixels8_y;
+        c->avg_no_rnd_pixels_tab[1][3] = avg_no_rnd_pixels8_xy;
 
 #ifdef QPEL