/*****************************************************************************
- * deblock.c: h264 encoder library
+ * deblock.c: deblocking
*****************************************************************************
- * Copyright (C) 2003-2008 x264 project
+ * Copyright (C) 2003-2015 x264 project
*
* Authors: Laurent Aimar <fenrir@via.ecp.fr>
* Loren Merritt <lorenm@u.washington.edu>
* Fiona Glaser <fiona@x264.com>
+ * Henrik Gramner <henrik@gramner.com>
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA.
+ *
+ * This program is also available under a commercial proprietary license.
+ * For more information, contact us at licensing@x264.com.
*****************************************************************************/
#include "common.h"
/* Deblocking filter */
-static const uint8_t i_alpha_table[52+12*2] =
+static const uint8_t i_alpha_table[52+12*3] =
{
+ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 4, 4, 5, 6,
255,255,
255,255,255,255,255,255,255,255,255,255,255,255,
};
-static const uint8_t i_beta_table[52+12*2] =
+static const uint8_t i_beta_table[52+12*3] =
{
+ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 2, 2, 2, 3,
18, 18,
18, 18, 18, 18, 18, 18, 18, 18, 18, 18, 18, 18,
};
-static const int8_t i_tc0_table[52+12*2][4] =
+static const int8_t i_tc0_table[52+12*3][4] =
{
{-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
{-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
{-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
{-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
+ {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
+ {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 },
{-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 0 }, {-1, 0, 0, 1 },
{-1, 0, 0, 1 }, {-1, 0, 0, 1 }, {-1, 0, 0, 1 }, {-1, 0, 1, 1 }, {-1, 0, 1, 1 }, {-1, 1, 1, 1 },
{-1, 1, 1, 1 }, {-1, 1, 1, 1 }, {-1, 1, 1, 1 }, {-1, 1, 1, 2 }, {-1, 1, 1, 2 }, {-1, 1, 1, 2 },
{-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 },
{-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 }, {-1,13,17,25 },
};
-#define alpha_table(x) i_alpha_table[(x)+12]
-#define beta_table(x) i_beta_table[(x)+12]
-#define tc0_table(x) i_tc0_table[(x)+12]
+#define alpha_table(x) i_alpha_table[(x)+24]
+#define beta_table(x) i_beta_table[(x)+24]
+#define tc0_table(x) i_tc0_table[(x)+24]
/* From ffmpeg */
-static inline void deblock_luma_c( uint8_t *pix, int xstride, int ystride, int alpha, int beta, int8_t *tc0 )
+static ALWAYS_INLINE void deblock_edge_luma_c( pixel *pix, intptr_t xstride, int alpha, int beta, int8_t tc0 )
+{
+ int p2 = pix[-3*xstride];
+ int p1 = pix[-2*xstride];
+ int p0 = pix[-1*xstride];
+ int q0 = pix[ 0*xstride];
+ int q1 = pix[ 1*xstride];
+ int q2 = pix[ 2*xstride];
+
+ if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
+ {
+ int tc = tc0;
+ int delta;
+ if( abs( p2 - p0 ) < beta )
+ {
+ if( tc0 )
+ pix[-2*xstride] = p1 + x264_clip3( (( p2 + ((p0 + q0 + 1) >> 1)) >> 1) - p1, -tc0, tc0 );
+ tc++;
+ }
+ if( abs( q2 - q0 ) < beta )
+ {
+ if( tc0 )
+ pix[ 1*xstride] = q1 + x264_clip3( (( q2 + ((p0 + q0 + 1) >> 1)) >> 1) - q1, -tc0, tc0 );
+ tc++;
+ }
+
+ delta = x264_clip3( (((q0 - p0 ) << 2) + (p1 - q1) + 4) >> 3, -tc, tc );
+ pix[-1*xstride] = x264_clip_pixel( p0 + delta ); /* p0' */
+ pix[ 0*xstride] = x264_clip_pixel( q0 - delta ); /* q0' */
+ }
+}
+static inline void deblock_luma_c( pixel *pix, intptr_t xstride, intptr_t ystride, int alpha, int beta, int8_t *tc0 )
{
for( int i = 0; i < 4; i++ )
{
pix += 4*ystride;
continue;
}
- for( int d = 0; d < 4; d++ )
- {
- int p2 = pix[-3*xstride];
- int p1 = pix[-2*xstride];
- int p0 = pix[-1*xstride];
- int q0 = pix[ 0*xstride];
- int q1 = pix[ 1*xstride];
- int q2 = pix[ 2*xstride];
-
- if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
- {
- int tc = tc0[i];
- int delta;
- if( abs( p2 - p0 ) < beta )
- {
- if( tc0[i] )
- pix[-2*xstride] = p1 + x264_clip3( (( p2 + ((p0 + q0 + 1) >> 1)) >> 1) - p1, -tc0[i], tc0[i] );
- tc++;
- }
- if( abs( q2 - q0 ) < beta )
- {
- if( tc0[i] )
- pix[ 1*xstride] = q1 + x264_clip3( (( q2 + ((p0 + q0 + 1) >> 1)) >> 1) - q1, -tc0[i], tc0[i] );
- tc++;
- }
-
- delta = x264_clip3( (((q0 - p0 ) << 2) + (p1 - q1) + 4) >> 3, -tc, tc );
- pix[-1*xstride] = x264_clip_uint8( p0 + delta ); /* p0' */
- pix[ 0*xstride] = x264_clip_uint8( q0 - delta ); /* q0' */
- }
- pix += ystride;
- }
+ for( int d = 0; d < 4; d++, pix += ystride )
+ deblock_edge_luma_c( pix, xstride, alpha, beta, tc0[i] );
}
}
-static void deblock_v_luma_c( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 )
+static void deblock_h_luma_mbaff_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
+{
+ for( int d = 0; d < 8; d++, pix += stride )
+ deblock_edge_luma_c( pix, 1, alpha, beta, tc0[d>>1] );
+}
+static void deblock_v_luma_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
{
deblock_luma_c( pix, stride, 1, alpha, beta, tc0 );
}
-static void deblock_h_luma_c( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 )
+static void deblock_h_luma_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
{
deblock_luma_c( pix, 1, stride, alpha, beta, tc0 );
}
-static inline void deblock_chroma_c( uint8_t *pix, int xstride, int ystride, int alpha, int beta, int8_t *tc0 )
+static ALWAYS_INLINE void deblock_edge_chroma_c( pixel *pix, intptr_t xstride, int alpha, int beta, int8_t tc )
+{
+ int p1 = pix[-2*xstride];
+ int p0 = pix[-1*xstride];
+ int q0 = pix[ 0*xstride];
+ int q1 = pix[ 1*xstride];
+
+ if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
+ {
+ int delta = x264_clip3( (((q0 - p0 ) << 2) + (p1 - q1) + 4) >> 3, -tc, tc );
+ pix[-1*xstride] = x264_clip_pixel( p0 + delta ); /* p0' */
+ pix[ 0*xstride] = x264_clip_pixel( q0 - delta ); /* q0' */
+ }
+}
+static ALWAYS_INLINE void deblock_chroma_c( pixel *pix, int height, intptr_t xstride, intptr_t ystride, int alpha, int beta, int8_t *tc0 )
{
for( int i = 0; i < 4; i++ )
{
int tc = tc0[i];
if( tc <= 0 )
{
- pix += 2*ystride;
+ pix += height*ystride;
continue;
}
- for( int d = 0; d < 2; d++ )
- {
- int p1 = pix[-2*xstride];
- int p0 = pix[-1*xstride];
- int q0 = pix[ 0*xstride];
- int q1 = pix[ 1*xstride];
-
- if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
- {
- int delta = x264_clip3( (((q0 - p0 ) << 2) + (p1 - q1) + 4) >> 3, -tc, tc );
- pix[-1*xstride] = x264_clip_uint8( p0 + delta ); /* p0' */
- pix[ 0*xstride] = x264_clip_uint8( q0 - delta ); /* q0' */
- }
- pix += ystride;
- }
+ for( int d = 0; d < height; d++, pix += ystride-2 )
+ for( int e = 0; e < 2; e++, pix++ )
+ deblock_edge_chroma_c( pix, xstride, alpha, beta, tc0[i] );
}
}
-static void deblock_v_chroma_c( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 )
+static void deblock_h_chroma_mbaff_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
{
- deblock_chroma_c( pix, stride, 1, alpha, beta, tc0 );
+ deblock_chroma_c( pix, 1, 2, stride, alpha, beta, tc0 );
}
-static void deblock_h_chroma_c( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 )
+static void deblock_v_chroma_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
{
- deblock_chroma_c( pix, 1, stride, alpha, beta, tc0 );
+ deblock_chroma_c( pix, 2, stride, 2, alpha, beta, tc0 );
+}
+static void deblock_h_chroma_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
+{
+ deblock_chroma_c( pix, 2, 2, stride, alpha, beta, tc0 );
+}
+static void deblock_h_chroma_422_c( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
+{
+ deblock_chroma_c( pix, 4, 2, stride, alpha, beta, tc0 );
}
-static inline void deblock_luma_intra_c( uint8_t *pix, int xstride, int ystride, int alpha, int beta )
+static ALWAYS_INLINE void deblock_edge_luma_intra_c( pixel *pix, intptr_t xstride, int alpha, int beta )
{
- for( int d = 0; d < 16; d++ )
+ int p2 = pix[-3*xstride];
+ int p1 = pix[-2*xstride];
+ int p0 = pix[-1*xstride];
+ int q0 = pix[ 0*xstride];
+ int q1 = pix[ 1*xstride];
+ int q2 = pix[ 2*xstride];
+
+ if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
{
- int p2 = pix[-3*xstride];
- int p1 = pix[-2*xstride];
- int p0 = pix[-1*xstride];
- int q0 = pix[ 0*xstride];
- int q1 = pix[ 1*xstride];
- int q2 = pix[ 2*xstride];
-
- if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
+ if( abs( p0 - q0 ) < ((alpha >> 2) + 2) )
{
- if(abs( p0 - q0 ) < ((alpha >> 2) + 2) )
+ if( abs( p2 - p0 ) < beta ) /* p0', p1', p2' */
{
- if( abs( p2 - p0 ) < beta ) /* p0', p1', p2' */
- {
- const int p3 = pix[-4*xstride];
- pix[-1*xstride] = ( p2 + 2*p1 + 2*p0 + 2*q0 + q1 + 4 ) >> 3;
- pix[-2*xstride] = ( p2 + p1 + p0 + q0 + 2 ) >> 2;
- pix[-3*xstride] = ( 2*p3 + 3*p2 + p1 + p0 + q0 + 4 ) >> 3;
- }
- else /* p0' */
- pix[-1*xstride] = ( 2*p1 + p0 + q1 + 2 ) >> 2;
- if( abs( q2 - q0 ) < beta ) /* q0', q1', q2' */
- {
- const int q3 = pix[3*xstride];
- pix[0*xstride] = ( p1 + 2*p0 + 2*q0 + 2*q1 + q2 + 4 ) >> 3;
- pix[1*xstride] = ( p0 + q0 + q1 + q2 + 2 ) >> 2;
- pix[2*xstride] = ( 2*q3 + 3*q2 + q1 + q0 + p0 + 4 ) >> 3;
- }
- else /* q0' */
- pix[0*xstride] = ( 2*q1 + q0 + p1 + 2 ) >> 2;
+ const int p3 = pix[-4*xstride];
+ pix[-1*xstride] = ( p2 + 2*p1 + 2*p0 + 2*q0 + q1 + 4 ) >> 3;
+ pix[-2*xstride] = ( p2 + p1 + p0 + q0 + 2 ) >> 2;
+ pix[-3*xstride] = ( 2*p3 + 3*p2 + p1 + p0 + q0 + 4 ) >> 3;
}
- else /* p0', q0' */
- {
+ else /* p0' */
pix[-1*xstride] = ( 2*p1 + p0 + q1 + 2 ) >> 2;
- pix[ 0*xstride] = ( 2*q1 + q0 + p1 + 2 ) >> 2;
+ if( abs( q2 - q0 ) < beta ) /* q0', q1', q2' */
+ {
+ const int q3 = pix[3*xstride];
+ pix[0*xstride] = ( p1 + 2*p0 + 2*q0 + 2*q1 + q2 + 4 ) >> 3;
+ pix[1*xstride] = ( p0 + q0 + q1 + q2 + 2 ) >> 2;
+ pix[2*xstride] = ( 2*q3 + 3*q2 + q1 + q0 + p0 + 4 ) >> 3;
}
+ else /* q0' */
+ pix[0*xstride] = ( 2*q1 + q0 + p1 + 2 ) >> 2;
+ }
+ else /* p0', q0' */
+ {
+ pix[-1*xstride] = ( 2*p1 + p0 + q1 + 2 ) >> 2;
+ pix[ 0*xstride] = ( 2*q1 + q0 + p1 + 2 ) >> 2;
}
- pix += ystride;
}
}
-static void deblock_v_luma_intra_c( uint8_t *pix, int stride, int alpha, int beta )
+static inline void deblock_luma_intra_c( pixel *pix, intptr_t xstride, intptr_t ystride, int alpha, int beta )
+{
+ for( int d = 0; d < 16; d++, pix += ystride )
+ deblock_edge_luma_intra_c( pix, xstride, alpha, beta );
+}
+static void deblock_h_luma_intra_mbaff_c( pixel *pix, intptr_t ystride, int alpha, int beta )
+{
+ for( int d = 0; d < 8; d++, pix += ystride )
+ deblock_edge_luma_intra_c( pix, 1, alpha, beta );
+}
+static void deblock_v_luma_intra_c( pixel *pix, intptr_t stride, int alpha, int beta )
{
deblock_luma_intra_c( pix, stride, 1, alpha, beta );
}
-static void deblock_h_luma_intra_c( uint8_t *pix, int stride, int alpha, int beta )
+static void deblock_h_luma_intra_c( pixel *pix, intptr_t stride, int alpha, int beta )
{
deblock_luma_intra_c( pix, 1, stride, alpha, beta );
}
-static inline void deblock_chroma_intra_c( uint8_t *pix, int xstride, int ystride, int alpha, int beta )
+static ALWAYS_INLINE void deblock_edge_chroma_intra_c( pixel *pix, intptr_t xstride, int alpha, int beta )
{
- for( int d = 0; d < 8; d++ )
- {
- int p1 = pix[-2*xstride];
- int p0 = pix[-1*xstride];
- int q0 = pix[ 0*xstride];
- int q1 = pix[ 1*xstride];
+ int p1 = pix[-2*xstride];
+ int p0 = pix[-1*xstride];
+ int q0 = pix[ 0*xstride];
+ int q1 = pix[ 1*xstride];
- if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
- {
- pix[-1*xstride] = (2*p1 + p0 + q1 + 2) >> 2; /* p0' */
- pix[ 0*xstride] = (2*q1 + q0 + p1 + 2) >> 2; /* q0' */
- }
- pix += ystride;
+ if( abs( p0 - q0 ) < alpha && abs( p1 - p0 ) < beta && abs( q1 - q0 ) < beta )
+ {
+ pix[-1*xstride] = (2*p1 + p0 + q1 + 2) >> 2; /* p0' */
+ pix[ 0*xstride] = (2*q1 + q0 + p1 + 2) >> 2; /* q0' */
}
}
-static void deblock_v_chroma_intra_c( uint8_t *pix, int stride, int alpha, int beta )
+static ALWAYS_INLINE void deblock_chroma_intra_c( pixel *pix, int width, int height, intptr_t xstride, intptr_t ystride, int alpha, int beta )
+{
+ for( int d = 0; d < height; d++, pix += ystride-2 )
+ for( int e = 0; e < width; e++, pix++ )
+ deblock_edge_chroma_intra_c( pix, xstride, alpha, beta );
+}
+static void deblock_h_chroma_intra_mbaff_c( pixel *pix, intptr_t stride, int alpha, int beta )
+{
+ deblock_chroma_intra_c( pix, 2, 4, 2, stride, alpha, beta );
+}
+static void deblock_v_chroma_intra_c( pixel *pix, intptr_t stride, int alpha, int beta )
+{
+ deblock_chroma_intra_c( pix, 1, 16, stride, 2, alpha, beta );
+}
+static void deblock_h_chroma_intra_c( pixel *pix, intptr_t stride, int alpha, int beta )
{
- deblock_chroma_intra_c( pix, stride, 1, alpha, beta );
+ deblock_chroma_intra_c( pix, 2, 8, 2, stride, alpha, beta );
}
-static void deblock_h_chroma_intra_c( uint8_t *pix, int stride, int alpha, int beta )
+static void deblock_h_chroma_422_intra_c( pixel *pix, intptr_t stride, int alpha, int beta )
{
- deblock_chroma_intra_c( pix, 1, stride, alpha, beta );
+ deblock_chroma_intra_c( pix, 2, 16, 2, stride, alpha, beta );
}
static void deblock_strength_c( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
- int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][4][4], int mvy_limit,
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4], int mvy_limit,
int bframe )
{
for( int dir = 0; dir < 2; dir++ )
}
}
-static inline void deblock_edge( x264_t *h, uint8_t *pix1, uint8_t *pix2, int i_stride, uint8_t bS[4], int i_qp, int b_chroma, x264_deblock_inter_t pf_inter )
+static ALWAYS_INLINE void deblock_edge( x264_t *h, pixel *pix, intptr_t i_stride, uint8_t bS[4], int i_qp,
+ int a, int b, int b_chroma, x264_deblock_inter_t pf_inter )
{
- int index_a = i_qp + h->sh.i_alpha_c0_offset;
- int alpha = alpha_table(index_a);
- int beta = beta_table(i_qp + h->sh.i_beta_offset);
+ int index_a = i_qp + a;
+ int index_b = i_qp + b;
+ int alpha = alpha_table(index_a) << (BIT_DEPTH-8);
+ int beta = beta_table(index_b) << (BIT_DEPTH-8);
int8_t tc[4];
if( !M32(bS) || !alpha || !beta )
return;
- tc[0] = tc0_table(index_a)[bS[0]] + b_chroma;
- tc[1] = tc0_table(index_a)[bS[1]] + b_chroma;
- tc[2] = tc0_table(index_a)[bS[2]] + b_chroma;
- tc[3] = tc0_table(index_a)[bS[3]] + b_chroma;
+ tc[0] = (tc0_table(index_a)[bS[0]] << (BIT_DEPTH-8)) + b_chroma;
+ tc[1] = (tc0_table(index_a)[bS[1]] << (BIT_DEPTH-8)) + b_chroma;
+ tc[2] = (tc0_table(index_a)[bS[2]] << (BIT_DEPTH-8)) + b_chroma;
+ tc[3] = (tc0_table(index_a)[bS[3]] << (BIT_DEPTH-8)) + b_chroma;
- pf_inter( pix1, i_stride, alpha, beta, tc );
- if( b_chroma )
- pf_inter( pix2, i_stride, alpha, beta, tc );
+ pf_inter( pix, i_stride, alpha, beta, tc );
}
-static inline void deblock_edge_intra( x264_t *h, uint8_t *pix1, uint8_t *pix2, int i_stride, uint8_t bS[4], int i_qp, int b_chroma, x264_deblock_intra_t pf_intra )
+static ALWAYS_INLINE void deblock_edge_intra( x264_t *h, pixel *pix, intptr_t i_stride, uint8_t bS[4], int i_qp,
+ int a, int b, int b_chroma, x264_deblock_intra_t pf_intra )
{
- int alpha = alpha_table(i_qp + h->sh.i_alpha_c0_offset);
- int beta = beta_table(i_qp + h->sh.i_beta_offset);
+ int index_a = i_qp + a;
+ int index_b = i_qp + b;
+ int alpha = alpha_table(index_a) << (BIT_DEPTH-8);
+ int beta = beta_table(index_b) << (BIT_DEPTH-8);
if( !alpha || !beta )
return;
- pf_intra( pix1, i_stride, alpha, beta );
- if( b_chroma )
- pf_intra( pix2, i_stride, alpha, beta );
+ pf_intra( pix, i_stride, alpha, beta );
+}
+
+static ALWAYS_INLINE void x264_macroblock_cache_load_neighbours_deblock( x264_t *h, int mb_x, int mb_y )
+{
+ int deblock_on_slice_edges = h->sh.i_disable_deblocking_filter_idc != 2;
+
+ h->mb.i_neighbour = 0;
+ h->mb.i_mb_xy = mb_y * h->mb.i_mb_stride + mb_x;
+ h->mb.b_interlaced = PARAM_INTERLACED && h->mb.field[h->mb.i_mb_xy];
+ h->mb.i_mb_top_y = mb_y - (1 << MB_INTERLACED);
+ h->mb.i_mb_top_xy = mb_x + h->mb.i_mb_stride*h->mb.i_mb_top_y;
+ h->mb.i_mb_left_xy[1] =
+ h->mb.i_mb_left_xy[0] = h->mb.i_mb_xy - 1;
+ if( SLICE_MBAFF )
+ {
+ if( mb_y&1 )
+ {
+ if( mb_x && h->mb.field[h->mb.i_mb_xy - 1] != MB_INTERLACED )
+ h->mb.i_mb_left_xy[0] -= h->mb.i_mb_stride;
+ }
+ else
+ {
+ if( h->mb.i_mb_top_xy >= 0 && MB_INTERLACED && !h->mb.field[h->mb.i_mb_top_xy] )
+ {
+ h->mb.i_mb_top_xy += h->mb.i_mb_stride;
+ h->mb.i_mb_top_y++;
+ }
+ if( mb_x && h->mb.field[h->mb.i_mb_xy - 1] != MB_INTERLACED )
+ h->mb.i_mb_left_xy[1] += h->mb.i_mb_stride;
+ }
+ }
+
+ if( mb_x > 0 && (deblock_on_slice_edges ||
+ h->mb.slice_table[h->mb.i_mb_left_xy[0]] == h->mb.slice_table[h->mb.i_mb_xy]) )
+ h->mb.i_neighbour |= MB_LEFT;
+ if( mb_y > MB_INTERLACED && (deblock_on_slice_edges
+ || h->mb.slice_table[h->mb.i_mb_top_xy] == h->mb.slice_table[h->mb.i_mb_xy]) )
+ h->mb.i_neighbour |= MB_TOP;
}
void x264_frame_deblock_row( x264_t *h, int mb_y )
{
- int b_interlaced = h->sh.b_mbaff;
- int qp_thresh = 15 - X264_MIN(h->sh.i_alpha_c0_offset, h->sh.i_beta_offset) - X264_MAX(0, h->param.analyse.i_chroma_qp_offset);
+ int b_interlaced = SLICE_MBAFF;
+ int a = h->sh.i_alpha_c0_offset - QP_BD_OFFSET;
+ int b = h->sh.i_beta_offset - QP_BD_OFFSET;
+ int qp_thresh = 15 - X264_MIN( a, b ) - X264_MAX( 0, h->pps->i_chroma_qp_index_offset );
int stridey = h->fdec->i_stride[0];
- int stride2y = stridey << b_interlaced;
int strideuv = h->fdec->i_stride[1];
- int stride2uv = strideuv << b_interlaced;
+ int chroma444 = CHROMA444;
+ int chroma_height = 16 >> CHROMA_V_SHIFT;
+ intptr_t uvdiff = chroma444 ? h->fdec->plane[2] - h->fdec->plane[1] : 1;
- for( int mb_x = 0; mb_x < h->sps->i_mb_width; mb_x += (~b_interlaced | mb_y)&1, mb_y ^= b_interlaced )
+ for( int mb_x = 0; mb_x < h->mb.i_mb_width; mb_x += (~b_interlaced | mb_y)&1, mb_y ^= b_interlaced )
{
x264_prefetch_fenc( h, h->fdec, mb_x, mb_y );
x264_macroblock_cache_load_neighbours_deblock( h, mb_x, mb_y );
int mb_xy = h->mb.i_mb_xy;
- int transform_8x8 = h->mb.mb_transform_size[h->mb.i_mb_xy];
+ int transform_8x8 = h->mb.mb_transform_size[mb_xy];
int intra_cur = IS_INTRA( h->mb.type[mb_xy] );
- uint8_t (*bs)[4][4] = h->deblock_strength[mb_y&b_interlaced][mb_x];
+ uint8_t (*bs)[8][4] = h->deblock_strength[mb_y&1][h->param.b_sliced_threads?mb_xy:mb_x];
+
+ pixel *pixy = h->fdec->plane[0] + 16*mb_y*stridey + 16*mb_x;
+ pixel *pixuv = h->fdec->plane[1] + chroma_height*mb_y*strideuv + 16*mb_x;
- uint8_t *pixy = h->fdec->plane[0] + 16*mb_y*stridey + 16*mb_x;
- uint8_t *pixu = h->fdec->plane[1] + 8*mb_y*strideuv + 8*mb_x;
- uint8_t *pixv = h->fdec->plane[2] + 8*mb_y*strideuv + 8*mb_x;
- if( b_interlaced && (mb_y&1) )
+ if( mb_y & MB_INTERLACED )
{
pixy -= 15*stridey;
- pixu -= 7*strideuv;
- pixv -= 7*strideuv;
+ pixuv -= (chroma_height-1)*strideuv;
}
+ int stride2y = stridey << MB_INTERLACED;
+ int stride2uv = strideuv << MB_INTERLACED;
int qp = h->mb.qp[mb_xy];
int qpc = h->chroma_qp_table[qp];
- int first_edge_only = h->mb.type[mb_xy] == P_SKIP || qp <= qp_thresh;
+ int first_edge_only = (h->mb.partition[mb_xy] == D_16x16 && !h->mb.cbp[mb_xy] && !intra_cur) || qp <= qp_thresh;
#define FILTER( intra, dir, edge, qp, chroma_qp )\
do\
{\
- deblock_edge##intra( h, pixy + 4*edge*(dir?stride2y:1), NULL,\
- stride2y, bs[dir][edge], qp, 0,\
- h->loopf.deblock_luma##intra[dir] );\
- if( !(edge & 1) )\
- deblock_edge##intra( h, pixu + 2*edge*(dir?stride2uv:1), pixv + 2*edge*(dir?stride2uv:1),\
- stride2uv, bs[dir][edge], chroma_qp, 1,\
+ if( !(edge & 1) || !transform_8x8 )\
+ {\
+ deblock_edge##intra( h, pixy + 4*edge*(dir?stride2y:1),\
+ stride2y, bs[dir][edge], qp, a, b, 0,\
+ h->loopf.deblock_luma##intra[dir] );\
+ if( CHROMA_FORMAT == CHROMA_444 )\
+ {\
+ deblock_edge##intra( h, pixuv + 4*edge*(dir?stride2uv:1),\
+ stride2uv, bs[dir][edge], chroma_qp, a, b, 0,\
+ h->loopf.deblock_luma##intra[dir] );\
+ deblock_edge##intra( h, pixuv + uvdiff + 4*edge*(dir?stride2uv:1),\
+ stride2uv, bs[dir][edge], chroma_qp, a, b, 0,\
+ h->loopf.deblock_luma##intra[dir] );\
+ }\
+ else if( CHROMA_FORMAT == CHROMA_420 && !(edge & 1) )\
+ {\
+ deblock_edge##intra( h, pixuv + edge*(dir?2*stride2uv:4),\
+ stride2uv, bs[dir][edge], chroma_qp, a, b, 1,\
+ h->loopf.deblock_chroma##intra[dir] );\
+ }\
+ }\
+ if( CHROMA_FORMAT == CHROMA_422 && (dir || !(edge & 1)) )\
+ {\
+ deblock_edge##intra( h, pixuv + edge*(dir?4*stride2uv:4),\
+ stride2uv, bs[dir][edge], chroma_qp, a, b, 1,\
h->loopf.deblock_chroma##intra[dir] );\
+ }\
} while(0)
if( h->mb.i_neighbour & MB_LEFT )
{
- int qpl = h->mb.qp[h->mb.i_mb_left_xy];
- int qp_left = (qp + qpl + 1) >> 1;
- int qpc_left = (h->chroma_qp_table[qp] + h->chroma_qp_table[qpl] + 1) >> 1;
- int intra_left = IS_INTRA( h->mb.type[h->mb.i_mb_left_xy] );
- if( intra_cur || intra_left )
- FILTER( _intra, 0, 0, qp_left, qpc_left );
+ if( b_interlaced && h->mb.field[h->mb.i_mb_left_xy[0]] != MB_INTERLACED )
+ {
+ int luma_qp[2];
+ int chroma_qp[2];
+ int left_qp[2];
+ x264_deblock_inter_t luma_deblock = h->loopf.deblock_luma_mbaff;
+ x264_deblock_inter_t chroma_deblock = h->loopf.deblock_chroma_mbaff;
+ x264_deblock_intra_t luma_intra_deblock = h->loopf.deblock_luma_intra_mbaff;
+ x264_deblock_intra_t chroma_intra_deblock = h->loopf.deblock_chroma_intra_mbaff;
+ int c = chroma444 ? 0 : 1;
+
+ left_qp[0] = h->mb.qp[h->mb.i_mb_left_xy[0]];
+ luma_qp[0] = (qp + left_qp[0] + 1) >> 1;
+ chroma_qp[0] = (qpc + h->chroma_qp_table[left_qp[0]] + 1) >> 1;
+ if( intra_cur || IS_INTRA( h->mb.type[h->mb.i_mb_left_xy[0]] ) )
+ {
+ deblock_edge_intra( h, pixy, 2*stridey, bs[0][0], luma_qp[0], a, b, 0, luma_intra_deblock );
+ deblock_edge_intra( h, pixuv, 2*strideuv, bs[0][0], chroma_qp[0], a, b, c, chroma_intra_deblock );
+ if( chroma444 )
+ deblock_edge_intra( h, pixuv + uvdiff, 2*strideuv, bs[0][0], chroma_qp[0], a, b, c, chroma_intra_deblock );
+ }
+ else
+ {
+ deblock_edge( h, pixy, 2*stridey, bs[0][0], luma_qp[0], a, b, 0, luma_deblock );
+ deblock_edge( h, pixuv, 2*strideuv, bs[0][0], chroma_qp[0], a, b, c, chroma_deblock );
+ if( chroma444 )
+ deblock_edge( h, pixuv + uvdiff, 2*strideuv, bs[0][0], chroma_qp[0], a, b, c, chroma_deblock );
+ }
+
+ int offy = MB_INTERLACED ? 4 : 0;
+ int offuv = MB_INTERLACED ? 4-CHROMA_V_SHIFT : 0;
+ left_qp[1] = h->mb.qp[h->mb.i_mb_left_xy[1]];
+ luma_qp[1] = (qp + left_qp[1] + 1) >> 1;
+ chroma_qp[1] = (qpc + h->chroma_qp_table[left_qp[1]] + 1) >> 1;
+ if( intra_cur || IS_INTRA( h->mb.type[h->mb.i_mb_left_xy[1]] ) )
+ {
+ deblock_edge_intra( h, pixy + (stridey<<offy), 2*stridey, bs[0][4], luma_qp[1], a, b, 0, luma_intra_deblock );
+ deblock_edge_intra( h, pixuv + (strideuv<<offuv), 2*strideuv, bs[0][4], chroma_qp[1], a, b, c, chroma_intra_deblock );
+ if( chroma444 )
+ deblock_edge_intra( h, pixuv + uvdiff + (strideuv<<offuv), 2*strideuv, bs[0][4], chroma_qp[1], a, b, c, chroma_intra_deblock );
+ }
+ else
+ {
+ deblock_edge( h, pixy + (stridey<<offy), 2*stridey, bs[0][4], luma_qp[1], a, b, 0, luma_deblock );
+ deblock_edge( h, pixuv + (strideuv<<offuv), 2*strideuv, bs[0][4], chroma_qp[1], a, b, c, chroma_deblock );
+ if( chroma444 )
+ deblock_edge( h, pixuv + uvdiff + (strideuv<<offuv), 2*strideuv, bs[0][4], chroma_qp[1], a, b, c, chroma_deblock );
+ }
+ }
else
- FILTER( , 0, 0, qp_left, qpc_left );
- }
+ {
+ int qpl = h->mb.qp[h->mb.i_mb_xy-1];
+ int qp_left = (qp + qpl + 1) >> 1;
+ int qpc_left = (qpc + h->chroma_qp_table[qpl] + 1) >> 1;
+ int intra_left = IS_INTRA( h->mb.type[h->mb.i_mb_xy-1] );
+ int intra_deblock = intra_cur || intra_left;
+
+ /* Any MB that was coded, or that analysis decided to skip, has quality commensurate with its QP.
+ * But if deblocking affects neighboring MBs that were force-skipped, blur might accumulate there.
+ * So reset their effective QP to max, to indicate that lack of guarantee. */
+ if( h->fdec->mb_info && M32( bs[0][0] ) )
+ {
+#define RESET_EFFECTIVE_QP(xy) h->fdec->effective_qp[xy] |= 0xff * !!(h->fdec->mb_info[xy] & X264_MBINFO_CONSTANT);
+ RESET_EFFECTIVE_QP(mb_xy);
+ RESET_EFFECTIVE_QP(h->mb.i_mb_left_xy[0]);
+ }
+ if( intra_deblock )
+ FILTER( _intra, 0, 0, qp_left, qpc_left );
+ else
+ FILTER( , 0, 0, qp_left, qpc_left );
+ }
+ }
if( !first_edge_only )
{
- if( !transform_8x8 ) FILTER( , 0, 1, qp, qpc );
- FILTER( , 0, 2, qp, qpc );
- if( !transform_8x8 ) FILTER( , 0, 3, qp, qpc );
+ FILTER( , 0, 1, qp, qpc );
+ FILTER( , 0, 2, qp, qpc );
+ FILTER( , 0, 3, qp, qpc );
}
if( h->mb.i_neighbour & MB_TOP )
{
- int qpt = h->mb.qp[h->mb.i_mb_top_xy];
- int qp_top = (qp + qpt + 1) >> 1;
- int qpc_top = (h->chroma_qp_table[qp] + h->chroma_qp_table[qpt] + 1) >> 1;
- int intra_top = IS_INTRA( h->mb.type[h->mb.i_mb_top_xy] );
- if( !b_interlaced && (intra_cur || intra_top) )
- FILTER( _intra, 1, 0, qp_top, qpc_top );
+ if( b_interlaced && !(mb_y&1) && !MB_INTERLACED && h->mb.field[h->mb.i_mb_top_xy] )
+ {
+ int mbn_xy = mb_xy - 2 * h->mb.i_mb_stride;
+
+ for( int j = 0; j < 2; j++, mbn_xy += h->mb.i_mb_stride )
+ {
+ int qpt = h->mb.qp[mbn_xy];
+ int qp_top = (qp + qpt + 1) >> 1;
+ int qpc_top = (qpc + h->chroma_qp_table[qpt] + 1) >> 1;
+ int intra_top = IS_INTRA( h->mb.type[mbn_xy] );
+ if( intra_cur || intra_top )
+ M32( bs[1][4*j] ) = 0x03030303;
+
+ // deblock the first horizontal edge of the even rows, then the first horizontal edge of the odd rows
+ deblock_edge( h, pixy + j*stridey, 2* stridey, bs[1][4*j], qp_top, a, b, 0, h->loopf.deblock_luma[1] );
+ if( chroma444 )
+ {
+ deblock_edge( h, pixuv + j*strideuv, 2*strideuv, bs[1][4*j], qpc_top, a, b, 0, h->loopf.deblock_luma[1] );
+ deblock_edge( h, pixuv + uvdiff + j*strideuv, 2*strideuv, bs[1][4*j], qpc_top, a, b, 0, h->loopf.deblock_luma[1] );
+ }
+ else
+ deblock_edge( h, pixuv + j*strideuv, 2*strideuv, bs[1][4*j], qpc_top, a, b, 1, h->loopf.deblock_chroma[1] );
+ }
+ }
else
{
- if( intra_top )
- memset( bs[1][0], 3, sizeof(bs[1][0]) );
- FILTER( , 1, 0, qp_top, qpc_top );
+ int qpt = h->mb.qp[h->mb.i_mb_top_xy];
+ int qp_top = (qp + qpt + 1) >> 1;
+ int qpc_top = (qpc + h->chroma_qp_table[qpt] + 1) >> 1;
+ int intra_top = IS_INTRA( h->mb.type[h->mb.i_mb_top_xy] );
+ int intra_deblock = intra_cur || intra_top;
+
+ /* This edge has been modified, reset effective qp to max. */
+ if( h->fdec->mb_info && M32( bs[1][0] ) )
+ {
+ RESET_EFFECTIVE_QP(mb_xy);
+ RESET_EFFECTIVE_QP(h->mb.i_mb_top_xy);
+ }
+
+ if( (!b_interlaced || (!MB_INTERLACED && !h->mb.field[h->mb.i_mb_top_xy])) && intra_deblock )
+ {
+ FILTER( _intra, 1, 0, qp_top, qpc_top );
+ }
+ else
+ {
+ if( intra_deblock )
+ M32( bs[1][0] ) = 0x03030303;
+ FILTER( , 1, 0, qp_top, qpc_top );
+ }
}
}
if( !first_edge_only )
{
- if( !transform_8x8 ) FILTER( , 1, 1, qp, qpc );
- FILTER( , 1, 2, qp, qpc );
- if( !transform_8x8 ) FILTER( , 1, 3, qp, qpc );
+ FILTER( , 1, 1, qp, qpc );
+ FILTER( , 1, 2, qp, qpc );
+ FILTER( , 1, 3, qp, qpc );
}
+
+ #undef FILTER
}
}
-#ifdef HAVE_MMX
-void x264_deblock_v_chroma_mmxext( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_h_chroma_mmxext( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_v_chroma_intra_mmxext( uint8_t *pix, int stride, int alpha, int beta );
-void x264_deblock_h_chroma_intra_mmxext( uint8_t *pix, int stride, int alpha, int beta );
-
-void x264_deblock_v_luma_sse2( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_h_luma_sse2( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_v_luma_intra_sse2( uint8_t *pix, int stride, int alpha, int beta );
-void x264_deblock_h_luma_intra_sse2( uint8_t *pix, int stride, int alpha, int beta );
-void x264_deblock_strength_mmxext( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
- int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][4][4],
- int mvy_limit, int bframe );
-void x264_deblock_strength_sse2 ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
- int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][4][4],
- int mvy_limit, int bframe );
-void x264_deblock_strength_ssse3 ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
- int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][4][4],
- int mvy_limit, int bframe );
-#ifdef ARCH_X86
-void x264_deblock_h_luma_mmxext( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_v8_luma_mmxext( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_h_luma_intra_mmxext( uint8_t *pix, int stride, int alpha, int beta );
-void x264_deblock_v8_luma_intra_mmxext( uint8_t *pix, int stride, int alpha, int beta );
-
-static void x264_deblock_v_luma_mmxext( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 )
-{
- x264_deblock_v8_luma_mmxext( pix, stride, alpha, beta, tc0 );
- x264_deblock_v8_luma_mmxext( pix+8, stride, alpha, beta, tc0+2 );
+/* For deblock-aware RD.
+ * TODO:
+ * deblock macroblock edges
+ * support analysis partitions smaller than 16x16
+ * deblock chroma for 4:2:0/4:2:2
+ * handle duplicate refs correctly
+ */
+void x264_macroblock_deblock( x264_t *h )
+{
+ int a = h->sh.i_alpha_c0_offset - QP_BD_OFFSET;
+ int b = h->sh.i_beta_offset - QP_BD_OFFSET;
+ int qp_thresh = 15 - X264_MIN( a, b ) - X264_MAX( 0, h->pps->i_chroma_qp_index_offset );
+ int intra_cur = IS_INTRA( h->mb.i_type );
+ int qp = h->mb.i_qp;
+ int qpc = h->mb.i_chroma_qp;
+ if( (h->mb.i_partition == D_16x16 && !h->mb.i_cbp_luma && !intra_cur) || qp <= qp_thresh )
+ return;
+
+ uint8_t (*bs)[8][4] = h->mb.cache.deblock_strength;
+ if( intra_cur )
+ {
+ M32( bs[0][1] ) = 0x03030303;
+ M64( bs[0][2] ) = 0x0303030303030303ULL;
+ M32( bs[1][1] ) = 0x03030303;
+ M64( bs[1][2] ) = 0x0303030303030303ULL;
+ }
+ else
+ h->loopf.deblock_strength( h->mb.cache.non_zero_count, h->mb.cache.ref, h->mb.cache.mv,
+ bs, 4 >> MB_INTERLACED, h->sh.i_type == SLICE_TYPE_B );
+
+ int transform_8x8 = h->mb.b_transform_8x8;
+
+ #define FILTER( dir, edge )\
+ do\
+ {\
+ deblock_edge( h, h->mb.pic.p_fdec[0] + 4*edge*(dir?FDEC_STRIDE:1),\
+ FDEC_STRIDE, bs[dir][edge], qp, a, b, 0,\
+ h->loopf.deblock_luma[dir] );\
+ if( CHROMA444 )\
+ {\
+ deblock_edge( h, h->mb.pic.p_fdec[1] + 4*edge*(dir?FDEC_STRIDE:1),\
+ FDEC_STRIDE, bs[dir][edge], qpc, a, b, 0,\
+ h->loopf.deblock_luma[dir] );\
+ deblock_edge( h, h->mb.pic.p_fdec[2] + 4*edge*(dir?FDEC_STRIDE:1),\
+ FDEC_STRIDE, bs[dir][edge], qpc, a, b, 0,\
+ h->loopf.deblock_luma[dir] );\
+ }\
+ } while(0)
+
+ if( !transform_8x8 ) FILTER( 0, 1 );
+ FILTER( 0, 2 );
+ if( !transform_8x8 ) FILTER( 0, 3 );
+
+ if( !transform_8x8 ) FILTER( 1, 1 );
+ FILTER( 1, 2 );
+ if( !transform_8x8 ) FILTER( 1, 3 );
+
+ #undef FILTER
+}
+
+#if HAVE_MMX
+void x264_deblock_v_luma_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_luma_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_chroma_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_chroma_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_mbaff_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_mbaff_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_422_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_422_sse2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_422_avx ( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_luma_intra_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_luma_intra_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_luma_intra_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_luma_intra_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_chroma_intra_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_chroma_intra_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_422_intra_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_422_intra_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_422_intra_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_strength_mmx2 ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+void x264_deblock_strength_sse2 ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+void x264_deblock_strength_ssse3( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+void x264_deblock_strength_avx ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+void x264_deblock_strength_avx2 ( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+
+void x264_deblock_h_chroma_intra_mbaff_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_mbaff_sse2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_mbaff_avx ( pixel *pix, intptr_t stride, int alpha, int beta );
+#if ARCH_X86
+void x264_deblock_h_luma_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v8_luma_mmx2( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_chroma_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_mbaff_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_intra_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v8_luma_intra_mmx2( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_chroma_intra_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_mbaff_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+
+#if HIGH_BIT_DEPTH
+void x264_deblock_v_luma_mmx2( pixel *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_luma_intra_mmx2( pixel *pix, intptr_t stride, int alpha, int beta );
+#else
+// FIXME this wrapper has a significant cpu cost
+static void x264_deblock_v_luma_mmx2( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 )
+{
+ x264_deblock_v8_luma_mmx2( pix, stride, alpha, beta, tc0 );
+ x264_deblock_v8_luma_mmx2( pix+8, stride, alpha, beta, tc0+2 );
}
-static void x264_deblock_v_luma_intra_mmxext( uint8_t *pix, int stride, int alpha, int beta )
+static void x264_deblock_v_luma_intra_mmx2( uint8_t *pix, intptr_t stride, int alpha, int beta )
{
- x264_deblock_v8_luma_intra_mmxext( pix, stride, alpha, beta );
- x264_deblock_v8_luma_intra_mmxext( pix+8, stride, alpha, beta );
+ x264_deblock_v8_luma_intra_mmx2( pix, stride, alpha, beta );
+ x264_deblock_v8_luma_intra_mmx2( pix+8, stride, alpha, beta );
}
+#endif // HIGH_BIT_DEPTH
#endif
#endif
-#ifdef ARCH_PPC
-void x264_deblock_v_luma_altivec( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
-void x264_deblock_h_luma_altivec( uint8_t *pix, int stride, int alpha, int beta, int8_t *tc0 );
+#if ARCH_PPC
+void x264_deblock_v_luma_altivec( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_altivec( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
#endif // ARCH_PPC
-#ifdef HAVE_ARMV6
-void x264_deblock_v_luma_neon( uint8_t *, int, int, int, int8_t * );
-void x264_deblock_h_luma_neon( uint8_t *, int, int, int, int8_t * );
-void x264_deblock_v_chroma_neon( uint8_t *, int, int, int, int8_t * );
-void x264_deblock_h_chroma_neon( uint8_t *, int, int, int, int8_t * );
+#if HAVE_ARMV6 || ARCH_AARCH64
+void x264_deblock_v_luma_neon ( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_neon ( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_chroma_neon( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_neon( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_strength_neon( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4],
+ int mvy_limit, int bframe );
+void x264_deblock_h_chroma_422_neon( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_mbaff_neon( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_intra_mbaff_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_422_intra_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_chroma_intra_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_luma_intra_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_luma_intra_neon( uint8_t *pix, intptr_t stride, int alpha, int beta );
#endif
-void x264_deblock_init( int cpu, x264_deblock_function_t *pf )
+#if !HIGH_BIT_DEPTH
+#if HAVE_MSA
+void x264_deblock_v_luma_msa( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_luma_msa( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_chroma_msa( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_h_chroma_msa( uint8_t *pix, intptr_t stride, int alpha, int beta, int8_t *tc0 );
+void x264_deblock_v_luma_intra_msa( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_luma_intra_msa( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_v_chroma_intra_msa( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_h_chroma_intra_msa( uint8_t *pix, intptr_t stride, int alpha, int beta );
+void x264_deblock_strength_msa( uint8_t nnz[X264_SCAN8_SIZE], int8_t ref[2][X264_SCAN8_LUMA_SIZE],
+ int16_t mv[2][X264_SCAN8_LUMA_SIZE][2], uint8_t bs[2][8][4], int mvy_limit,
+ int bframe );
+#endif
+#endif
+
+void x264_deblock_init( int cpu, x264_deblock_function_t *pf, int b_mbaff )
{
pf->deblock_luma[1] = deblock_v_luma_c;
pf->deblock_luma[0] = deblock_h_luma_c;
pf->deblock_chroma[1] = deblock_v_chroma_c;
- pf->deblock_chroma[0] = deblock_h_chroma_c;
+ pf->deblock_h_chroma_420 = deblock_h_chroma_c;
+ pf->deblock_h_chroma_422 = deblock_h_chroma_422_c;
pf->deblock_luma_intra[1] = deblock_v_luma_intra_c;
pf->deblock_luma_intra[0] = deblock_h_luma_intra_c;
pf->deblock_chroma_intra[1] = deblock_v_chroma_intra_c;
- pf->deblock_chroma_intra[0] = deblock_h_chroma_intra_c;
+ pf->deblock_h_chroma_420_intra = deblock_h_chroma_intra_c;
+ pf->deblock_h_chroma_422_intra = deblock_h_chroma_422_intra_c;
+ pf->deblock_luma_mbaff = deblock_h_luma_mbaff_c;
+ pf->deblock_chroma_420_mbaff = deblock_h_chroma_mbaff_c;
+ pf->deblock_luma_intra_mbaff = deblock_h_luma_intra_mbaff_c;
+ pf->deblock_chroma_420_intra_mbaff = deblock_h_chroma_intra_mbaff_c;
pf->deblock_strength = deblock_strength_c;
-#ifdef HAVE_MMX
- if( cpu&X264_CPU_MMXEXT )
+#if HAVE_MMX
+ if( cpu&X264_CPU_MMX2 )
{
- pf->deblock_chroma[1] = x264_deblock_v_chroma_mmxext;
- pf->deblock_chroma[0] = x264_deblock_h_chroma_mmxext;
- pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_mmxext;
- pf->deblock_chroma_intra[0] = x264_deblock_h_chroma_intra_mmxext;
-#ifdef ARCH_X86
- pf->deblock_luma[1] = x264_deblock_v_luma_mmxext;
- pf->deblock_luma[0] = x264_deblock_h_luma_mmxext;
- pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_mmxext;
- pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_mmxext;
+#if ARCH_X86
+ pf->deblock_luma[1] = x264_deblock_v_luma_mmx2;
+ pf->deblock_luma[0] = x264_deblock_h_luma_mmx2;
+ pf->deblock_chroma[1] = x264_deblock_v_chroma_mmx2;
+ pf->deblock_h_chroma_420 = x264_deblock_h_chroma_mmx2;
+ pf->deblock_chroma_420_mbaff = x264_deblock_h_chroma_mbaff_mmx2;
+ pf->deblock_h_chroma_422 = x264_deblock_h_chroma_422_mmx2;
+ pf->deblock_h_chroma_422_intra = x264_deblock_h_chroma_422_intra_mmx2;
+ pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_mmx2;
+ pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_mmx2;
+ pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_mmx2;
+ pf->deblock_h_chroma_420_intra = x264_deblock_h_chroma_intra_mmx2;
+ pf->deblock_chroma_420_intra_mbaff = x264_deblock_h_chroma_intra_mbaff_mmx2;
+#endif
+#if !HIGH_BIT_DEPTH
+ pf->deblock_chroma_420_intra_mbaff = x264_deblock_h_chroma_intra_mbaff_mmx2;
#endif
- pf->deblock_strength = x264_deblock_strength_mmxext;
+ pf->deblock_strength = x264_deblock_strength_mmx2;
if( cpu&X264_CPU_SSE2 )
{
pf->deblock_strength = x264_deblock_strength_sse2;
+ pf->deblock_h_chroma_420 = x264_deblock_h_chroma_sse2;
+ pf->deblock_h_chroma_422 = x264_deblock_h_chroma_422_sse2;
+ pf->deblock_h_chroma_422_intra = x264_deblock_h_chroma_422_intra_sse2;
+ pf->deblock_chroma_420_mbaff = x264_deblock_h_chroma_mbaff_sse2;
+ pf->deblock_luma[1] = x264_deblock_v_luma_sse2;
+ pf->deblock_luma[0] = x264_deblock_h_luma_sse2;
+ pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_sse2;
+ pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_sse2;
if( !(cpu&X264_CPU_STACK_MOD4) )
{
- pf->deblock_luma[1] = x264_deblock_v_luma_sse2;
- pf->deblock_luma[0] = x264_deblock_h_luma_sse2;
- pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_sse2;
- pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_sse2;
+ pf->deblock_chroma[1] = x264_deblock_v_chroma_sse2;
+ pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_sse2;
+ pf->deblock_h_chroma_420_intra = x264_deblock_h_chroma_intra_sse2;
+#if HIGH_BIT_DEPTH
+ pf->deblock_chroma_420_intra_mbaff= x264_deblock_h_chroma_intra_mbaff_sse2;
+#endif
}
}
if( cpu&X264_CPU_SSSE3 )
pf->deblock_strength = x264_deblock_strength_ssse3;
+ if( cpu&X264_CPU_AVX )
+ {
+ pf->deblock_strength = x264_deblock_strength_avx;
+ pf->deblock_h_chroma_420 = x264_deblock_h_chroma_avx;
+ pf->deblock_h_chroma_422 = x264_deblock_h_chroma_422_avx;
+ pf->deblock_h_chroma_422_intra = x264_deblock_h_chroma_422_intra_avx;
+ pf->deblock_luma[1] = x264_deblock_v_luma_avx;
+ pf->deblock_luma[0] = x264_deblock_h_luma_avx;
+ pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_avx;
+ pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_avx;
+ if( !(cpu&X264_CPU_STACK_MOD4) )
+ {
+ pf->deblock_chroma[1] = x264_deblock_v_chroma_avx;
+ pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_avx;
+ pf->deblock_h_chroma_420_intra = x264_deblock_h_chroma_intra_avx;
+#if HIGH_BIT_DEPTH
+ pf->deblock_chroma_420_mbaff = x264_deblock_h_chroma_mbaff_avx;
+ pf->deblock_chroma_420_intra_mbaff = x264_deblock_h_chroma_intra_mbaff_avx;
+#endif
+ }
+ }
+ if( cpu&X264_CPU_AVX2 )
+ {
+ pf->deblock_strength = x264_deblock_strength_avx2;
+ }
}
#endif
-#ifdef HAVE_ALTIVEC
+#if !HIGH_BIT_DEPTH
+#if HAVE_ALTIVEC
if( cpu&X264_CPU_ALTIVEC )
{
pf->deblock_luma[1] = x264_deblock_v_luma_altivec;
pf->deblock_luma[0] = x264_deblock_h_luma_altivec;
- }
+ }
#endif // HAVE_ALTIVEC
-#ifdef HAVE_ARMV6
- if( cpu&X264_CPU_NEON )
- {
+#if HAVE_ARMV6 || ARCH_AARCH64
+ if( cpu&X264_CPU_NEON )
+ {
pf->deblock_luma[1] = x264_deblock_v_luma_neon;
pf->deblock_luma[0] = x264_deblock_h_luma_neon;
pf->deblock_chroma[1] = x264_deblock_v_chroma_neon;
- pf->deblock_chroma[0] = x264_deblock_h_chroma_neon;
- }
+ pf->deblock_h_chroma_420 = x264_deblock_h_chroma_neon;
+ pf->deblock_h_chroma_422 = x264_deblock_h_chroma_422_neon;
+ pf->deblock_chroma_420_mbaff = x264_deblock_h_chroma_mbaff_neon;
+ pf->deblock_chroma_420_intra_mbaff = x264_deblock_h_chroma_intra_mbaff_neon;
+ pf->deblock_h_chroma_420_intra = x264_deblock_h_chroma_intra_neon;
+ pf->deblock_h_chroma_422_intra = x264_deblock_h_chroma_422_intra_neon;
+ pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_neon;
+ pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_neon;
+ pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_neon;
+ pf->deblock_strength = x264_deblock_strength_neon;
+ }
+#endif
+
+#if HAVE_MSA
+ if( cpu&X264_CPU_MSA )
+ {
+ pf->deblock_luma[1] = x264_deblock_v_luma_msa;
+ pf->deblock_luma[0] = x264_deblock_h_luma_msa;
+ pf->deblock_chroma[1] = x264_deblock_v_chroma_msa;
+ pf->deblock_h_chroma_420 = x264_deblock_h_chroma_msa;
+ pf->deblock_luma_intra[1] = x264_deblock_v_luma_intra_msa;
+ pf->deblock_luma_intra[0] = x264_deblock_h_luma_intra_msa;
+ pf->deblock_chroma_intra[1] = x264_deblock_v_chroma_intra_msa;
+ pf->deblock_h_chroma_420_intra = x264_deblock_h_chroma_intra_msa;
+ pf->deblock_strength = x264_deblock_strength_msa;
+ }
#endif
+#endif // !HIGH_BIT_DEPTH
+
+ /* These functions are equivalent, so don't duplicate them. */
+ pf->deblock_chroma_422_mbaff = pf->deblock_h_chroma_420;
+ pf->deblock_chroma_422_intra_mbaff = pf->deblock_h_chroma_420_intra;
}