* subpel_iters[i_subpel_refine] = { refine_hpel, refine_qpel, me_hpel, me_qpel }
* where me_* are the number of EPZS iterations run on all candidate block types,
* and refine_* are run only on the winner. */
-const static int subpel_iterations[][4] =
+static const int subpel_iterations[][4] =
{{1,0,0,0},
{1,1,0,0},
{1,2,0,0},
#define COST_MV( mx, my ) \
{ \
- int cost = h->pixf.sad[i_pixel]( m->p_fenc, m->i_stride, \
- &p_fref[(my)*m->i_stride+(mx)], m->i_stride ) + \
- m->lm * ( bs_size_se(((mx)<<2) - m->mvp[0] ) + \
- bs_size_se(((my)<<2) - m->mvp[1] ) ); \
+ int cost = h->pixf.sad[i_pixel]( m->p_fenc[0], m->i_stride[0], \
+ &p_fref[(my)*m->i_stride[0]+(mx)], m->i_stride[0] ) \
+ + p_cost_mvx[ (mx)<<2 ] \
+ + p_cost_mvy[ (my)<<2 ]; \
if( cost < bcost ) \
{ \
bcost = cost; \
void x264_me_search_ref( x264_t *h, x264_me_t *m, int (*mvc)[2], int i_mvc, int *p_fullpel_thresh )
{
const int i_pixel = m->i_pixel;
+ const unsigned int i_me_range = h->param.analyse.i_me_range;
+ const int b_chroma_me = h->mb.b_chroma_me && i_pixel <= PIXEL_8x8;
int bmx, bmy, bcost;
- int omx, omy;
- uint8_t *p_fref = m->p_fref;
- int i_iter;
-
- const int mv_x_min = h->mb.mv_min[0] + 16;
- const int mv_y_min = h->mb.mv_min[1] + 16;
- const int mv_x_max = h->mb.mv_max[0] - 16;
- const int mv_y_max = h->mb.mv_max[1] - 16;
-
- /* init with mvp */
- /* XXX: We don't need to clamp because the way diamond work, we will
- * never go outside padded picture, and predict mv won't compute vector
- * with componant magnitude greater.
- * XXX: if some vector can go outside, (accelerator, ....) you need to clip
- * them yourself */
- bmx = x264_clip3( ( m->mvp[0] + 2 ) >> 2, mv_x_min, mv_x_max );
- bmy = x264_clip3( ( m->mvp[1] + 2 ) >> 2, mv_y_min, mv_y_max );
-
- bcost = h->pixf.sad[i_pixel]( m->p_fenc, m->i_stride,
- &p_fref[bmy * m->i_stride + bmx], m->i_stride );
+ int omx, omy, pmx, pmy;
+ uint8_t *p_fref = m->p_fref[0];
+ int i, j;
+
+ int mv_x_min = h->mb.mv_min_fpel[0];
+ int mv_y_min = h->mb.mv_min_fpel[1];
+ int mv_x_max = h->mb.mv_max_fpel[0];
+ int mv_y_max = h->mb.mv_max_fpel[1];
+
+ const int16_t *p_cost_mvx = m->p_cost_mv - m->mvp[0];
+ const int16_t *p_cost_mvy = m->p_cost_mv - m->mvp[1];
+
+ if( h->mb.i_me_method == X264_ME_UMH )
+ {
+ /* clamp mvp to inside frame+padding, so that we don't have to check it each iteration */
+ p_cost_mvx = m->p_cost_mv - x264_clip3( m->mvp[0], h->mb.mv_min[0], h->mb.mv_max[0] );
+ p_cost_mvy = m->p_cost_mv - x264_clip3( m->mvp[1], h->mb.mv_min[1], h->mb.mv_max[1] );
+ }
+
+ bmx = pmx = x264_clip3( ( m->mvp[0] + 2 ) >> 2, mv_x_min, mv_x_max );
+ bmy = pmy = x264_clip3( ( m->mvp[1] + 2 ) >> 2, mv_y_min, mv_y_max );
+ bcost = COST_MAX;
+ COST_MV( bmx, bmy );
+ /* I don't know why this helps */
+ bcost -= p_cost_mvx[ bmx<<2 ] + p_cost_mvy[ bmy<<2 ];
/* try extra predictors if provided */
- for( i_iter = 0; i_iter < i_mvc; i_iter++ )
+ for( i = 0; i < i_mvc; i++ )
{
- const int mx = x264_clip3( ( mvc[i_iter][0] + 2 ) >> 2, mv_x_min, mv_x_max );
- const int my = x264_clip3( ( mvc[i_iter][1] + 2 ) >> 2, mv_y_min, mv_y_max );
+ const int mx = x264_clip3( ( mvc[i][0] + 2 ) >> 2, mv_x_min, mv_x_max );
+ const int my = x264_clip3( ( mvc[i][1] + 2 ) >> 2, mv_y_min, mv_y_max );
if( mx != bmx || my != bmy )
COST_MV( mx, my );
}
COST_MV( 0, 0 );
- if( h->param.analyse.i_subpel_refine >= 2 )
+ mv_x_max += 8;
+ mv_y_max += 8;
+ mv_x_min -= 8;
+ mv_y_min -= 8;
+
+ switch( h->mb.i_me_method )
{
- /* hexagon search */
- /* Don't need to test mv_range each time, we won't go outside picture+padding */
- omx = bmx;
- omy = bmy;
- for( i_iter = 0; i_iter < 8; i_iter++ )
+ case X264_ME_DIA:
+ /* diamond search, radius 1 */
+#define DIA1_ITER(mx, my)\
+ {\
+ omx = mx;\
+ omy = my;\
+ COST_MV( omx , omy-1 );\
+ COST_MV( omx , omy+1 );\
+ COST_MV( omx-1, omy );\
+ COST_MV( omx+1, omy );\
+ }
+
+ for( i = 0; i < i_me_range; i++ )
{
- COST_MV( omx-2, omy );
- COST_MV( omx-1, omy+2 );
- COST_MV( omx+1, omy+2 );
- COST_MV( omx+2, omy );
- COST_MV( omx+1, omy-2 );
- COST_MV( omx-1, omy-2 );
+ DIA1_ITER( bmx, bmy );
+ if( bmx == omx && bmy == omy )
+ break;
+ }
+ break;
+
+ case X264_ME_HEX:
+ /* hexagon search, radius 2 */
+#define HEX2_ITER(mx, my)\
+ {\
+ omx = mx;\
+ omy = my;\
+ COST_MV( omx-2, omy );\
+ COST_MV( omx-1, omy+2 );\
+ COST_MV( omx+1, omy+2 );\
+ COST_MV( omx+2, omy );\
+ COST_MV( omx+1, omy-2 );\
+ COST_MV( omx-1, omy-2 );\
+ }
+ for( i = 0; i < i_me_range/2; i++ )
+ {
+ HEX2_ITER( bmx, bmy );
if( bmx == omx && bmy == omy )
break;
- omx = bmx;
- omy = bmy;
}
-
/* square refine */
+ DIA1_ITER( bmx, bmy );
COST_MV( omx-1, omy-1 );
- COST_MV( omx-1, omy );
COST_MV( omx-1, omy+1 );
- COST_MV( omx , omy-1 );
- COST_MV( omx , omy+1 );
COST_MV( omx+1, omy-1 );
- COST_MV( omx+1, omy );
COST_MV( omx+1, omy+1 );
- }
- else
- {
- /* diamond search */
- for( i_iter = 0; i_iter < 16; i_iter++ )
+ break;
+
+ case X264_ME_UMH:
+ /* Uneven-cross Multi-Hexagon-grid Search
+ * as in JM, except without early termination */
+
+ DIA1_ITER( pmx, pmy );
+ if( pmx || pmy )
+ DIA1_ITER( 0, 0 );
+ DIA1_ITER( bmx, bmy );
+
+ if(i_pixel == PIXEL_4x4)
+ goto umh_small_hex;
+
+ /* cross */
+ omx = bmx; omy = bmy;
+ for( i = 1; i < i_me_range; i+=2 )
+ {
+ if( omx + i <= mv_x_max )
+ COST_MV( omx + i, omy );
+ if( omx - i >= mv_x_min )
+ COST_MV( omx - i, omy );
+ }
+ for( i = 1; i < i_me_range/2; i+=2 )
+ {
+ if( omy + i <= mv_y_max )
+ COST_MV( omx, omy + i );
+ if( omy - i >= mv_y_min )
+ COST_MV( omx, omy - i );
+ }
+
+ /* 5x5 ESA */
+ omx = bmx; omy = bmy;
+ for( i = 0; i < 24; i++ )
+ {
+ static const int square2_x[24] = {1,1,0,-1,-1,-1, 0, 1, 2,2,2,2,1,0,-1,-2,-2,-2,-2,-2,-1, 0, 1, 2};
+ static const int square2_y[24] = {0,1,1, 1, 0,-1,-1,-1,-1,0,1,2,2,2, 2, 2, 1, 0,-1,-2,-2,-2,-2,-2};
+ COST_MV( omx + square2_x[i], omy + square2_y[i] );
+ }
+ /* hexagon grid */
+ omx = bmx; omy = bmy;
+ for( i = 1; i <= i_me_range/4; i++ )
+ {
+ int bounds_check = 4*i > X264_MIN4( mv_x_max-omx, mv_y_max-omy, omx-mv_x_min, omy-mv_y_min );
+ for( j = 0; j < 16; j++ )
+ {
+ static const int hex4_x[16] = {0,-2,-4,-4,-4,-4,-4,-2, 0, 2, 4, 4,4,4,4,2};
+ static const int hex4_y[16] = {4, 3, 2, 1, 0,-1,-2,-3,-4,-3,-2,-1,0,1,2,3};
+ int mx = omx + hex4_x[j]*i;
+ int my = omy + hex4_y[j]*i;
+ if( !bounds_check || ( mx >= mv_x_min && mx <= mv_x_max
+ && my >= mv_y_min && my <= mv_y_max ) )
+ COST_MV( mx, my );
+ }
+ }
+umh_small_hex:
+ /* iterative search */
+ for( i = 0; i < i_me_range; i++ )
{
- omx = bmx;
- omy = bmy;
- COST_MV( omx , omy-1 );
- COST_MV( omx , omy+1 );
- COST_MV( omx-1, omy );
- COST_MV( omx+1, omy );
+ HEX2_ITER( bmx, bmy );
if( bmx == omx && bmy == omy )
break;
}
+ for( i = 0; i < i_me_range; i++ )
+ {
+ DIA1_ITER( bmx, bmy );
+ if( bmx == omx && bmy == omy )
+ break;
+ }
+ break;
+
+ case X264_ME_ESA:
+ {
+ const int min_x = X264_MAX( bmx - i_me_range, mv_x_min);
+ const int min_y = X264_MAX( bmy - i_me_range, mv_y_min);
+ const int max_x = X264_MIN( bmx + i_me_range, mv_x_max);
+ const int max_y = X264_MIN( bmy + i_me_range, mv_y_max);
+ for( omy = min_y; omy <= max_y; omy++ )
+ for( omx = min_x; omx <= max_x; omx++ )
+ {
+ COST_MV( omx, omy );
+ }
+ }
+ break;
}
/* -> qpel mv */
m->mv[1] = bmy << 2;
/* compute the real cost */
- m->cost = h->pixf.satd[i_pixel]( m->p_fenc, m->i_stride,
- &p_fref[bmy * m->i_stride + bmx], m->i_stride ) +
- m->lm * ( bs_size_se( m->mv[0] - m->mvp[0] ) +
- bs_size_se( m->mv[1] - m->mvp[1] ) );
+ m->cost_mv = p_cost_mvx[ m->mv[0] ] + p_cost_mvy[ m->mv[1] ];
+ m->cost = h->pixf.satd[i_pixel]( m->p_fenc[0], m->i_stride[0],
+ &p_fref[bmy * m->i_stride[0] + bmx], m->i_stride[0] )
+ + m->cost_mv;
+ if( b_chroma_me )
+ {
+ const int bw = x264_pixel_size[m->i_pixel].w;
+ const int bh = x264_pixel_size[m->i_pixel].h;
+ DECLARE_ALIGNED( uint8_t, pix[8*8*2], 16 );
+ h->mc.mc_chroma( m->p_fref[4], m->i_stride[1], pix, 8, m->mv[0], m->mv[1], bw/2, bh/2 );
+ h->mc.mc_chroma( m->p_fref[5], m->i_stride[1], pix+8*8, 8, m->mv[0], m->mv[1], bw/2, bh/2 );
+ m->cost += h->pixf.satd[i_pixel+3]( m->p_fenc[1], m->i_stride[1], pix, 8 )
+ + h->pixf.satd[i_pixel+3]( m->p_fenc[2], m->i_stride[1], pix+8*8, 8 );
+ }
/* subpel refine */
- if( h->param.analyse.i_subpel_refine >= 3 )
+ if( h->mb.i_subpel_refine >= 3 )
{
int hpel, qpel;
*p_fullpel_thresh = m->cost;
}
- hpel = subpel_iterations[h->param.analyse.i_subpel_refine][2];
- qpel = subpel_iterations[h->param.analyse.i_subpel_refine][3];
+ hpel = subpel_iterations[h->mb.i_subpel_refine][2];
+ qpel = subpel_iterations[h->mb.i_subpel_refine][3];
refine_subpel( h, m, hpel, qpel );
}
}
void x264_me_refine_qpel( x264_t *h, x264_me_t *m )
{
- int hpel = subpel_iterations[h->param.analyse.i_subpel_refine][0];
- int qpel = subpel_iterations[h->param.analyse.i_subpel_refine][1];
+ int hpel = subpel_iterations[h->mb.i_subpel_refine][0];
+ int qpel = subpel_iterations[h->mb.i_subpel_refine][1];
// if( hpel || qpel )
refine_subpel( h, m, hpel, qpel );
}
+#define COST_MV( mx, my ) \
+{ \
+ int stride = 16; \
+ uint8_t *src = h->mc.get_ref( m->p_fref, m->i_stride[0], pix, &stride, mx, my, bw, bh ); \
+ int cost = h->pixf.satd[i_pixel]( m->p_fenc[0], m->i_stride[0], src, stride ) \
+ + p_cost_mvx[ mx ] + p_cost_mvy[ my ]; \
+ if( b_chroma_me && cost < bcost ) \
+ { \
+ h->mc.mc_chroma( m->p_fref[4], m->i_stride[1], pix, 8, mx, my, bw/2, bh/2 ); \
+ cost += h->pixf.satd[i_pixel+3]( m->p_fenc[1], m->i_stride[1], pix, 8 ); \
+ if( cost < bcost ) \
+ { \
+ h->mc.mc_chroma( m->p_fref[5], m->i_stride[1], pix, 8, mx, my, bw/2, bh/2 ); \
+ cost += h->pixf.satd[i_pixel+3]( m->p_fenc[2], m->i_stride[1], pix, 8 ); \
+ } \
+ } \
+ if( cost < bcost ) \
+ { \
+ bcost = cost; \
+ bmx = mx; \
+ bmy = my; \
+ } \
+}
+
static void refine_subpel( x264_t *h, x264_me_t *m, int hpel_iters, int qpel_iters )
{
const int bw = x264_pixel_size[m->i_pixel].w;
const int bh = x264_pixel_size[m->i_pixel].h;
+ const int16_t *p_cost_mvx = m->p_cost_mv - m->mvp[0];
+ const int16_t *p_cost_mvy = m->p_cost_mv - m->mvp[1];
+ const int i_pixel = m->i_pixel;
+ const int b_chroma_me = h->mb.b_chroma_me && i_pixel <= PIXEL_8x8;
- DECLARE_ALIGNED( uint8_t, pix[4][16*16], 16 );
- int cost[4];
- int best;
+ DECLARE_ALIGNED( uint8_t, pix[16*16], 16 );
int step, i;
int bmx = m->mv[0];
int bmy = m->mv[1];
+ int bcost = m->cost;
+
+ /* try the subpel component of the predicted mv if it's close to
+ * the result of the fullpel search */
+ if( hpel_iters )
+ {
+ int mx = X264_ABS(bmx - m->mvp[0]) < 4 ? m->mvp[0] : bmx;
+ int my = X264_ABS(bmy - m->mvp[1]) < 4 ? m->mvp[1] : bmy;
+ if( mx != bmx || my != bmy )
+ COST_MV( mx, my );
+ }
for( step = 2; step >= 1; step-- )
{
for( i = step>1 ? hpel_iters : qpel_iters; i > 0; i-- )
{
- h->mc[MC_LUMA]( m->p_fref, m->i_stride, pix[0], 16, bmx + 0, bmy - step, bw, bh );
- h->mc[MC_LUMA]( m->p_fref, m->i_stride, pix[1], 16, bmx + 0, bmy + step, bw, bh );
- h->mc[MC_LUMA]( m->p_fref, m->i_stride, pix[2], 16, bmx - step, bmy + 0, bw, bh );
- h->mc[MC_LUMA]( m->p_fref, m->i_stride, pix[3], 16, bmx + step, bmy + 0, bw, bh );
-
- cost[0] = h->pixf.satd[m->i_pixel]( m->p_fenc, m->i_stride, pix[0], 16 ) +
- m->lm * ( bs_size_se( bmx + 0 - m->mvp[0] ) + bs_size_se( bmy - step - m->mvp[1] ) );
- cost[1] = h->pixf.satd[m->i_pixel]( m->p_fenc, m->i_stride, pix[1], 16 ) +
- m->lm * ( bs_size_se( bmx + 0 - m->mvp[0] ) + bs_size_se( bmy + step - m->mvp[1] ) );
- cost[2] = h->pixf.satd[m->i_pixel]( m->p_fenc, m->i_stride, pix[2], 16 ) +
- m->lm * ( bs_size_se( bmx - step - m->mvp[0] ) + bs_size_se( bmy + 0 - m->mvp[1] ) );
- cost[3] = h->pixf.satd[m->i_pixel]( m->p_fenc, m->i_stride, pix[3], 16 ) +
- m->lm * ( bs_size_se( bmx + step - m->mvp[0] ) + bs_size_se( bmy + 0 - m->mvp[1] ) );
-
- best = 0;
- if( cost[1] < cost[0] ) best = 1;
- if( cost[2] < cost[best] ) best = 2;
- if( cost[3] < cost[best] ) best = 3;
-
- if( cost[best] < m->cost )
- {
- m->cost = cost[best];
- if( best == 0 ) bmy -= step;
- else if( best == 1 ) bmy += step;
- else if( best == 2 ) bmx -= step;
- else if( best == 3 ) bmx += step;
- }
- else break;
+ int omx = bmx;
+ int omy = bmy;
+ COST_MV( omx, omy - step );
+ COST_MV( omx, omy + step );
+ COST_MV( omx - step, omy );
+ COST_MV( omx + step, omy );
+ if( bmx == omx && bmy == omy )
+ break;
}
}
+ m->cost = bcost;
m->mv[0] = bmx;
m->mv[1] = bmy;
+ m->cost_mv = p_cost_mvx[ bmx ] + p_cost_mvy[ bmy ];
}