1 /*****************************************************************************
2 * slicetype.c: lookahead analysis
3 *****************************************************************************
4 * Copyright (C) 2005-2010 x264 project
6 * Authors: Fiona Glaser <fiona@x264.com>
7 * Loren Merritt <lorenm@u.washington.edu>
8 * Dylan Yudaken <dyudaken@gmail.com>
10 * This program is free software; you can redistribute it and/or modify
11 * it under the terms of the GNU General Public License as published by
12 * the Free Software Foundation; either version 2 of the License, or
13 * (at your option) any later version.
15 * This program is distributed in the hope that it will be useful,
16 * but WITHOUT ANY WARRANTY; without even the implied warranty of
17 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
18 * GNU General Public License for more details.
20 * You should have received a copy of the GNU General Public License
21 * along with this program; if not, write to the Free Software
22 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02111, USA.
24 * This program is also available under a commercial proprietary license.
25 * For more information, contact us at licensing@x264.com.
26 *****************************************************************************/
30 #include "common/common.h"
31 #include "macroblock.h"
34 // Indexed by pic_struct values
35 static const uint8_t delta_tfi_divisor[10] = { 0, 2, 1, 1, 2, 2, 3, 3, 4, 6 };
37 static int x264_slicetype_frame_cost( x264_t *h, x264_mb_analysis_t *a,
38 x264_frame_t **frames, int p0, int p1, int b,
39 int b_intra_penalty );
41 static void x264_lowres_context_init( x264_t *h, x264_mb_analysis_t *a )
43 a->i_qp = X264_LOOKAHEAD_QP;
44 a->i_lambda = x264_lambda_tab[ a->i_qp ];
45 x264_mb_analyse_load_costs( h, a );
46 if( h->param.analyse.i_subpel_refine > 1 )
48 h->mb.i_me_method = X264_MIN( X264_ME_HEX, h->param.analyse.i_me_method );
49 h->mb.i_subpel_refine = 4;
53 h->mb.i_me_method = X264_ME_DIA;
54 h->mb.i_subpel_refine = 2;
56 h->mb.b_chroma_me = 0;
59 /* makes a non-h264 weight (i.e. fix7), into an h264 weight */
60 static void x264_weight_get_h264( int weight_nonh264, int offset, x264_weight_t *w )
64 w->i_scale = weight_nonh264;
65 while( w->i_denom > 0 && (w->i_scale > 127 || !(w->i_scale & 1)) )
70 w->i_scale = X264_MIN( w->i_scale, 127 );
73 static NOINLINE pixel *x264_weight_cost_init_luma( x264_t *h, x264_frame_t *fenc, x264_frame_t *ref, pixel *dest )
75 int ref0_distance = fenc->i_frame - ref->i_frame - 1;
76 /* Note: this will never run during lookahead as weights_analyse is only called if no
77 * motion search has been done. */
78 if( fenc->lowres_mvs[0][ref0_distance][0][0] != 0x7FFF )
80 int i_stride = fenc->i_stride_lowres;
81 int i_lines = fenc->i_lines_lowres;
82 int i_width = fenc->i_width_lowres;
86 for( int y = 0; y < i_lines; y += 8, p += i_stride*8 )
87 for( int x = 0; x < i_width; x += 8, i_mb_xy++ )
89 int mvx = fenc->lowres_mvs[0][ref0_distance][i_mb_xy][0];
90 int mvy = fenc->lowres_mvs[0][ref0_distance][i_mb_xy][1];
91 h->mc.mc_luma( p+x, i_stride, ref->lowres, i_stride,
92 mvx+(x<<2), mvy+(y<<2), 8, 8, weight_none );
98 return ref->lowres[0];
101 /* How data is organized for chroma weightp:
104 * fenc = ref + offset
105 * v = u + stride * chroma height
106 * We'll need more room if we do 4:2:2 or 4:4:4. */
108 static NOINLINE void x264_weight_cost_init_chroma( x264_t *h, x264_frame_t *fenc, x264_frame_t *ref, pixel *dstu, pixel *dstv )
110 int ref0_distance = fenc->i_frame - ref->i_frame - 1;
111 int i_stride = fenc->i_stride[1];
112 int i_offset = i_stride / 2;
113 int i_lines = fenc->i_lines[1];
114 int i_width = fenc->i_width[1];
115 int cw = h->mb.i_mb_width << 3;
116 int ch = h->mb.i_mb_height << 3;
118 if( fenc->lowres_mvs[0][ref0_distance][0][0] != 0x7FFF )
120 for( int y = 0, mb_xy = 0, pel_offset_y = 0; y < i_lines; y += 8, pel_offset_y = y*i_stride )
121 for( int x = 0, pel_offset_x = 0; x < i_width; x += 8, mb_xy++, pel_offset_x += 8 )
123 pixel *pixu = dstu + pel_offset_y + pel_offset_x;
124 pixel *pixv = dstv + pel_offset_y + pel_offset_x;
125 pixel *src1 = ref->plane[1] + pel_offset_y + pel_offset_x*2; /* NV12 */
126 int mvx = fenc->lowres_mvs[0][ref0_distance][mb_xy][0];
127 int mvy = fenc->lowres_mvs[0][ref0_distance][mb_xy][1];
128 h->mc.mc_chroma( pixu, pixv, i_stride, src1, i_stride, mvx, mvy, 8, 8 );
132 h->mc.plane_copy_deinterleave( dstu, i_stride, dstv, i_stride, ref->plane[1], i_stride, cw, ch );
133 h->mc.plane_copy_deinterleave( dstu+i_offset, i_stride, dstv+i_offset, i_stride, fenc->plane[1], i_stride, cw, ch );
137 static int x264_weight_slice_header_cost( x264_t *h, x264_weight_t *w, int b_chroma )
139 /* Add cost of weights in the slice header. */
140 int lambda = x264_lambda_tab[X264_LOOKAHEAD_QP];
142 if( h->param.i_slice_count )
143 numslices = h->param.i_slice_count;
144 else if( h->param.i_slice_max_mbs )
145 numslices = (h->mb.i_mb_width * h->mb.i_mb_height + h->param.i_slice_max_mbs-1) / h->param.i_slice_max_mbs;
148 /* FIXME: find a way to account for --slice-max-size?
149 * Multiply by 2 as there will be a duplicate. 10 bits added as if there is a weighted frame, then an additional duplicate is used.
150 * Cut denom cost in half if chroma, since it's shared between the two chroma planes. */
151 int denom_cost = bs_size_ue( w[0].i_denom ) * (2 - b_chroma);
152 return lambda * numslices * ( 10 + denom_cost + 2 * (bs_size_se( w[0].i_scale ) + bs_size_se( w[0].i_offset )) );
155 static NOINLINE unsigned int x264_weight_cost_luma( x264_t *h, x264_frame_t *fenc, pixel *src, x264_weight_t *w )
157 unsigned int cost = 0;
158 int i_stride = fenc->i_stride_lowres;
159 int i_lines = fenc->i_lines_lowres;
160 int i_width = fenc->i_width_lowres;
161 pixel *fenc_plane = fenc->lowres[0];
162 ALIGNED_ARRAY_16( pixel, buf,[8*8] );
168 for( int y = 0; y < i_lines; y += 8, pixoff = y*i_stride )
169 for( int x = 0; x < i_width; x += 8, i_mb++, pixoff += 8)
171 w->weightfn[8>>2]( buf, 8, &src[pixoff], i_stride, w, 8 );
172 cost += X264_MIN( h->pixf.mbcmp[PIXEL_8x8]( buf, 8, &fenc_plane[pixoff], i_stride ), fenc->i_intra_cost[i_mb] );
174 cost += x264_weight_slice_header_cost( h, w, 0 );
177 for( int y = 0; y < i_lines; y += 8, pixoff = y*i_stride )
178 for( int x = 0; x < i_width; x += 8, i_mb++, pixoff += 8 )
179 cost += X264_MIN( h->pixf.mbcmp[PIXEL_8x8]( &src[pixoff], i_stride, &fenc_plane[pixoff], i_stride ), fenc->i_intra_cost[i_mb] );
184 static NOINLINE unsigned int x264_weight_cost_chroma( x264_t *h, x264_frame_t *fenc, pixel *ref, x264_weight_t *w )
186 unsigned int cost = 0;
187 int i_stride = fenc->i_stride[1];
188 int i_offset = i_stride / 2;
189 int i_lines = fenc->i_lines[1];
190 int i_width = fenc->i_width[1];
191 pixel *src = ref + i_offset;
192 ALIGNED_ARRAY_16( pixel, buf, [8*8] );
194 ALIGNED_16( static pixel flat[9] ) = {0,0,0,0,0,0,0,0,1}; //hack for win32
197 for( int y = 0; y < i_lines; y += 8, pixoff = y*i_stride )
198 for( int x = 0; x < i_width; x += 8, pixoff += 8 )
200 w->weightfn[8>>2]( buf, 8, &ref[pixoff], i_stride, w, 8 );
201 /* The naive and seemingly sensible algorithm is to use mbcmp as in luma.
202 * But testing shows that for chroma the DC coefficient is by far the most
203 * important part of the coding cost. Thus a more useful chroma weight is
204 * obtained by comparing each block's DC coefficient instead of the actual
207 * FIXME: add a (faster) asm sum function to replace sad. */
208 cost += abs( h->pixf.sad_aligned[PIXEL_8x8]( buf, 8, flat, 0 ) -
209 h->pixf.sad_aligned[PIXEL_8x8]( &src[pixoff], i_stride, flat, 0 ) );
211 cost += x264_weight_slice_header_cost( h, w, 1 );
214 for( int y = 0; y < i_lines; y += 8, pixoff = y*i_stride )
215 for( int x = 0; x < i_width; x += 8, pixoff += 8 )
216 cost += abs( h->pixf.sad_aligned[PIXEL_8x8]( &ref[pixoff], i_stride, flat, 0 ) -
217 h->pixf.sad_aligned[PIXEL_8x8]( &src[pixoff], i_stride, flat, 0 ) );
222 void x264_weights_analyse( x264_t *h, x264_frame_t *fenc, x264_frame_t *ref, int b_lookahead )
224 int i_delta_index = fenc->i_frame - ref->i_frame - 1;
225 /* epsilon is chosen to require at least a numerator of 127 (with denominator = 128) */
226 const float epsilon = 1.f/128.f;
227 x264_weight_t *weights = fenc->weight[0];
228 SET_WEIGHT( weights[0], 0, 1, 0, 0 );
229 SET_WEIGHT( weights[1], 0, 1, 0, 0 );
230 SET_WEIGHT( weights[2], 0, 1, 0, 0 );
231 /* Don't check chroma in lookahead, or if there wasn't a luma weight. */
232 for( int plane = 0; plane <= 2 && !( plane && ( !weights[0].weightfn || b_lookahead ) ); plane++ )
235 int minoff, minscale, mindenom;
236 unsigned int minscore, origscore;
238 float fenc_var = fenc->i_pixel_ssd[plane] + !ref->i_pixel_ssd[plane];
239 float ref_var = ref->i_pixel_ssd[plane] + !ref->i_pixel_ssd[plane];
240 float guess_scale = sqrtf( fenc_var / ref_var );
241 float fenc_mean = (float)fenc->i_pixel_sum[plane] / (fenc->i_lines[!!plane] * fenc->i_width[!!plane]);
242 float ref_mean = (float) ref->i_pixel_sum[plane] / (fenc->i_lines[!!plane] * fenc->i_width[!!plane]);
245 if( fabsf( ref_mean - fenc_mean ) < 0.5f && fabsf( 1.f - guess_scale ) < epsilon )
247 SET_WEIGHT( weights[plane], 0, 1, 0, 0 );
253 weights[plane].i_denom = 6;
254 weights[plane].i_scale = x264_clip3( round( guess_scale * 64 ), 0, 255 );
255 if( weights[plane].i_scale > 127 )
257 weights[1].weightfn = weights[2].weightfn = NULL;
262 x264_weight_get_h264( round( guess_scale * 128 ), 0, &weights[plane] );
265 mindenom = weights[plane].i_denom;
266 minscale = weights[plane].i_scale;
272 if( !fenc->b_intra_calculated )
274 x264_mb_analysis_t a;
275 x264_lowres_context_init( h, &a );
276 x264_slicetype_frame_cost( h, &a, &fenc, 0, 0, 0, 0 );
278 mcbuf = x264_weight_cost_init_luma( h, fenc, ref, h->mb.p_weight_buf[0] );
279 origscore = minscore = x264_weight_cost_luma( h, fenc, mcbuf, NULL );
283 pixel *dstu = h->mb.p_weight_buf[0];
284 pixel *dstv = h->mb.p_weight_buf[0]+fenc->i_stride[1]*fenc->i_lines[1];
285 /* Only initialize chroma data once. */
287 x264_weight_cost_init_chroma( h, fenc, ref, dstu, dstv );
288 mcbuf = plane == 1 ? dstu : dstv;
289 origscore = minscore = x264_weight_cost_chroma( h, fenc, mcbuf, NULL );
295 // This gives a slight improvement due to rounding errors but only tests
296 // one offset on lookahead.
297 // TODO: currently searches only offset +1. try other offsets/multipliers/combinations thereof?
298 offset_search = x264_clip3( fenc_mean - ref_mean * minscale / (1 << mindenom) + 0.5f * b_lookahead, -128, 126 );
299 for( int i_off = offset_search; i_off <= offset_search+!b_lookahead; i_off++ )
301 SET_WEIGHT( weights[plane], 1, minscale, mindenom, i_off );
304 s = x264_weight_cost_chroma( h, fenc, mcbuf, &weights[plane] );
306 s = x264_weight_cost_luma( h, fenc, mcbuf, &weights[plane] );
307 COPY3_IF_LT( minscore, s, minoff, i_off, found, 1 );
311 /* FIXME: More analysis can be done here on SAD vs. SATD termination. */
312 /* 0.2% termination derived experimentally to avoid weird weights in frames that are mostly intra. */
313 if( !found || (minscale == 1 << mindenom && minoff == 0) || (float)minscore / origscore > 0.998f )
315 SET_WEIGHT( weights[plane], 0, 1, 0, 0 );
319 SET_WEIGHT( weights[plane], 1, minscale, mindenom, minoff );
321 if( h->param.analyse.i_weighted_pred == X264_WEIGHTP_FAKE && weights[0].weightfn && !plane )
322 fenc->f_weighted_cost_delta[i_delta_index] = (float)minscore / origscore;
325 //FIXME, what is the correct way to deal with this?
326 if( weights[1].weightfn && weights[2].weightfn && weights[1].i_denom != weights[2].i_denom )
328 int denom = X264_MIN( weights[1].i_denom, weights[2].i_denom );
330 for( i = 1; i <= 2; i++ )
332 weights[i].i_scale = x264_clip3( weights[i].i_scale >> ( weights[i].i_denom - denom ), 0, 255 );
333 weights[i].i_denom = denom;
334 h->mc.weight_cache( h, &weights[i] );
338 if( weights[0].weightfn && b_lookahead )
340 //scale lowres in lookahead for slicetype_frame_cost
341 pixel *src = ref->buffer_lowres[0];
342 pixel *dst = h->mb.p_weight_buf[0];
343 int width = ref->i_width_lowres + PADH*2;
344 int height = ref->i_lines_lowres + PADV*2;
345 x264_weight_scale_plane( h, dst, ref->i_stride_lowres, src, ref->i_stride_lowres,
346 width, height, &weights[0] );
347 fenc->weighted[0] = h->mb.p_weight_buf[0] + PADH + ref->i_stride_lowres * PADV;
351 static void x264_slicetype_mb_cost( x264_t *h, x264_mb_analysis_t *a,
352 x264_frame_t **frames, int p0, int p1, int b,
353 int dist_scale_factor, int do_search[2], const x264_weight_t *w )
355 x264_frame_t *fref0 = frames[p0];
356 x264_frame_t *fref1 = frames[p1];
357 x264_frame_t *fenc = frames[b];
358 const int b_bidir = (b < p1);
359 const int i_mb_x = h->mb.i_mb_x;
360 const int i_mb_y = h->mb.i_mb_y;
361 const int i_mb_stride = h->mb.i_mb_width;
362 const int i_mb_xy = i_mb_x + i_mb_y * i_mb_stride;
363 const int i_stride = fenc->i_stride_lowres;
364 const int i_pel_offset = 8 * (i_mb_x + i_mb_y * i_stride);
365 const int i_bipred_weight = h->param.analyse.b_weighted_bipred ? 64 - (dist_scale_factor>>2) : 32;
366 int16_t (*fenc_mvs[2])[2] = { &frames[b]->lowres_mvs[0][b-p0-1][i_mb_xy], &frames[b]->lowres_mvs[1][p1-b-1][i_mb_xy] };
367 int (*fenc_costs[2]) = { &frames[b]->lowres_mv_costs[0][b-p0-1][i_mb_xy], &frames[b]->lowres_mv_costs[1][p1-b-1][i_mb_xy] };
368 int b_frame_score_mb = (i_mb_x > 0 && i_mb_x < h->mb.i_mb_width - 1 &&
369 i_mb_y > 0 && i_mb_y < h->mb.i_mb_height - 1) ||
370 h->mb.i_mb_width <= 2 || h->mb.i_mb_height <= 2;
372 ALIGNED_ARRAY_16( pixel, pix1,[9*FDEC_STRIDE] );
373 pixel *pix2 = pix1+8;
375 int i_bcost = COST_MAX;
378 h->mb.pic.p_fenc[0] = h->mb.pic.fenc_buf;
379 h->mc.copy[PIXEL_8x8]( h->mb.pic.p_fenc[0], FENC_STRIDE, &fenc->lowres[0][i_pel_offset], i_stride, 8 );
382 goto lowres_intra_mb;
384 // no need for h->mb.mv_min[]
385 h->mb.mv_min_fpel[0] = -8*h->mb.i_mb_x - 4;
386 h->mb.mv_max_fpel[0] = 8*( h->mb.i_mb_width - h->mb.i_mb_x - 1 ) + 4;
387 h->mb.mv_min_spel[0] = 4*( h->mb.mv_min_fpel[0] - 8 );
388 h->mb.mv_max_spel[0] = 4*( h->mb.mv_max_fpel[0] + 8 );
389 if( h->mb.i_mb_x >= h->mb.i_mb_width - 2 )
391 h->mb.mv_min_fpel[1] = -8*h->mb.i_mb_y - 4;
392 h->mb.mv_max_fpel[1] = 8*( h->mb.i_mb_height - h->mb.i_mb_y - 1 ) + 4;
393 h->mb.mv_min_spel[1] = 4*( h->mb.mv_min_fpel[1] - 8 );
394 h->mb.mv_max_spel[1] = 4*( h->mb.mv_max_fpel[1] + 8 );
397 #define LOAD_HPELS_LUMA(dst, src) \
399 (dst)[0] = &(src)[0][i_pel_offset]; \
400 (dst)[1] = &(src)[1][i_pel_offset]; \
401 (dst)[2] = &(src)[2][i_pel_offset]; \
402 (dst)[3] = &(src)[3][i_pel_offset]; \
404 #define LOAD_WPELS_LUMA(dst,src) \
405 (dst) = &(src)[i_pel_offset];
407 #define CLIP_MV( mv ) \
409 mv[0] = x264_clip3( mv[0], h->mb.mv_min_spel[0], h->mb.mv_max_spel[0] ); \
410 mv[1] = x264_clip3( mv[1], h->mb.mv_min_spel[1], h->mb.mv_max_spel[1] ); \
412 #define TRY_BIDIR( mv0, mv1, penalty ) \
415 if( h->param.analyse.i_subpel_refine <= 1 ) \
417 int hpel_idx1 = (((mv0)[0]&2)>>1) + ((mv0)[1]&2); \
418 int hpel_idx2 = (((mv1)[0]&2)>>1) + ((mv1)[1]&2); \
419 pixel *src1 = m[0].p_fref[hpel_idx1] + ((mv0)[0]>>2) + ((mv0)[1]>>2) * m[0].i_stride[0]; \
420 pixel *src2 = m[1].p_fref[hpel_idx2] + ((mv1)[0]>>2) + ((mv1)[1]>>2) * m[1].i_stride[0]; \
421 h->mc.avg[PIXEL_8x8]( pix1, 16, src1, m[0].i_stride[0], src2, m[1].i_stride[0], i_bipred_weight ); \
425 int stride1 = 16, stride2 = 16; \
426 pixel *src1, *src2; \
427 src1 = h->mc.get_ref( pix1, &stride1, m[0].p_fref, m[0].i_stride[0], \
428 (mv0)[0], (mv0)[1], 8, 8, w ); \
429 src2 = h->mc.get_ref( pix2, &stride2, m[1].p_fref, m[1].i_stride[0], \
430 (mv1)[0], (mv1)[1], 8, 8, w ); \
431 h->mc.avg[PIXEL_8x8]( pix1, 16, src1, stride1, src2, stride2, i_bipred_weight ); \
433 i_cost = penalty * a->i_lambda + h->pixf.mbcmp[PIXEL_8x8]( \
434 m[0].p_fenc[0], FENC_STRIDE, pix1, 16 ); \
435 COPY2_IF_LT( i_bcost, i_cost, list_used, 3 ); \
438 m[0].i_pixel = PIXEL_8x8;
439 m[0].p_cost_mv = a->p_cost_mv;
440 m[0].i_stride[0] = i_stride;
441 m[0].p_fenc[0] = h->mb.pic.p_fenc[0];
444 LOAD_HPELS_LUMA( m[0].p_fref, fref0->lowres );
445 m[0].p_fref_w = m[0].p_fref[0];
447 LOAD_WPELS_LUMA( m[0].p_fref_w, fenc->weighted[0] );
451 int16_t *mvr = fref1->lowres_mvs[0][p1-p0-1][i_mb_xy];
452 ALIGNED_ARRAY_8( int16_t, dmv,[2],[2] );
454 m[1].i_pixel = PIXEL_8x8;
455 m[1].p_cost_mv = a->p_cost_mv;
456 m[1].i_stride[0] = i_stride;
457 m[1].p_fenc[0] = h->mb.pic.p_fenc[0];
459 m[1].weight = weight_none;
460 LOAD_HPELS_LUMA( m[1].p_fref, fref1->lowres );
461 m[1].p_fref_w = m[1].p_fref[0];
463 dmv[0][0] = ( mvr[0] * dist_scale_factor + 128 ) >> 8;
464 dmv[0][1] = ( mvr[1] * dist_scale_factor + 128 ) >> 8;
465 dmv[1][0] = dmv[0][0] - mvr[0];
466 dmv[1][1] = dmv[0][1] - mvr[1];
469 if( h->param.analyse.i_subpel_refine <= 1 )
470 M64( dmv ) &= ~0x0001000100010001ULL; /* mv & ~1 */
472 TRY_BIDIR( dmv[0], dmv[1], 0 );
476 h->mc.avg[PIXEL_8x8]( pix1, 16, m[0].p_fref[0], m[0].i_stride[0], m[1].p_fref[0], m[1].i_stride[0], i_bipred_weight );
477 i_cost = h->pixf.mbcmp[PIXEL_8x8]( m[0].p_fenc[0], FENC_STRIDE, pix1, 16 );
478 COPY2_IF_LT( i_bcost, i_cost, list_used, 3 );
482 for( int l = 0; l < 1 + b_bidir; l++ )
487 int16_t (*fenc_mv)[2] = fenc_mvs[l];
488 ALIGNED_4( int16_t mvc[4][2] );
490 /* Reverse-order MV prediction. */
493 #define MVC(mv) { CP32( mvc[i_mvc], mv ); i_mvc++; }
494 if( i_mb_x < h->mb.i_mb_width - 1 )
496 if( i_mb_y < h->mb.i_mb_height - 1 )
498 MVC( fenc_mv[i_mb_stride] );
500 MVC( fenc_mv[i_mb_stride-1] );
501 if( i_mb_x < h->mb.i_mb_width - 1 )
502 MVC( fenc_mv[i_mb_stride+1] );
506 CP32( m[l].mvp, mvc[0] );
508 x264_median_mv( m[l].mvp, mvc[0], mvc[1], mvc[2] );
510 /* Fast skip for cases of near-zero residual. Shortcut: don't bother except in the mv0 case,
511 * since anything else is likely to have enough residual to not trigger the skip. */
512 if( !M32( m[l].mvp ) )
514 m[l].cost = h->pixf.mbcmp[PIXEL_8x8]( m[l].p_fenc[0], FENC_STRIDE, m[l].p_fref[0], m[l].i_stride[0] );
522 x264_me_search( h, &m[l], mvc, i_mvc );
523 m[l].cost -= 2 * a->i_lambda; // remove mvcost from skip mbs
525 m[l].cost += 5 * a->i_lambda;
528 CP32( fenc_mvs[l], m[l].mv );
529 *fenc_costs[l] = m[l].cost;
533 CP32( m[l].mv, fenc_mvs[l] );
534 m[l].cost = *fenc_costs[l];
536 COPY2_IF_LT( i_bcost, m[l].cost, list_used, l+1 );
539 if( b_bidir && ( M32( m[0].mv ) || M32( m[1].mv ) ) )
540 TRY_BIDIR( m[0].mv, m[1].mv, 5 );
543 if( !fenc->b_intra_calculated )
545 ALIGNED_ARRAY_16( pixel, edge,[33] );
546 pixel *pix = &pix1[8+FDEC_STRIDE - 1];
547 pixel *src = &fenc->lowres[0][i_pel_offset - 1];
548 const int intra_penalty = 5 * a->i_lambda;
551 memcpy( pix-FDEC_STRIDE, src-i_stride, 17 * sizeof(pixel) );
552 for( int i = 0; i < 8; i++ )
553 pix[i*FDEC_STRIDE] = src[i*i_stride];
556 h->pixf.intra_mbcmp_x3_8x8c( h->mb.pic.p_fenc[0], pix, satds );
557 int i_icost = X264_MIN3( satds[0], satds[1], satds[2] );
559 if( h->param.analyse.i_subpel_refine > 1 )
561 h->predict_8x8c[I_PRED_CHROMA_P]( pix );
562 int satd = h->pixf.mbcmp[PIXEL_8x8]( pix, FDEC_STRIDE, h->mb.pic.p_fenc[0], FENC_STRIDE );
563 i_icost = X264_MIN( i_icost, satd );
564 h->predict_8x8_filter( pix, edge, ALL_NEIGHBORS, ALL_NEIGHBORS );
565 for( int i = 3; i < 9; i++ )
567 h->predict_8x8[i]( pix, edge );
568 satd = h->pixf.mbcmp[PIXEL_8x8]( pix, FDEC_STRIDE, h->mb.pic.p_fenc[0], FENC_STRIDE );
569 i_icost = X264_MIN( i_icost, satd );
573 i_icost += intra_penalty;
574 fenc->i_intra_cost[i_mb_xy] = i_icost;
575 if( b_frame_score_mb )
577 int *row_satd_intra = frames[b]->i_row_satds[0][0];
578 int i_icost_aq = i_icost;
579 if( h->param.rc.i_aq_mode )
580 i_icost_aq = (i_icost_aq * frames[b]->i_inv_qscale_factor[i_mb_xy] + 128) >> 8;
581 fenc->i_cost_est[0][0] += i_icost;
582 fenc->i_cost_est_aq[0][0] += i_icost_aq;
583 row_satd_intra[h->mb.i_mb_y] += i_icost_aq;
587 /* forbid intra-mbs in B-frames, because it's rare and not worth checking */
588 /* FIXME: Should we still forbid them now that we cache intra scores? */
591 int i_icost = fenc->i_intra_cost[i_mb_xy];
592 int b_intra = i_icost < i_bcost;
598 if( b_frame_score_mb )
599 fenc->i_intra_mbs[b-p0] += b_intra;
602 /* In an I-frame, we've already added the results above in the intra section. */
605 int i_bcost_aq = i_bcost;
606 if( h->param.rc.i_aq_mode )
607 i_bcost_aq = (i_bcost_aq * frames[b]->i_inv_qscale_factor[i_mb_xy] + 128) >> 8;
608 fenc->i_row_satds[b-p0][p1-b][h->mb.i_mb_y] += i_bcost_aq;
609 if( b_frame_score_mb )
611 /* Don't use AQ-weighted costs for slicetype decision, only for ratecontrol. */
612 frames[b]->i_cost_est[b-p0][p1-b] += i_bcost;
613 frames[b]->i_cost_est_aq[b-p0][p1-b] += i_bcost_aq;
617 fenc->lowres_costs[b-p0][p1-b][i_mb_xy] = X264_MIN( i_bcost, LOWRES_COST_MASK ) + (list_used << LOWRES_COST_SHIFT);
622 (h->mb.i_mb_width > 2 && h->mb.i_mb_height > 2 ?\
623 (h->mb.i_mb_width - 2) * (h->mb.i_mb_height - 2) :\
624 h->mb.i_mb_width * h->mb.i_mb_height)
626 static int x264_slicetype_frame_cost( x264_t *h, x264_mb_analysis_t *a,
627 x264_frame_t **frames, int p0, int p1, int b,
628 int b_intra_penalty )
632 const x264_weight_t *w = weight_none;
633 /* Check whether we already evaluated this frame
634 * If we have tried this frame as P, then we have also tried
635 * the preceding frames as B. (is this still true?) */
636 /* Also check that we already calculated the row SATDs for the current frame. */
637 if( frames[b]->i_cost_est[b-p0][p1-b] >= 0 && (!h->param.rc.i_vbv_buffer_size || frames[b]->i_row_satds[b-p0][p1-b][0] != -1) )
638 i_score = frames[b]->i_cost_est[b-p0][p1-b];
641 int dist_scale_factor = 128;
642 int *row_satd = frames[b]->i_row_satds[b-p0][p1-b];
643 int *row_satd_intra = frames[b]->i_row_satds[0][0];
645 /* For each list, check to see whether we have lowres motion-searched this reference frame before. */
646 do_search[0] = b != p0 && frames[b]->lowres_mvs[0][b-p0-1][0][0] == 0x7FFF;
647 do_search[1] = b != p1 && frames[b]->lowres_mvs[1][p1-b-1][0][0] == 0x7FFF;
650 if( h->param.analyse.i_weighted_pred && b == p1 )
653 x264_weights_analyse( h, frames[b], frames[p0], 1 );
654 w = frames[b]->weight[0];
656 frames[b]->lowres_mvs[0][b-p0-1][0][0] = 0;
658 if( do_search[1] ) frames[b]->lowres_mvs[1][p1-b-1][0][0] = 0;
661 frames[b]->i_intra_mbs[b-p0] = 0;
662 if( !frames[b]->b_intra_calculated )
664 frames[b]->i_cost_est[0][0] = 0;
665 frames[b]->i_cost_est_aq[0][0] = 0;
668 dist_scale_factor = ( ((b-p0) << 8) + ((p1-p0) >> 1) ) / (p1-p0);
670 frames[b]->i_cost_est[b-p0][p1-b] = 0;
671 frames[b]->i_cost_est_aq[b-p0][p1-b] = 0;
673 /* Lowres lookahead goes backwards because the MVs are used as predictors in the main encode.
674 * This considerably improves MV prediction overall. */
676 /* The edge mbs seem to reduce the predictive quality of the
677 * whole frame's score, but are needed for a spatial distribution. */
678 if( h->param.rc.b_mb_tree || h->param.rc.i_vbv_buffer_size ||
679 h->mb.i_mb_width <= 2 || h->mb.i_mb_height <= 2 )
681 for( h->mb.i_mb_y = h->mb.i_mb_height - 1; h->mb.i_mb_y >= 0; h->mb.i_mb_y-- )
683 row_satd[h->mb.i_mb_y] = 0;
684 if( !frames[b]->b_intra_calculated )
685 row_satd_intra[h->mb.i_mb_y] = 0;
686 for( h->mb.i_mb_x = h->mb.i_mb_width - 1; h->mb.i_mb_x >= 0; h->mb.i_mb_x-- )
687 x264_slicetype_mb_cost( h, a, frames, p0, p1, b, dist_scale_factor, do_search, w );
692 for( h->mb.i_mb_y = h->mb.i_mb_height - 2; h->mb.i_mb_y >= 1; h->mb.i_mb_y-- )
693 for( h->mb.i_mb_x = h->mb.i_mb_width - 2; h->mb.i_mb_x >= 1; h->mb.i_mb_x-- )
694 x264_slicetype_mb_cost( h, a, frames, p0, p1, b, dist_scale_factor, do_search, w );
697 i_score = frames[b]->i_cost_est[b-p0][p1-b];
699 i_score = (uint64_t)i_score * 100 / (120 + h->param.i_bframe_bias);
701 frames[b]->b_intra_calculated = 1;
703 frames[b]->i_cost_est[b-p0][p1-b] = i_score;
707 if( b_intra_penalty )
709 // arbitrary penalty for I-blocks after B-frames
711 i_score += i_score * frames[b]->i_intra_mbs[b-p0] / (nmb * 8);
716 /* If MB-tree changes the quantizers, we need to recalculate the frame cost without
717 * re-running lookahead. */
718 static int x264_slicetype_frame_cost_recalculate( x264_t *h, x264_frame_t **frames, int p0, int p1, int b )
721 int *row_satd = frames[b]->i_row_satds[b-p0][p1-b];
722 float *qp_offset = IS_X264_TYPE_B(frames[b]->i_type) ? frames[b]->f_qp_offset_aq : frames[b]->f_qp_offset;
724 for( h->mb.i_mb_y = h->mb.i_mb_height - 1; h->mb.i_mb_y >= 0; h->mb.i_mb_y-- )
726 row_satd[ h->mb.i_mb_y ] = 0;
727 for( h->mb.i_mb_x = h->mb.i_mb_width - 1; h->mb.i_mb_x >= 0; h->mb.i_mb_x-- )
729 int i_mb_xy = h->mb.i_mb_x + h->mb.i_mb_y*h->mb.i_mb_stride;
730 int i_mb_cost = frames[b]->lowres_costs[b-p0][p1-b][i_mb_xy] & LOWRES_COST_MASK;
731 float qp_adj = qp_offset[i_mb_xy];
732 i_mb_cost = (i_mb_cost * x264_exp2fix8(qp_adj) + 128) >> 8;
733 row_satd[ h->mb.i_mb_y ] += i_mb_cost;
734 if( (h->mb.i_mb_y > 0 && h->mb.i_mb_y < h->mb.i_mb_height - 1 &&
735 h->mb.i_mb_x > 0 && h->mb.i_mb_x < h->mb.i_mb_width - 1) ||
736 h->mb.i_mb_width <= 2 || h->mb.i_mb_height <= 2 )
738 i_score += i_mb_cost;
745 static void x264_macroblock_tree_finish( x264_t *h, x264_frame_t *frame, int ref0_distance )
748 float weightdelta = 0.0;
749 if( ref0_distance && frame->f_weighted_cost_delta[ref0_distance-1] > 0 )
750 weightdelta = (1.0 - frame->f_weighted_cost_delta[ref0_distance-1]);
752 /* Allow the strength to be adjusted via qcompress, since the two
753 * concepts are very similar. */
754 float strength = 5.0f * (1.0f - h->param.rc.f_qcompress);
755 for( int mb_index = 0; mb_index < h->mb.i_mb_count; mb_index++ )
757 int intra_cost = (frame->i_intra_cost[mb_index] * frame->i_inv_qscale_factor[mb_index]+128)>>8;
760 int propagate_cost = frame->i_propagate_cost[mb_index];
761 float log2_ratio = x264_log2(intra_cost + propagate_cost) - x264_log2(intra_cost) + weightdelta;
762 frame->f_qp_offset[mb_index] = frame->f_qp_offset_aq[mb_index] - strength * log2_ratio;
767 static void x264_macroblock_tree_propagate( x264_t *h, x264_frame_t **frames, int p0, int p1, int b, int referenced )
769 uint16_t *ref_costs[2] = {frames[p0]->i_propagate_cost,frames[p1]->i_propagate_cost};
770 int dist_scale_factor = ( ((b-p0) << 8) + ((p1-p0) >> 1) ) / (p1-p0);
771 int i_bipred_weight = h->param.analyse.b_weighted_bipred ? 64 - (dist_scale_factor>>2) : 32;
772 int16_t (*mvs[2])[2] = { frames[b]->lowres_mvs[0][b-p0-1], frames[b]->lowres_mvs[1][p1-b-1] };
773 int bipred_weights[2] = {i_bipred_weight, 64 - i_bipred_weight};
774 int *buf = h->scratch_buffer;
775 uint16_t *propagate_cost = frames[b]->i_propagate_cost;
777 /* For non-reffed frames the source costs are always zero, so just memset one row and re-use it. */
779 memset( frames[b]->i_propagate_cost, 0, h->mb.i_mb_width * sizeof(uint16_t) );
781 for( h->mb.i_mb_y = 0; h->mb.i_mb_y < h->mb.i_mb_height; h->mb.i_mb_y++ )
783 int mb_index = h->mb.i_mb_y*h->mb.i_mb_stride;
784 h->mc.mbtree_propagate_cost( buf, propagate_cost,
785 frames[b]->i_intra_cost+mb_index, frames[b]->lowres_costs[b-p0][p1-b]+mb_index,
786 frames[b]->i_inv_qscale_factor+mb_index, h->mb.i_mb_width );
788 propagate_cost += h->mb.i_mb_width;
789 for( h->mb.i_mb_x = 0; h->mb.i_mb_x < h->mb.i_mb_width; h->mb.i_mb_x++, mb_index++ )
791 int propagate_amount = buf[h->mb.i_mb_x];
792 /* Don't propagate for an intra block. */
793 if( propagate_amount > 0 )
795 /* Access width-2 bitfield. */
796 int lists_used = frames[b]->lowres_costs[b-p0][p1-b][mb_index] >> LOWRES_COST_SHIFT;
797 /* Follow the MVs to the previous frame(s). */
798 for( int list = 0; list < 2; list++ )
799 if( (lists_used >> list)&1 )
801 #define CLIP_ADD(s,x) (s) = X264_MIN((s)+(x),(1<<16)-1)
802 int listamount = propagate_amount;
803 /* Apply bipred weighting. */
804 if( lists_used == 3 )
805 listamount = (listamount * bipred_weights[list] + 32) >> 6;
807 /* Early termination for simple case of mv0. */
808 if( !M32( mvs[list][mb_index] ) )
810 CLIP_ADD( ref_costs[list][mb_index], listamount );
814 int x = mvs[list][mb_index][0];
815 int y = mvs[list][mb_index][1];
816 int mbx = (x>>5)+h->mb.i_mb_x;
817 int mby = (y>>5)+h->mb.i_mb_y;
818 int idx0 = mbx + mby * h->mb.i_mb_stride;
820 int idx2 = idx0 + h->mb.i_mb_stride;
821 int idx3 = idx0 + h->mb.i_mb_stride + 1;
824 int idx0weight = (32-y)*(32-x);
825 int idx1weight = (32-y)*x;
826 int idx2weight = y*(32-x);
827 int idx3weight = y*x;
829 /* We could just clip the MVs, but pixels that lie outside the frame probably shouldn't
831 if( mbx < h->mb.i_mb_width-1 && mby < h->mb.i_mb_height-1 && mbx >= 0 && mby >= 0 )
833 CLIP_ADD( ref_costs[list][idx0], (listamount*idx0weight+512)>>10 );
834 CLIP_ADD( ref_costs[list][idx1], (listamount*idx1weight+512)>>10 );
835 CLIP_ADD( ref_costs[list][idx2], (listamount*idx2weight+512)>>10 );
836 CLIP_ADD( ref_costs[list][idx3], (listamount*idx3weight+512)>>10 );
838 else /* Check offsets individually */
840 if( mbx < h->mb.i_mb_width && mby < h->mb.i_mb_height && mbx >= 0 && mby >= 0 )
841 CLIP_ADD( ref_costs[list][idx0], (listamount*idx0weight+512)>>10 );
842 if( mbx+1 < h->mb.i_mb_width && mby < h->mb.i_mb_height && mbx+1 >= 0 && mby >= 0 )
843 CLIP_ADD( ref_costs[list][idx1], (listamount*idx1weight+512)>>10 );
844 if( mbx < h->mb.i_mb_width && mby+1 < h->mb.i_mb_height && mbx >= 0 && mby+1 >= 0 )
845 CLIP_ADD( ref_costs[list][idx2], (listamount*idx2weight+512)>>10 );
846 if( mbx+1 < h->mb.i_mb_width && mby+1 < h->mb.i_mb_height && mbx+1 >= 0 && mby+1 >= 0 )
847 CLIP_ADD( ref_costs[list][idx3], (listamount*idx3weight+512)>>10 );
854 if( h->param.rc.i_vbv_buffer_size && h->param.rc.i_lookahead && referenced )
855 x264_macroblock_tree_finish( h, frames[b], b == p1 ? b - p0 : 0 );
858 static void x264_macroblock_tree( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int num_frames, int b_intra )
861 int last_nonb, cur_nonb = 1;
866 x264_slicetype_frame_cost( h, a, frames, 0, 0, 0, 0 );
868 while( i > 0 && frames[i]->i_type == X264_TYPE_B )
872 /* Lookaheadless MB-tree is not a theoretically distinct case; the same extrapolation could
873 * be applied to the end of a lookahead buffer of any size. However, it's most needed when
874 * lookahead=0, so that's what's currently implemented. */
875 if( !h->param.rc.i_lookahead )
879 memset( frames[0]->i_propagate_cost, 0, h->mb.i_mb_count * sizeof(uint16_t) );
880 memcpy( frames[0]->f_qp_offset, frames[0]->f_qp_offset_aq, h->mb.i_mb_count * sizeof(float) );
883 XCHG( uint16_t*, frames[last_nonb]->i_propagate_cost, frames[0]->i_propagate_cost );
884 memset( frames[0]->i_propagate_cost, 0, h->mb.i_mb_count * sizeof(uint16_t) );
888 if( last_nonb < idx )
890 memset( frames[last_nonb]->i_propagate_cost, 0, h->mb.i_mb_count * sizeof(uint16_t) );
896 while( frames[cur_nonb]->i_type == X264_TYPE_B && cur_nonb > 0 )
900 x264_slicetype_frame_cost( h, a, frames, cur_nonb, last_nonb, last_nonb, 0 );
901 memset( frames[cur_nonb]->i_propagate_cost, 0, h->mb.i_mb_count * sizeof(uint16_t) );
902 bframes = last_nonb - cur_nonb - 1;
903 if( h->param.i_bframe_pyramid && bframes > 1 )
905 int middle = (bframes + 1)/2 + cur_nonb;
906 x264_slicetype_frame_cost( h, a, frames, cur_nonb, last_nonb, middle, 0 );
907 memset( frames[middle]->i_propagate_cost, 0, h->mb.i_mb_count * sizeof(uint16_t) );
908 while( i > cur_nonb )
910 int p0 = i > middle ? middle : cur_nonb;
911 int p1 = i < middle ? middle : last_nonb;
914 x264_slicetype_frame_cost( h, a, frames, p0, p1, i, 0 );
915 x264_macroblock_tree_propagate( h, frames, p0, p1, i, 0 );
919 x264_macroblock_tree_propagate( h, frames, cur_nonb, last_nonb, middle, 1 );
923 while( i > cur_nonb )
925 x264_slicetype_frame_cost( h, a, frames, cur_nonb, last_nonb, i, 0 );
926 x264_macroblock_tree_propagate( h, frames, cur_nonb, last_nonb, i, 0 );
930 x264_macroblock_tree_propagate( h, frames, cur_nonb, last_nonb, last_nonb, 1 );
931 last_nonb = cur_nonb;
934 if( !h->param.rc.i_lookahead )
936 x264_macroblock_tree_propagate( h, frames, 0, last_nonb, last_nonb, 1 );
937 XCHG( uint16_t*, frames[last_nonb]->i_propagate_cost, frames[0]->i_propagate_cost );
940 x264_macroblock_tree_finish( h, frames[last_nonb], last_nonb );
941 if( h->param.i_bframe_pyramid && bframes > 1 && !h->param.rc.i_vbv_buffer_size )
942 x264_macroblock_tree_finish( h, frames[last_nonb+(bframes+1)/2], 0 );
945 static int x264_vbv_frame_cost( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int p0, int p1, int b )
947 int cost = x264_slicetype_frame_cost( h, a, frames, p0, p1, b, 0 );
948 if( h->param.rc.i_aq_mode )
950 if( h->param.rc.b_mb_tree )
951 return x264_slicetype_frame_cost_recalculate( h, frames, p0, p1, b );
953 return frames[b]->i_cost_est_aq[b-p0][p1-b];
958 static void x264_calculate_durations( x264_t *h, x264_frame_t *cur_frame, x264_frame_t *prev_frame, int *i_cpb_delay, int *i_coded_fields )
960 cur_frame->i_cpb_delay = *i_cpb_delay;
961 cur_frame->i_dpb_output_delay = cur_frame->i_field_cnt - *i_coded_fields;
963 // add a correction term for frame reordering
964 cur_frame->i_dpb_output_delay += h->sps->vui.i_num_reorder_frames*2;
966 // fix possible negative dpb_output_delay because of pulldown changes and reordering
967 if( cur_frame->i_dpb_output_delay < 0 )
969 cur_frame->i_cpb_delay += cur_frame->i_dpb_output_delay;
970 cur_frame->i_dpb_output_delay = 0;
972 prev_frame->i_cpb_duration += cur_frame->i_dpb_output_delay;
975 if( cur_frame->b_keyframe )
978 *i_cpb_delay += cur_frame->i_duration;
979 *i_coded_fields += cur_frame->i_duration;
980 cur_frame->i_cpb_duration = cur_frame->i_duration;
983 static void x264_vbv_lookahead( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int num_frames, int keyframe )
985 int last_nonb = 0, cur_nonb = 1, idx = 0;
986 x264_frame_t *prev_frame = NULL;
987 int prev_frame_idx = 0;
988 while( cur_nonb < num_frames && frames[cur_nonb]->i_type == X264_TYPE_B )
990 int next_nonb = keyframe ? last_nonb : cur_nonb;
992 if( frames[cur_nonb]->i_coded_fields_lookahead >= 0 )
994 h->i_coded_fields_lookahead = frames[cur_nonb]->i_coded_fields_lookahead;
995 h->i_cpb_delay_lookahead = frames[cur_nonb]->i_cpb_delay_lookahead;
998 while( cur_nonb < num_frames )
1000 /* P/I cost: This shouldn't include the cost of next_nonb */
1001 if( next_nonb != cur_nonb )
1003 int p0 = IS_X264_TYPE_I( frames[cur_nonb]->i_type ) ? cur_nonb : last_nonb;
1004 frames[next_nonb]->i_planned_satd[idx] = x264_vbv_frame_cost( h, a, frames, p0, cur_nonb, cur_nonb );
1005 frames[next_nonb]->i_planned_type[idx] = frames[cur_nonb]->i_type;
1006 frames[cur_nonb]->i_coded_fields_lookahead = h->i_coded_fields_lookahead;
1007 frames[cur_nonb]->i_cpb_delay_lookahead = h->i_cpb_delay_lookahead;
1008 x264_calculate_durations( h, frames[cur_nonb], prev_frame, &h->i_cpb_delay_lookahead, &h->i_coded_fields_lookahead );
1011 frames[next_nonb]->f_planned_cpb_duration[prev_frame_idx] = (double)prev_frame->i_cpb_duration *
1012 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1014 frames[next_nonb]->f_planned_cpb_duration[idx] = (double)frames[cur_nonb]->i_cpb_duration *
1015 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1016 prev_frame = frames[cur_nonb];
1017 prev_frame_idx = idx;
1020 /* Handle the B-frames: coded order */
1021 for( int i = last_nonb+1; i < cur_nonb; i++, idx++ )
1023 frames[next_nonb]->i_planned_satd[idx] = x264_vbv_frame_cost( h, a, frames, last_nonb, cur_nonb, i );
1024 frames[next_nonb]->i_planned_type[idx] = X264_TYPE_B;
1025 frames[i]->i_coded_fields_lookahead = h->i_coded_fields_lookahead;
1026 frames[i]->i_cpb_delay_lookahead = h->i_cpb_delay_lookahead;
1027 x264_calculate_durations( h, frames[i], prev_frame, &h->i_cpb_delay_lookahead, &h->i_coded_fields_lookahead );
1030 frames[next_nonb]->f_planned_cpb_duration[prev_frame_idx] = (double)prev_frame->i_cpb_duration *
1031 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1033 frames[next_nonb]->f_planned_cpb_duration[idx] = (double)frames[i]->i_cpb_duration *
1034 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1035 prev_frame = frames[i];
1036 prev_frame_idx = idx;
1038 last_nonb = cur_nonb;
1040 while( cur_nonb <= num_frames && frames[cur_nonb]->i_type == X264_TYPE_B )
1043 frames[next_nonb]->i_planned_type[idx] = X264_TYPE_AUTO;
1046 static int x264_slicetype_path_cost( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, char *path, int threshold )
1051 path--; /* Since the 1st path element is really the second frame */
1055 /* Find the location of the next P-frame. */
1056 while( path[next_p] != 'P' )
1059 /* Add the cost of the P-frame found above */
1060 cost += x264_slicetype_frame_cost( h, a, frames, cur_p, next_p, next_p, 0 );
1061 /* Early terminate if the cost we have found is larger than the best path cost so far */
1062 if( cost > threshold )
1065 if( h->param.i_bframe_pyramid && next_p - cur_p > 2 )
1067 int middle = cur_p + (next_p - cur_p)/2;
1068 cost += x264_slicetype_frame_cost( h, a, frames, cur_p, next_p, middle, 0 );
1069 for( int next_b = loc; next_b < middle && cost < threshold; next_b++ )
1070 cost += x264_slicetype_frame_cost( h, a, frames, cur_p, middle, next_b, 0 );
1071 for( int next_b = middle+1; next_b < next_p && cost < threshold; next_b++ )
1072 cost += x264_slicetype_frame_cost( h, a, frames, middle, next_p, next_b, 0 );
1075 for( int next_b = loc; next_b < next_p && cost < threshold; next_b++ )
1076 cost += x264_slicetype_frame_cost( h, a, frames, cur_p, next_p, next_b, 0 );
1084 /* Viterbi/trellis slicetype decision algorithm. */
1085 /* Uses strings due to the fact that the speed of the control functions is
1086 negligible compared to the cost of running slicetype_frame_cost, and because
1087 it makes debugging easier. */
1088 static void x264_slicetype_path( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int length, char (*best_paths)[X264_LOOKAHEAD_MAX] )
1090 char paths[2][X264_LOOKAHEAD_MAX];
1091 int num_paths = X264_MIN( h->param.i_bframe+1, length );
1092 int best_cost = COST_MAX;
1095 /* Iterate over all currently possible paths */
1096 for( int path = 0; path < num_paths; path++ )
1098 /* Add suffixes to the current path */
1099 int len = length - (path + 1);
1100 memcpy( paths[idx], best_paths[len % (X264_BFRAME_MAX+1)], len );
1101 memset( paths[idx]+len, 'B', path );
1102 strcpy( paths[idx]+len+path, "P" );
1104 /* Calculate the actual cost of the current path */
1105 int cost = x264_slicetype_path_cost( h, a, frames, paths[idx], best_cost );
1106 if( cost < best_cost )
1113 /* Store the best path. */
1114 memcpy( best_paths[length % (X264_BFRAME_MAX+1)], paths[idx^1], length );
1117 static int scenecut_internal( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int p0, int p1, int print )
1119 x264_frame_t *frame = frames[p1];
1120 x264_slicetype_frame_cost( h, a, frames, p0, p1, p1, 0 );
1122 int icost = frame->i_cost_est[0][0];
1123 int pcost = frame->i_cost_est[p1-p0][0];
1125 int i_gop_size = frame->i_frame - h->lookahead->i_last_keyframe;
1126 float f_thresh_max = h->param.i_scenecut_threshold / 100.0;
1127 /* magic numbers pulled out of thin air */
1128 float f_thresh_min = f_thresh_max * 0.25;
1131 if( h->param.i_keyint_min == h->param.i_keyint_max )
1132 f_thresh_min = f_thresh_max;
1133 if( i_gop_size <= h->param.i_keyint_min / 4 || h->param.b_intra_refresh )
1134 f_bias = f_thresh_min / 4;
1135 else if( i_gop_size <= h->param.i_keyint_min )
1136 f_bias = f_thresh_min * i_gop_size / h->param.i_keyint_min;
1139 f_bias = f_thresh_min
1140 + ( f_thresh_max - f_thresh_min )
1141 * ( i_gop_size - h->param.i_keyint_min )
1142 / ( h->param.i_keyint_max - h->param.i_keyint_min );
1145 res = pcost >= (1.0 - f_bias) * icost;
1148 int imb = frame->i_intra_mbs[p1-p0];
1149 int pmb = NUM_MBS - imb;
1150 x264_log( h, X264_LOG_DEBUG, "scene cut at %d Icost:%d Pcost:%d ratio:%.4f bias:%.4f gop:%d (imb:%d pmb:%d)\n",
1152 icost, pcost, 1. - (double)pcost / icost,
1153 f_bias, i_gop_size, imb, pmb );
1158 static int scenecut( x264_t *h, x264_mb_analysis_t *a, x264_frame_t **frames, int p0, int p1, int real_scenecut, int num_frames, int i_max_search )
1160 /* Only do analysis during a normal scenecut check. */
1161 if( real_scenecut && h->param.i_bframe )
1163 int origmaxp1 = p0 + 1;
1164 /* Look ahead to avoid coding short flashes as scenecuts. */
1165 if( h->param.i_bframe_adaptive == X264_B_ADAPT_TRELLIS )
1166 /* Don't analyse any more frames than the trellis would have covered. */
1167 origmaxp1 += h->param.i_bframe;
1170 int maxp1 = X264_MIN( origmaxp1, num_frames );
1172 /* Where A and B are scenes: AAAAAABBBAAAAAA
1173 * If BBB is shorter than (maxp1-p0), it is detected as a flash
1174 * and not considered a scenecut. */
1175 for( int curp1 = p1; curp1 <= maxp1; curp1++ )
1176 if( !scenecut_internal( h, a, frames, p0, curp1, 0 ) )
1177 /* Any frame in between p0 and cur_p1 cannot be a real scenecut. */
1178 for( int i = curp1; i > p0; i-- )
1179 frames[i]->b_scenecut = 0;
1181 /* Where A-F are scenes: AAAAABBCCDDEEFFFFFF
1182 * If each of BB ... EE are shorter than (maxp1-p0), they are
1183 * detected as flashes and not considered scenecuts.
1184 * Instead, the first F frame becomes a scenecut.
1185 * If the video ends before F, no frame becomes a scenecut. */
1186 for( int curp0 = p0; curp0 <= maxp1; curp0++ )
1187 if( origmaxp1 > i_max_search || (curp0 < maxp1 && scenecut_internal( h, a, frames, curp0, maxp1, 0 )) )
1188 /* If cur_p0 is the p0 of a scenecut, it cannot be the p1 of a scenecut. */
1189 frames[curp0]->b_scenecut = 0;
1192 /* Ignore frames that are part of a flash, i.e. cannot be real scenecuts. */
1193 if( !frames[p1]->b_scenecut )
1195 return scenecut_internal( h, a, frames, p0, p1, real_scenecut );
1198 void x264_slicetype_analyse( x264_t *h, int keyframe )
1200 x264_mb_analysis_t a;
1201 x264_frame_t *frames[X264_LOOKAHEAD_MAX+3] = { NULL, };
1202 int num_frames, orig_num_frames, keyint_limit, framecnt;
1203 int i_mb_count = NUM_MBS;
1204 int cost1p0, cost2p0, cost1b1, cost2p1;
1205 int i_max_search = X264_MIN( h->lookahead->next.i_size, X264_LOOKAHEAD_MAX );
1206 int vbv_lookahead = h->param.rc.i_vbv_buffer_size && h->param.rc.i_lookahead;
1207 if( h->param.b_deterministic )
1208 i_max_search = X264_MIN( i_max_search, h->lookahead->i_slicetype_length + !keyframe );
1210 assert( h->frames.b_have_lowres );
1212 if( !h->lookahead->last_nonb )
1214 frames[0] = h->lookahead->last_nonb;
1215 for( framecnt = 0; framecnt < i_max_search && h->lookahead->next.list[framecnt]->i_type == X264_TYPE_AUTO; framecnt++ )
1216 frames[framecnt+1] = h->lookahead->next.list[framecnt];
1220 if( h->param.rc.b_mb_tree )
1221 x264_macroblock_tree( h, &a, frames, 0, keyframe );
1225 keyint_limit = h->param.i_keyint_max - frames[0]->i_frame + h->lookahead->i_last_keyframe - 1;
1226 orig_num_frames = num_frames = h->param.b_intra_refresh ? framecnt : X264_MIN( framecnt, keyint_limit );
1228 x264_lowres_context_init( h, &a );
1230 /* This is important psy-wise: if we have a non-scenecut keyframe,
1231 * there will be significant visual artifacts if the frames just before
1232 * go down in quality due to being referenced less, despite it being
1233 * more RD-optimal. */
1234 if( (h->param.analyse.b_psy && h->param.rc.b_mb_tree) || vbv_lookahead )
1235 num_frames = framecnt;
1236 else if( num_frames == 0 )
1238 frames[1]->i_type = X264_TYPE_I;
1242 int num_bframes = 0;
1243 int num_analysed_frames = num_frames;
1245 if( h->param.i_scenecut_threshold && scenecut( h, &a, frames, 0, 1, 1, orig_num_frames, i_max_search ) )
1247 frames[1]->i_type = X264_TYPE_I;
1251 if( h->param.i_bframe )
1253 if( h->param.i_bframe_adaptive == X264_B_ADAPT_TRELLIS )
1255 if( num_frames > 1 )
1257 char best_paths[X264_BFRAME_MAX+1][X264_LOOKAHEAD_MAX] = {"","P"};
1258 int best_path_index = (num_frames-1) % (X264_BFRAME_MAX+1);
1260 /* Perform the frametype analysis. */
1261 for( int j = 2; j < num_frames; j++ )
1262 x264_slicetype_path( h, &a, frames, j, best_paths );
1264 num_bframes = strspn( best_paths[best_path_index], "B" );
1265 /* Load the results of the analysis into the frame types. */
1266 for( int j = 1; j < num_frames; j++ )
1267 frames[j]->i_type = best_paths[best_path_index][j-1] == 'B' ? X264_TYPE_B : X264_TYPE_P;
1269 frames[num_frames]->i_type = X264_TYPE_P;
1271 else if( h->param.i_bframe_adaptive == X264_B_ADAPT_FAST )
1273 for( int i = 0; i <= num_frames-2; )
1275 cost2p1 = x264_slicetype_frame_cost( h, &a, frames, i+0, i+2, i+2, 1 );
1276 if( frames[i+2]->i_intra_mbs[2] > i_mb_count / 2 )
1278 frames[i+1]->i_type = X264_TYPE_P;
1279 frames[i+2]->i_type = X264_TYPE_P;
1284 cost1b1 = x264_slicetype_frame_cost( h, &a, frames, i+0, i+2, i+1, 0 );
1285 cost1p0 = x264_slicetype_frame_cost( h, &a, frames, i+0, i+1, i+1, 0 );
1286 cost2p0 = x264_slicetype_frame_cost( h, &a, frames, i+1, i+2, i+2, 0 );
1288 if( cost1p0 + cost2p0 < cost1b1 + cost2p1 )
1290 frames[i+1]->i_type = X264_TYPE_P;
1295 // arbitrary and untuned
1296 #define INTER_THRESH 300
1297 #define P_SENS_BIAS (50 - h->param.i_bframe_bias)
1298 frames[i+1]->i_type = X264_TYPE_B;
1301 for( j = i+2; j <= X264_MIN( i+h->param.i_bframe, num_frames-1 ); j++ )
1303 int pthresh = X264_MAX(INTER_THRESH - P_SENS_BIAS * (j-i-1), INTER_THRESH/10);
1304 int pcost = x264_slicetype_frame_cost( h, &a, frames, i+0, j+1, j+1, 1 );
1305 if( pcost > pthresh*i_mb_count || frames[j+1]->i_intra_mbs[j-i+1] > i_mb_count/3 )
1307 frames[j]->i_type = X264_TYPE_B;
1309 frames[j]->i_type = X264_TYPE_P;
1312 frames[num_frames]->i_type = X264_TYPE_P;
1314 while( num_bframes < num_frames && frames[num_bframes+1]->i_type == X264_TYPE_B )
1319 num_bframes = X264_MIN(num_frames-1, h->param.i_bframe);
1320 for( int j = 1; j < num_frames; j++ )
1321 frames[j]->i_type = (j%(num_bframes+1)) ? X264_TYPE_B : X264_TYPE_P;
1322 frames[num_frames]->i_type = X264_TYPE_P;
1325 /* Check scenecut on the first minigop. */
1326 for( int j = 1; j < num_bframes+1; j++ )
1327 if( h->param.i_scenecut_threshold && scenecut( h, &a, frames, j, j+1, 0, orig_num_frames, i_max_search ) )
1329 frames[j]->i_type = X264_TYPE_P;
1330 num_analysed_frames = j;
1334 reset_start = keyframe ? 1 : X264_MIN( num_bframes+2, num_analysed_frames+1 );
1338 for( int j = 1; j <= num_frames; j++ )
1339 frames[j]->i_type = X264_TYPE_P;
1340 reset_start = !keyframe + 1;
1344 /* Perform the actual macroblock tree analysis.
1345 * Don't go farther than the maximum keyframe interval; this helps in short GOPs. */
1346 if( h->param.rc.b_mb_tree )
1347 x264_macroblock_tree( h, &a, frames, X264_MIN(num_frames, h->param.i_keyint_max), keyframe );
1349 /* Enforce keyframe limit. */
1350 if( !h->param.b_intra_refresh )
1351 for( int i = keyint_limit+1; i <= num_frames; i += h->param.i_keyint_max )
1353 frames[i]->i_type = X264_TYPE_I;
1354 reset_start = X264_MIN( reset_start, i+1 );
1355 if( h->param.i_open_gop == X264_OPEN_GOP_BLURAY )
1356 while( IS_X264_TYPE_B( frames[i-1]->i_type ) )
1361 x264_vbv_lookahead( h, &a, frames, num_frames, keyframe );
1363 /* Restore frametypes for all frames that haven't actually been decided yet. */
1364 for( int j = reset_start; j <= num_frames; j++ )
1365 frames[j]->i_type = X264_TYPE_AUTO;
1368 void x264_slicetype_decide( x264_t *h )
1370 x264_frame_t *frames[X264_BFRAME_MAX+2];
1375 if( !h->lookahead->next.i_size )
1378 int lookahead_size = h->lookahead->next.i_size;
1380 for( int i = 0; i < h->lookahead->next.i_size; i++ )
1382 if( h->param.b_vfr_input )
1384 if( lookahead_size-- > 1 )
1385 h->lookahead->next.list[i]->i_duration = 2 * (h->lookahead->next.list[i+1]->i_pts - h->lookahead->next.list[i]->i_pts);
1387 h->lookahead->next.list[i]->i_duration = h->i_prev_duration;
1390 h->lookahead->next.list[i]->i_duration = delta_tfi_divisor[h->lookahead->next.list[i]->i_pic_struct];
1391 h->i_prev_duration = h->lookahead->next.list[i]->i_duration;
1392 h->lookahead->next.list[i]->f_duration = (double)h->lookahead->next.list[i]->i_duration
1393 * h->sps->vui.i_num_units_in_tick
1394 / h->sps->vui.i_time_scale;
1396 if( h->lookahead->next.list[i]->i_frame > h->i_disp_fields_last_frame && lookahead_size > 0 )
1398 h->lookahead->next.list[i]->i_field_cnt = h->i_disp_fields;
1399 h->i_disp_fields += h->lookahead->next.list[i]->i_duration;
1400 h->i_disp_fields_last_frame = h->lookahead->next.list[i]->i_frame;
1402 else if( lookahead_size == 0 )
1404 h->lookahead->next.list[i]->i_field_cnt = h->i_disp_fields;
1405 h->lookahead->next.list[i]->i_duration = h->i_prev_duration;
1409 if( h->param.rc.b_stat_read )
1411 /* Use the frame types from the first pass */
1412 for( int i = 0; i < h->lookahead->next.i_size; i++ )
1413 h->lookahead->next.list[i]->i_type =
1414 x264_ratecontrol_slice_type( h, h->lookahead->next.list[i]->i_frame );
1416 else if( (h->param.i_bframe && h->param.i_bframe_adaptive)
1417 || h->param.i_scenecut_threshold
1418 || h->param.rc.b_mb_tree
1419 || (h->param.rc.i_vbv_buffer_size && h->param.rc.i_lookahead) )
1420 x264_slicetype_analyse( h, 0 );
1422 for( bframes = 0, brefs = 0;; bframes++ )
1424 frm = h->lookahead->next.list[bframes];
1425 if( frm->i_type == X264_TYPE_BREF && h->param.i_bframe_pyramid < X264_B_PYRAMID_NORMAL &&
1426 brefs == h->param.i_bframe_pyramid )
1428 frm->i_type = X264_TYPE_B;
1429 x264_log( h, X264_LOG_WARNING, "B-ref at frame %d incompatible with B-pyramid %s \n",
1430 frm->i_frame, x264_b_pyramid_names[h->param.i_bframe_pyramid] );
1432 /* pyramid with multiple B-refs needs a big enough dpb that the preceding P-frame stays available.
1433 smaller dpb could be supported by smart enough use of mmco, but it's easier just to forbid it. */
1434 else if( frm->i_type == X264_TYPE_BREF && h->param.i_bframe_pyramid == X264_B_PYRAMID_NORMAL &&
1435 brefs && h->param.i_frame_reference <= (brefs+3) )
1437 frm->i_type = X264_TYPE_B;
1438 x264_log( h, X264_LOG_WARNING, "B-ref at frame %d incompatible with B-pyramid %s and %d reference frames\n",
1439 frm->i_frame, x264_b_pyramid_names[h->param.i_bframe_pyramid], h->param.i_frame_reference );
1442 if( frm->i_type == X264_TYPE_KEYFRAME )
1443 frm->i_type = h->param.i_open_gop ? X264_TYPE_I : X264_TYPE_IDR;
1445 /* Limit GOP size */
1446 if( (!h->param.b_intra_refresh || frm->i_frame == 0) && frm->i_frame - h->lookahead->i_last_keyframe >= h->param.i_keyint_max )
1448 if( frm->i_type == X264_TYPE_AUTO || frm->i_type == X264_TYPE_I )
1449 frm->i_type = h->param.i_open_gop && h->lookahead->i_last_keyframe >= 0 ? X264_TYPE_I : X264_TYPE_IDR;
1450 int warn = frm->i_type != X264_TYPE_IDR;
1451 if( warn && h->param.i_open_gop )
1452 warn &= frm->i_type != X264_TYPE_I;
1454 x264_log( h, X264_LOG_WARNING, "specified frame type (%d) at %d is not compatible with keyframe interval\n", frm->i_type, frm->i_frame );
1456 if( frm->i_type == X264_TYPE_I && frm->i_frame - h->lookahead->i_last_keyframe >= h->param.i_keyint_min )
1458 if( h->param.i_open_gop )
1460 h->lookahead->i_last_keyframe = frm->i_frame; // Use display order
1461 if( h->param.i_open_gop == X264_OPEN_GOP_BLURAY )
1462 h->lookahead->i_last_keyframe -= bframes; // Use bluray order
1463 frm->b_keyframe = 1;
1466 frm->i_type = X264_TYPE_IDR;
1468 if( frm->i_type == X264_TYPE_IDR )
1471 h->lookahead->i_last_keyframe = frm->i_frame;
1472 frm->b_keyframe = 1;
1476 h->lookahead->next.list[bframes]->i_type = X264_TYPE_P;
1480 if( bframes == h->param.i_bframe ||
1481 !h->lookahead->next.list[bframes+1] )
1483 if( IS_X264_TYPE_B( frm->i_type ) )
1484 x264_log( h, X264_LOG_WARNING, "specified frame type is not compatible with max B-frames\n" );
1485 if( frm->i_type == X264_TYPE_AUTO
1486 || IS_X264_TYPE_B( frm->i_type ) )
1487 frm->i_type = X264_TYPE_P;
1490 if( frm->i_type == X264_TYPE_BREF )
1493 if( frm->i_type == X264_TYPE_AUTO )
1494 frm->i_type = X264_TYPE_B;
1496 else if( !IS_X264_TYPE_B( frm->i_type ) ) break;
1500 h->lookahead->next.list[bframes-1]->b_last_minigop_bframe = 1;
1501 h->lookahead->next.list[bframes]->i_bframes = bframes;
1503 /* insert a bref into the sequence */
1504 if( h->param.i_bframe_pyramid && bframes > 1 && !brefs )
1506 h->lookahead->next.list[bframes/2]->i_type = X264_TYPE_BREF;
1510 /* calculate the frame costs ahead of time for x264_rc_analyse_slice while we still have lowres */
1511 if( h->param.rc.i_rc_method != X264_RC_CQP )
1513 x264_mb_analysis_t a;
1515 p1 = b = bframes + 1;
1517 x264_lowres_context_init( h, &a );
1519 frames[0] = h->lookahead->last_nonb;
1520 memcpy( &frames[1], h->lookahead->next.list, (bframes+1) * sizeof(x264_frame_t*) );
1521 if( IS_X264_TYPE_I( h->lookahead->next.list[bframes]->i_type ) )
1526 x264_slicetype_frame_cost( h, &a, frames, p0, p1, b, 0 );
1528 if( (p0 != p1 || bframes) && h->param.rc.i_vbv_buffer_size )
1530 /* We need the intra costs for row SATDs. */
1531 x264_slicetype_frame_cost( h, &a, frames, b, b, b, 0 );
1533 /* We need B-frame costs for row SATDs. */
1535 for( b = 1; b <= bframes; b++ )
1537 if( frames[b]->i_type == X264_TYPE_B )
1538 for( p1 = b; frames[p1]->i_type == X264_TYPE_B; )
1542 x264_slicetype_frame_cost( h, &a, frames, p0, p1, b, 0 );
1543 if( frames[b]->i_type == X264_TYPE_BREF )
1549 /* Analyse for weighted P frames */
1550 if( !h->param.rc.b_stat_read && h->lookahead->next.list[bframes]->i_type == X264_TYPE_P
1551 && h->param.analyse.i_weighted_pred >= X264_WEIGHTP_SIMPLE )
1554 x264_weights_analyse( h, h->lookahead->next.list[bframes], h->lookahead->last_nonb, 0 );
1557 /* shift sequence to coded order.
1558 use a small temporary list to avoid shifting the entire next buffer around */
1559 int i_coded = h->lookahead->next.list[0]->i_frame;
1562 int idx_list[] = { brefs+1, 1 };
1563 for( int i = 0; i < bframes; i++ )
1565 int idx = idx_list[h->lookahead->next.list[i]->i_type == X264_TYPE_BREF]++;
1566 frames[idx] = h->lookahead->next.list[i];
1567 frames[idx]->i_reordered_pts = h->lookahead->next.list[idx]->i_pts;
1569 frames[0] = h->lookahead->next.list[bframes];
1570 frames[0]->i_reordered_pts = h->lookahead->next.list[0]->i_pts;
1571 memcpy( h->lookahead->next.list, frames, (bframes+1) * sizeof(x264_frame_t*) );
1574 for( int i = 0; i <= bframes; i++ )
1576 h->lookahead->next.list[i]->i_coded = i_coded++;
1579 x264_calculate_durations( h, h->lookahead->next.list[i], h->lookahead->next.list[i-1], &h->i_cpb_delay, &h->i_coded_fields );
1580 h->lookahead->next.list[0]->f_planned_cpb_duration[i-1] = (double)h->lookahead->next.list[i-1]->i_cpb_duration *
1581 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1584 x264_calculate_durations( h, h->lookahead->next.list[i], NULL, &h->i_cpb_delay, &h->i_coded_fields );
1586 h->lookahead->next.list[0]->f_planned_cpb_duration[i] = (double)h->lookahead->next.list[i]->i_cpb_duration *
1587 h->sps->vui.i_num_units_in_tick / h->sps->vui.i_time_scale;
1591 int x264_rc_analyse_slice( x264_t *h )
1597 if( IS_X264_TYPE_I(h->fenc->i_type) )
1599 else if( h->fenc->i_type == X264_TYPE_P )
1600 p1 = b = h->fenc->i_bframes + 1;
1603 p1 = (h->fref1[0]->i_poc - h->fref0[0]->i_poc)/2;
1604 b = (h->fenc->i_poc - h->fref0[0]->i_poc)/2;
1606 /* We don't need to assign p0/p1 since we are not performing any real analysis here. */
1607 x264_frame_t **frames = &h->fenc - b;
1609 /* cost should have been already calculated by x264_slicetype_decide */
1610 cost = frames[b]->i_cost_est[b-p0][p1-b];
1611 assert( cost >= 0 );
1613 if( h->param.rc.b_mb_tree && !h->param.rc.b_stat_read )
1615 cost = x264_slicetype_frame_cost_recalculate( h, frames, p0, p1, b );
1616 if( b && h->param.rc.i_vbv_buffer_size )
1617 x264_slicetype_frame_cost_recalculate( h, frames, b, b, b );
1619 /* In AQ, use the weighted score instead. */
1620 else if( h->param.rc.i_aq_mode )
1621 cost = frames[b]->i_cost_est_aq[b-p0][p1-b];
1623 h->fenc->i_row_satd = h->fenc->i_row_satds[b-p0][p1-b];
1624 h->fdec->i_row_satd = h->fdec->i_row_satds[b-p0][p1-b];
1625 h->fdec->i_satd = cost;
1626 memcpy( h->fdec->i_row_satd, h->fenc->i_row_satd, h->mb.i_mb_height * sizeof(int) );
1627 if( !IS_X264_TYPE_I(h->fenc->i_type) )
1628 memcpy( h->fdec->i_row_satds[0][0], h->fenc->i_row_satds[0][0], h->mb.i_mb_height * sizeof(int) );
1630 if( h->param.b_intra_refresh && h->param.rc.i_vbv_buffer_size && h->fenc->i_type == X264_TYPE_P )
1632 int ip_factor = 256 * h->param.rc.f_ip_factor; /* fix8 */
1633 for( int y = 0; y < h->mb.i_mb_height; y++ )
1635 int mb_xy = y * h->mb.i_mb_stride;
1636 for( int x = h->fdec->i_pir_start_col; x <= h->fdec->i_pir_end_col; x++, mb_xy++ )
1638 int intra_cost = (h->fenc->i_intra_cost[mb_xy] * ip_factor + 128) >> 8;
1639 int inter_cost = h->fenc->lowres_costs[b-p0][p1-b][mb_xy] & LOWRES_COST_MASK;
1640 int diff = intra_cost - inter_cost;
1641 if( h->param.rc.i_aq_mode )
1642 h->fdec->i_row_satd[y] += (diff * frames[b]->i_inv_qscale_factor[mb_xy] + 128) >> 8;
1644 h->fdec->i_row_satd[y] += diff;