4 * Copyright (C) 2012 - 2013 Guillaume Martres
5 * Copyright (C) 2013 Anand Meher Kotra
7 * This file is part of Libav.
9 * Libav is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU Lesser General Public
11 * License as published by the Free Software Foundation; either
12 * version 2.1 of the License, or (at your option) any later version.
14 * Libav is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * Lesser General Public License for more details.
19 * You should have received a copy of the GNU Lesser General Public
20 * License along with Libav; if not, write to the Free Software
21 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
26 static const uint8_t l0_l1_cand_idx[12][2] = {
41 void ff_hevc_set_neighbour_available(HEVCContext *s, int x0, int y0,
44 HEVCLocalContext *lc = &s->HEVClc;
45 int x0b = x0 & ((1 << s->sps->log2_ctb_size) - 1);
46 int y0b = y0 & ((1 << s->sps->log2_ctb_size) - 1);
48 lc->na.cand_up = (lc->ctb_up_flag || y0b);
49 lc->na.cand_left = (lc->ctb_left_flag || x0b);
50 lc->na.cand_up_left = (!x0b && !y0b) ? lc->ctb_up_left_flag : lc->na.cand_left && lc->na.cand_up;
51 lc->na.cand_up_right_sap =
52 ((x0b + nPbW) == (1 << s->sps->log2_ctb_size)) ?
53 lc->ctb_up_right_flag && !y0b : lc->na.cand_up;
54 lc->na.cand_up_right =
55 ((x0b + nPbW) == (1 << s->sps->log2_ctb_size) ?
56 lc->ctb_up_right_flag && !y0b : lc->na.cand_up )
57 && (x0 + nPbW) < lc->end_of_tiles_x;
58 lc->na.cand_bottom_left = ((y0 + nPbH) >= lc->end_of_tiles_y) ? 0 : lc->na.cand_left;
62 * 6.4.1 Derivation process for z-scan order block availability
64 static int z_scan_block_avail(HEVCContext *s, int xCurr, int yCurr,
67 #define MIN_TB_ADDR_ZS(x, y) \
68 s->pps->min_tb_addr_zs[(y) * s->sps->min_tb_width + (x)]
69 int Curr = MIN_TB_ADDR_ZS(xCurr >> s->sps->log2_min_tb_size,
70 yCurr >> s->sps->log2_min_tb_size);
73 if (xN < 0 || yN < 0 ||
74 xN >= s->sps->width ||
78 N = MIN_TB_ADDR_ZS(xN >> s->sps->log2_min_tb_size,
79 yN >> s->sps->log2_min_tb_size);
84 static int same_prediction_block(HEVCLocalContext *lc, int log2_cb_size,
85 int x0, int y0, int nPbW, int nPbH,
86 int xA1, int yA1, int partIdx)
88 return !(nPbW << 1 == 1 << log2_cb_size &&
89 nPbH << 1 == 1 << log2_cb_size && partIdx == 1 &&
90 lc->cu.x + nPbW > xA1 &&
91 lc->cu.y + nPbH <= yA1);
95 * 6.4.2 Derivation process for prediction block availability
97 static int check_prediction_block_available(HEVCContext *s, int log2_cb_size,
98 int x0, int y0, int nPbW, int nPbH,
99 int xA1, int yA1, int partIdx)
101 HEVCLocalContext *lc = &s->HEVClc;
103 if (lc->cu.x < xA1 && lc->cu.y < yA1 &&
104 (lc->cu.x + (1 << log2_cb_size)) > xA1 &&
105 (lc->cu.y + (1 << log2_cb_size)) > yA1)
106 return same_prediction_block(lc, log2_cb_size, x0, y0,
107 nPbW, nPbH, xA1, yA1, partIdx);
109 return z_scan_block_avail(s, x0, y0, xA1, yA1);
112 //check if the two luma locations belong to the same mostion estimation region
113 static int isDiffMER(HEVCContext *s, int xN, int yN, int xP, int yP)
115 uint8_t plevel = s->pps->log2_parallel_merge_level;
117 return xN >> plevel == xP >> plevel &&
118 yN >> plevel == yP >> plevel;
121 #define MATCH_MV(x) (AV_RN32A(&A.x) == AV_RN32A(&B.x))
122 #define MATCH(x) (A.x == B.x)
124 // check if the mv's and refidx are the same between A and B
125 static int compareMVrefidx(struct MvField A, struct MvField B)
127 if (A.pred_flag[0] && A.pred_flag[1] && B.pred_flag[0] && B.pred_flag[1])
128 return MATCH(ref_idx[0]) && MATCH_MV(mv[0]) &&
129 MATCH(ref_idx[1]) && MATCH_MV(mv[1]);
131 if (A.pred_flag[0] && !A.pred_flag[1] && B.pred_flag[0] && !B.pred_flag[1])
132 return MATCH(ref_idx[0]) && MATCH_MV(mv[0]);
134 if (!A.pred_flag[0] && A.pred_flag[1] && !B.pred_flag[0] && B.pred_flag[1])
135 return MATCH(ref_idx[1]) && MATCH_MV(mv[1]);
140 static av_always_inline void mv_scale(Mv *dst, Mv *src, int td, int tb)
142 int tx, scale_factor;
144 td = av_clip_int8_c(td);
145 tb = av_clip_int8_c(tb);
146 tx = (0x4000 + abs(td / 2)) / td;
147 scale_factor = av_clip_c((tb * tx + 32) >> 6, -4096, 4095);
148 dst->x = av_clip_int16_c((scale_factor * src->x + 127 +
149 (scale_factor * src->x < 0)) >> 8);
150 dst->y = av_clip_int16_c((scale_factor * src->y + 127 +
151 (scale_factor * src->y < 0)) >> 8);
154 static int check_mvset(Mv *mvLXCol, Mv *mvCol,
156 RefPicList *refPicList, int X, int refIdxLx,
157 RefPicList *refPicList_col, int listCol, int refidxCol)
159 int cur_lt = refPicList[X].isLongTerm[refIdxLx];
160 int col_lt = refPicList_col[listCol].isLongTerm[refidxCol];
161 int col_poc_diff, cur_poc_diff;
163 if (cur_lt != col_lt) {
169 col_poc_diff = colPic - refPicList_col[listCol].list[refidxCol];
170 cur_poc_diff = poc - refPicList[X].list[refIdxLx];
173 col_poc_diff = 1; // error resilience
175 if (cur_lt || col_poc_diff == cur_poc_diff) {
176 mvLXCol->x = mvCol->x;
177 mvLXCol->y = mvCol->y;
179 mv_scale(mvLXCol, mvCol, col_poc_diff, cur_poc_diff);
184 #define CHECK_MVSET(l) \
185 check_mvset(mvLXCol, temp_col.mv + l, \
187 refPicList, X, refIdxLx, \
188 refPicList_col, L ## l, temp_col.ref_idx[l])
190 // derive the motion vectors section 8.5.3.1.8
191 static int derive_temporal_colocated_mvs(HEVCContext *s, MvField temp_col,
192 int refIdxLx, Mv *mvLXCol, int X,
193 int colPic, RefPicList *refPicList_col)
195 RefPicList *refPicList = s->ref->refPicList;
197 if (temp_col.is_intra) {
203 if (temp_col.pred_flag[0] == 0)
204 return CHECK_MVSET(1);
205 else if (temp_col.pred_flag[0] == 1 && temp_col.pred_flag[1] == 0)
206 return CHECK_MVSET(0);
207 else if (temp_col.pred_flag[0] == 1 && temp_col.pred_flag[1] == 1) {
208 int check_diffpicount = 0;
210 for (i = 0; i < refPicList[0].nb_refs; i++) {
211 if (refPicList[0].list[i] > s->poc)
214 for (i = 0; i < refPicList[1].nb_refs; i++) {
215 if (refPicList[1].list[i] > s->poc)
218 if (check_diffpicount == 0 && X == 0)
219 return CHECK_MVSET(0);
220 else if (check_diffpicount == 0 && X == 1)
221 return CHECK_MVSET(1);
223 if (s->sh.collocated_list == L1)
224 return CHECK_MVSET(0);
226 return CHECK_MVSET(1);
233 #define TAB_MVF(x, y) \
234 tab_mvf[(y) * min_pu_width + x]
236 #define TAB_MVF_PU(v) \
237 TAB_MVF(x ## v ## _pu, y ## v ## _pu)
239 #define DERIVE_TEMPORAL_COLOCATED_MVS \
240 derive_temporal_colocated_mvs(s, temp_col, \
241 refIdxLx, mvLXCol, X, colPic, \
242 ff_hevc_get_ref_list(s, ref, x, y))
245 * 8.5.3.1.7 temporal luma motion vector prediction
247 static int temporal_luma_motion_vector(HEVCContext *s, int x0, int y0,
248 int nPbW, int nPbH, int refIdxLx,
253 int x, y, x_pu, y_pu;
254 int min_pu_width = s->sps->min_pu_width;
255 int availableFlagLXCol = 0;
258 HEVCFrame *ref = s->ref->collocated_ref;
263 tab_mvf = ref->tab_mvf;
266 //bottom right collocated motion vector
271 (y0 >> s->sps->log2_ctb_size) == (y >> s->sps->log2_ctb_size) &&
272 y < s->sps->height &&
276 ff_thread_await_progress(&ref->tf, y, 0);
277 x_pu = x >> s->sps->log2_min_pu_size;
278 y_pu = y >> s->sps->log2_min_pu_size;
279 temp_col = TAB_MVF(x_pu, y_pu);
280 availableFlagLXCol = DERIVE_TEMPORAL_COLOCATED_MVS;
283 // derive center collocated motion vector
284 if (tab_mvf && !availableFlagLXCol) {
285 x = x0 + (nPbW >> 1);
286 y = y0 + (nPbH >> 1);
289 ff_thread_await_progress(&ref->tf, y, 0);
290 x_pu = x >> s->sps->log2_min_pu_size;
291 y_pu = y >> s->sps->log2_min_pu_size;
292 temp_col = TAB_MVF(x_pu, y_pu);
293 availableFlagLXCol = DERIVE_TEMPORAL_COLOCATED_MVS;
295 return availableFlagLXCol;
298 #define AVAILABLE(cand, v) \
299 (cand && !TAB_MVF_PU(v).is_intra)
301 #define PRED_BLOCK_AVAILABLE(v) \
302 check_prediction_block_available(s, log2_cb_size, \
303 x0, y0, nPbW, nPbH, \
304 x ## v, y ## v, part_idx)
306 #define COMPARE_MV_REFIDX(a, b) \
307 compareMVrefidx(TAB_MVF_PU(a), TAB_MVF_PU(b))
310 * 8.5.3.1.2 Derivation process for spatial merging candidates
312 static void derive_spatial_merge_candidates(HEVCContext *s, int x0, int y0,
315 int singleMCLFlag, int part_idx,
317 struct MvField mergecandlist[])
319 HEVCLocalContext *lc = &s->HEVClc;
320 RefPicList *refPicList = s->ref->refPicList;
321 MvField *tab_mvf = s->ref->tab_mvf;
323 const int min_pu_width = s->sps->min_pu_width;
325 const int cand_bottom_left = lc->na.cand_bottom_left;
326 const int cand_left = lc->na.cand_left;
327 const int cand_up_left = lc->na.cand_up_left;
328 const int cand_up = lc->na.cand_up;
329 const int cand_up_right = lc->na.cand_up_right_sap;
331 const int xA1 = x0 - 1;
332 const int yA1 = y0 + nPbH - 1;
333 const int xA1_pu = xA1 >> s->sps->log2_min_pu_size;
334 const int yA1_pu = yA1 >> s->sps->log2_min_pu_size;
336 const int xB1 = x0 + nPbW - 1;
337 const int yB1 = y0 - 1;
338 const int xB1_pu = xB1 >> s->sps->log2_min_pu_size;
339 const int yB1_pu = yB1 >> s->sps->log2_min_pu_size;
341 const int xB0 = x0 + nPbW;
342 const int yB0 = y0 - 1;
343 const int xB0_pu = xB0 >> s->sps->log2_min_pu_size;
344 const int yB0_pu = yB0 >> s->sps->log2_min_pu_size;
346 const int xA0 = x0 - 1;
347 const int yA0 = y0 + nPbH;
348 const int xA0_pu = xA0 >> s->sps->log2_min_pu_size;
349 const int yA0_pu = yA0 >> s->sps->log2_min_pu_size;
351 const int xB2 = x0 - 1;
352 const int yB2 = y0 - 1;
353 const int xB2_pu = xB2 >> s->sps->log2_min_pu_size;
354 const int yB2_pu = yB2 >> s->sps->log2_min_pu_size;
356 const int nb_refs = (s->sh.slice_type == P_SLICE) ?
357 s->sh.nb_refs[0] : FFMIN(s->sh.nb_refs[0], s->sh.nb_refs[1]);
363 int nb_merge_cand = 0;
364 int nb_orig_merge_cand = 0;
374 //first left spatial merge candidate
375 is_available_a1 = AVAILABLE(cand_left, A1);
377 if (!singleMCLFlag && part_idx == 1 &&
378 (lc->cu.part_mode == PART_Nx2N ||
379 lc->cu.part_mode == PART_nLx2N ||
380 lc->cu.part_mode == PART_nRx2N) ||
381 isDiffMER(s, xA1, yA1, x0, y0)) {
385 if (is_available_a1) {
386 mergecandlist[0] = TAB_MVF_PU(A1);
392 // above spatial merge candidate
393 is_available_b1 = AVAILABLE(cand_up, B1);
395 if (!singleMCLFlag && part_idx == 1 &&
396 (lc->cu.part_mode == PART_2NxN ||
397 lc->cu.part_mode == PART_2NxnU ||
398 lc->cu.part_mode == PART_2NxnD) ||
399 isDiffMER(s, xB1, yB1, x0, y0)) {
403 if (is_available_a1 && is_available_b1)
404 check_MER = !COMPARE_MV_REFIDX(B1, A1);
406 if (is_available_b1 && check_MER)
407 mergecandlist[nb_merge_cand++] = TAB_MVF_PU(B1);
409 // above right spatial merge candidate
411 check_B0 = PRED_BLOCK_AVAILABLE(B0);
413 is_available_b0 = check_B0 && AVAILABLE(cand_up_right, B0);
415 if (isDiffMER(s, xB0, yB0, x0, y0))
418 if (is_available_b1 && is_available_b0)
419 check_MER = !COMPARE_MV_REFIDX(B0, B1);
421 if (is_available_b0 && check_MER) {
422 mergecandlist[nb_merge_cand] = TAB_MVF_PU(B0);
423 if (merge_idx == nb_merge_cand)
428 // left bottom spatial merge candidate
430 check_A0 = PRED_BLOCK_AVAILABLE(A0);
432 is_available_a0 = check_A0 && AVAILABLE(cand_bottom_left, A0);
434 if (isDiffMER(s, xA0, yA0, x0, y0))
437 if (is_available_a1 && is_available_a0)
438 check_MER = !COMPARE_MV_REFIDX(A0, A1);
440 if (is_available_a0 && check_MER) {
441 mergecandlist[nb_merge_cand] = TAB_MVF_PU(A0);
442 if (merge_idx == nb_merge_cand)
447 // above left spatial merge candidate
450 is_available_b2 = AVAILABLE(cand_up_left, B2);
452 if (isDiffMER(s, xB2, yB2, x0, y0))
455 if (is_available_a1 && is_available_b2)
456 check_MER = !COMPARE_MV_REFIDX(B2, A1);
458 if (is_available_b1 && is_available_b2)
459 check_MER_1 = !COMPARE_MV_REFIDX(B2, B1);
461 if (is_available_b2 && check_MER && check_MER_1 && nb_merge_cand != 4) {
462 mergecandlist[nb_merge_cand] = TAB_MVF_PU(B2);
463 if (merge_idx == nb_merge_cand)
468 // temporal motion vector candidate
469 if (s->sh.slice_temporal_mvp_enabled_flag &&
470 nb_merge_cand < s->sh.max_num_merge_cand) {
471 Mv mv_l0_col, mv_l1_col;
472 int available_l0 = temporal_luma_motion_vector(s, x0, y0, nPbW, nPbH,
474 int available_l1 = (s->sh.slice_type == B_SLICE) ?
475 temporal_luma_motion_vector(s, x0, y0, nPbW, nPbH,
476 0, &mv_l1_col, 1) : 0;
478 if (available_l0 || available_l1) {
479 mergecandlist[nb_merge_cand].is_intra = 0;
480 mergecandlist[nb_merge_cand].pred_flag[0] = available_l0;
481 mergecandlist[nb_merge_cand].pred_flag[1] = available_l1;
483 mergecandlist[nb_merge_cand].mv[0] = mv_l0_col;
484 mergecandlist[nb_merge_cand].ref_idx[0] = 0;
487 mergecandlist[nb_merge_cand].mv[1] = mv_l1_col;
488 mergecandlist[nb_merge_cand].ref_idx[1] = 0;
490 if (merge_idx == nb_merge_cand)
496 nb_orig_merge_cand = nb_merge_cand;
498 // combined bi-predictive merge candidates (applies for B slices)
499 if (s->sh.slice_type == B_SLICE && nb_orig_merge_cand > 1 &&
500 nb_orig_merge_cand < s->sh.max_num_merge_cand) {
503 for (comb_idx = 0; nb_merge_cand < s->sh.max_num_merge_cand &&
504 comb_idx < nb_orig_merge_cand * (nb_orig_merge_cand - 1); comb_idx++) {
505 int l0_cand_idx = l0_l1_cand_idx[comb_idx][0];
506 int l1_cand_idx = l0_l1_cand_idx[comb_idx][1];
507 MvField l0_cand = mergecandlist[l0_cand_idx];
508 MvField l1_cand = mergecandlist[l1_cand_idx];
510 if (l0_cand.pred_flag[0] && l1_cand.pred_flag[1] &&
511 (refPicList[0].list[l0_cand.ref_idx[0]] !=
512 refPicList[1].list[l1_cand.ref_idx[1]] ||
513 AV_RN32A(&l0_cand.mv[0]) != AV_RN32A(&l1_cand.mv[1]))) {
514 mergecandlist[nb_merge_cand].ref_idx[0] = l0_cand.ref_idx[0];
515 mergecandlist[nb_merge_cand].ref_idx[1] = l1_cand.ref_idx[1];
516 mergecandlist[nb_merge_cand].pred_flag[0] = 1;
517 mergecandlist[nb_merge_cand].pred_flag[1] = 1;
518 AV_COPY32(&mergecandlist[nb_merge_cand].mv[0], &l0_cand.mv[0]);
519 AV_COPY32(&mergecandlist[nb_merge_cand].mv[1], &l1_cand.mv[1]);
520 mergecandlist[nb_merge_cand].is_intra = 0;
521 if (merge_idx == nb_merge_cand)
528 // append Zero motion vector candidates
529 while (nb_merge_cand < s->sh.max_num_merge_cand) {
530 mergecandlist[nb_merge_cand].pred_flag[0] = 1;
531 mergecandlist[nb_merge_cand].pred_flag[1] = s->sh.slice_type == B_SLICE;
532 AV_ZERO32(mergecandlist[nb_merge_cand].mv + 0);
533 AV_ZERO32(mergecandlist[nb_merge_cand].mv + 1);
534 mergecandlist[nb_merge_cand].is_intra = 0;
535 mergecandlist[nb_merge_cand].ref_idx[0] = zero_idx < nb_refs ? zero_idx : 0;
536 mergecandlist[nb_merge_cand].ref_idx[1] = zero_idx < nb_refs ? zero_idx : 0;
538 if (merge_idx == nb_merge_cand)
546 * 8.5.3.1.1 Derivation process of luma Mvs for merge mode
548 void ff_hevc_luma_mv_merge_mode(HEVCContext *s, int x0, int y0, int nPbW,
549 int nPbH, int log2_cb_size, int part_idx,
550 int merge_idx, MvField *mv)
552 int singleMCLFlag = 0;
553 int nCS = 1 << log2_cb_size;
554 LOCAL_ALIGNED(4, MvField, mergecand_list, [MRG_MAX_NUM_CANDS]);
557 HEVCLocalContext *lc = &s->HEVClc;
559 if (s->pps->log2_parallel_merge_level > 2 && nCS == 8) {
568 ff_hevc_set_neighbour_available(s, x0, y0, nPbW, nPbH);
569 derive_spatial_merge_candidates(s, x0, y0, nPbW, nPbH, log2_cb_size,
570 singleMCLFlag, part_idx,
571 merge_idx, mergecand_list);
573 if (mergecand_list[merge_idx].pred_flag[0] == 1 &&
574 mergecand_list[merge_idx].pred_flag[1] == 1 &&
575 (nPbW2 + nPbH2) == 12) {
576 mergecand_list[merge_idx].ref_idx[1] = -1;
577 mergecand_list[merge_idx].pred_flag[1] = 0;
580 *mv = mergecand_list[merge_idx];
583 static av_always_inline void dist_scale(HEVCContext *s, Mv *mv,
584 int min_pu_width, int x, int y,
585 int elist, int ref_idx_curr, int ref_idx)
587 RefPicList *refPicList = s->ref->refPicList;
588 MvField *tab_mvf = s->ref->tab_mvf;
589 int ref_pic_elist = refPicList[elist].list[TAB_MVF(x, y).ref_idx[elist]];
590 int ref_pic_curr = refPicList[ref_idx_curr].list[ref_idx];
592 if (ref_pic_elist != ref_pic_curr) {
593 int poc_diff = s->poc - ref_pic_elist;
596 mv_scale(mv, mv, poc_diff, s->poc - ref_pic_curr);
600 static int mv_mp_mode_mx(HEVCContext *s, int x, int y, int pred_flag_index,
601 Mv *mv, int ref_idx_curr, int ref_idx)
603 MvField *tab_mvf = s->ref->tab_mvf;
604 int min_pu_width = s->sps->min_pu_width;
606 RefPicList *refPicList = s->ref->refPicList;
608 if (TAB_MVF(x, y).pred_flag[pred_flag_index] == 1 &&
609 refPicList[pred_flag_index].list[TAB_MVF(x, y).ref_idx[pred_flag_index]] == refPicList[ref_idx_curr].list[ref_idx]) {
610 *mv = TAB_MVF(x, y).mv[pred_flag_index];
616 static int mv_mp_mode_mx_lt(HEVCContext *s, int x, int y, int pred_flag_index,
617 Mv *mv, int ref_idx_curr, int ref_idx)
619 MvField *tab_mvf = s->ref->tab_mvf;
620 int min_pu_width = s->sps->min_pu_width;
622 RefPicList *refPicList = s->ref->refPicList;
623 int currIsLongTerm = refPicList[ref_idx_curr].isLongTerm[ref_idx];
626 refPicList[pred_flag_index].isLongTerm[(TAB_MVF(x, y).ref_idx[pred_flag_index])];
628 if (TAB_MVF(x, y).pred_flag[pred_flag_index] &&
629 colIsLongTerm == currIsLongTerm) {
630 *mv = TAB_MVF(x, y).mv[pred_flag_index];
632 dist_scale(s, mv, min_pu_width, x, y,
633 pred_flag_index, ref_idx_curr, ref_idx);
639 #define MP_MX(v, pred, mx) \
640 mv_mp_mode_mx(s, x ## v ## _pu, y ## v ## _pu, pred, \
641 &mx, ref_idx_curr, ref_idx)
643 #define MP_MX_LT(v, pred, mx) \
644 mv_mp_mode_mx_lt(s, x ## v ## _pu, y ## v ## _pu, pred, \
645 &mx, ref_idx_curr, ref_idx)
647 void ff_hevc_luma_mv_mvp_mode(HEVCContext *s, int x0, int y0, int nPbW,
648 int nPbH, int log2_cb_size, int part_idx,
649 int merge_idx, MvField *mv,
650 int mvp_lx_flag, int LX)
652 HEVCLocalContext *lc = &s->HEVClc;
653 MvField *tab_mvf = s->ref->tab_mvf;
654 int isScaledFlag_L0 = 0;
655 int availableFlagLXA0 = 0;
656 int availableFlagLXB0 = 0;
657 int numMVPCandLX = 0;
658 int min_pu_width = s->sps->min_pu_width;
673 int xB1_pu = 0, yB1_pu = 0;
674 int is_available_b1 = 0;
677 int xB2_pu = 0, yB2_pu = 0;
678 int is_available_b2 = 0;
679 Mv mvpcand_list[2] = { { 0 } };
682 int ref_idx_curr = 0;
684 int pred_flag_index_l0;
685 int pred_flag_index_l1;
686 int x0b = x0 & ((1 << s->sps->log2_ctb_size) - 1);
687 int y0b = y0 & ((1 << s->sps->log2_ctb_size) - 1);
689 int cand_up = (lc->ctb_up_flag || y0b);
690 int cand_left = (lc->ctb_left_flag || x0b);
692 (!x0b && !y0b) ? lc->ctb_up_left_flag : cand_left && cand_up;
694 (x0b + nPbW == (1 << s->sps->log2_ctb_size) ||
695 x0 + nPbW >= lc->end_of_tiles_x) ? lc->ctb_up_right_flag && !y0b
697 int cand_bottom_left = (y0 + nPbH >= lc->end_of_tiles_y) ? 0 : cand_left;
700 ref_idx = mv->ref_idx[LX];
701 pred_flag_index_l0 = LX;
702 pred_flag_index_l1 = !LX;
704 // left bottom spatial candidate
707 xA0_pu = xA0 >> s->sps->log2_min_pu_size;
708 yA0_pu = yA0 >> s->sps->log2_min_pu_size;
710 is_available_a0 = PRED_BLOCK_AVAILABLE(A0) && AVAILABLE(cand_bottom_left, A0);
712 //left spatial merge candidate
715 xA1_pu = xA1 >> s->sps->log2_min_pu_size;
716 yA1_pu = yA1 >> s->sps->log2_min_pu_size;
718 is_available_a1 = AVAILABLE(cand_left, A1);
719 if (is_available_a0 || is_available_a1)
722 if (is_available_a0) {
723 availableFlagLXA0 = MP_MX(A0, pred_flag_index_l0, mxA);
724 if (!availableFlagLXA0)
725 availableFlagLXA0 = MP_MX(A0, pred_flag_index_l1, mxA);
728 if (is_available_a1 && !availableFlagLXA0) {
729 availableFlagLXA0 = MP_MX(A1, pred_flag_index_l0, mxA);
730 if (!availableFlagLXA0)
731 availableFlagLXA0 = MP_MX(A1, pred_flag_index_l1, mxA);
734 if (is_available_a0 && !availableFlagLXA0) {
735 availableFlagLXA0 = MP_MX_LT(A0, pred_flag_index_l0, mxA);
736 if (!availableFlagLXA0)
737 availableFlagLXA0 = MP_MX_LT(A0, pred_flag_index_l1, mxA);
740 if (is_available_a1 && !availableFlagLXA0) {
741 availableFlagLXA0 = MP_MX_LT(A1, pred_flag_index_l0, mxA);
742 if (!availableFlagLXA0)
743 availableFlagLXA0 = MP_MX_LT(A1, pred_flag_index_l1, mxA);
746 if (availableFlagLXA0 && !mvp_lx_flag) {
752 // above right spatial merge candidate
755 xB0_pu = xB0 >> s->sps->log2_min_pu_size;
756 yB0_pu = yB0 >> s->sps->log2_min_pu_size;
758 is_available_b0 = PRED_BLOCK_AVAILABLE(B0) && AVAILABLE(cand_up_right, B0);
760 if (is_available_b0) {
761 availableFlagLXB0 = MP_MX(B0, pred_flag_index_l0, mxB);
762 if (!availableFlagLXB0)
763 availableFlagLXB0 = MP_MX(B0, pred_flag_index_l1, mxB);
766 if (!availableFlagLXB0) {
767 // above spatial merge candidate
770 xB1_pu = xB1 >> s->sps->log2_min_pu_size;
771 yB1_pu = yB1 >> s->sps->log2_min_pu_size;
773 is_available_b1 = AVAILABLE(cand_up, B1);
775 if (is_available_b1) {
776 availableFlagLXB0 = MP_MX(B1, pred_flag_index_l0, mxB);
777 if (!availableFlagLXB0)
778 availableFlagLXB0 = MP_MX(B1, pred_flag_index_l1, mxB);
782 if (!availableFlagLXB0) {
783 // above left spatial merge candidate
786 xB2_pu = xB2 >> s->sps->log2_min_pu_size;
787 yB2_pu = yB2 >> s->sps->log2_min_pu_size;
788 is_available_b2 = AVAILABLE(cand_up_left, B2);
790 if (is_available_b2) {
791 availableFlagLXB0 = MP_MX(B2, pred_flag_index_l0, mxB);
792 if (!availableFlagLXB0)
793 availableFlagLXB0 = MP_MX(B2, pred_flag_index_l1, mxB);
797 if (isScaledFlag_L0 == 0) {
798 if (availableFlagLXB0) {
799 availableFlagLXA0 = 1;
802 availableFlagLXB0 = 0;
805 if (is_available_b0) {
806 availableFlagLXB0 = MP_MX_LT(B0, pred_flag_index_l0, mxB);
807 if (!availableFlagLXB0)
808 availableFlagLXB0 = MP_MX_LT(B0, pred_flag_index_l1, mxB);
811 if (is_available_b1 && !availableFlagLXB0) {
812 availableFlagLXB0 = MP_MX_LT(B1, pred_flag_index_l0, mxB);
813 if (!availableFlagLXB0)
814 availableFlagLXB0 = MP_MX_LT(B1, pred_flag_index_l1, mxB);
817 if (is_available_b2 && !availableFlagLXB0) {
818 availableFlagLXB0 = MP_MX_LT(B2, pred_flag_index_l0, mxB);
819 if (!availableFlagLXB0)
820 availableFlagLXB0 = MP_MX_LT(B2, pred_flag_index_l1, mxB);
824 if (availableFlagLXA0)
825 mvpcand_list[numMVPCandLX++] = mxA;
827 if (availableFlagLXB0 && (!availableFlagLXA0 || mxA.x != mxB.x || mxA.y != mxB.y))
828 mvpcand_list[numMVPCandLX++] = mxB;
830 //temporal motion vector prediction candidate
831 if (numMVPCandLX < 2 && s->sh.slice_temporal_mvp_enabled_flag &&
832 mvp_lx_flag == numMVPCandLX) {
834 int available_col = temporal_luma_motion_vector(s, x0, y0, nPbW,
838 mvpcand_list[numMVPCandLX++] = mv_col;
841 // insert zero motion vectors when the number of available candidates are less than 2
842 while (numMVPCandLX < 2)
843 mvpcand_list[numMVPCandLX++] = (Mv){ 0, 0 };
845 mv->mv[LX].x = mvpcand_list[mvp_lx_flag].x;
846 mv->mv[LX].y = mvpcand_list[mvp_lx_flag].y;