2 * This file is part of FFmpeg.
4 * FFmpeg is free software; you can redistribute it and/or
5 * modify it under the terms of the GNU Lesser General Public
6 * License as published by the Free Software Foundation; either
7 * version 2.1 of the License, or (at your option) any later version.
9 * FFmpeg is distributed in the hope that it will be useful,
10 * but WITHOUT ANY WARRANTY; without even the implied warranty of
11 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
12 * Lesser General Public License for more details.
14 * You should have received a copy of the GNU Lesser General Public
15 * License along with FFmpeg; if not, write to the Free Software
16 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
19 static int FUNC(rbsp_trailing_bits)(CodedBitstreamContext *ctx, RWContext *rw)
23 fixed(1, rbsp_stop_one_bit, 1);
24 while (byte_alignment(rw) != 0)
25 fixed(1, rbsp_alignment_zero_bit, 0);
30 static int FUNC(nal_unit_header)(CodedBitstreamContext *ctx, RWContext *rw,
31 H264RawNALUnitHeader *current,
32 uint32_t valid_type_mask)
36 fixed(1, forbidden_zero_bit, 0);
40 if (!(1 << current->nal_unit_type & valid_type_mask)) {
41 av_log(ctx->log_ctx, AV_LOG_ERROR, "Invalid NAL unit type %d.\n",
42 current->nal_unit_type);
43 return AVERROR_INVALIDDATA;
46 if (current->nal_unit_type == 14 ||
47 current->nal_unit_type == 20 ||
48 current->nal_unit_type == 21) {
49 if (current->nal_unit_type != 21)
50 flag(svc_extension_flag);
52 flag(avc_3d_extension_flag);
54 if (current->svc_extension_flag) {
55 av_log(ctx->log_ctx, AV_LOG_ERROR, "SVC not supported.\n");
56 return AVERROR_PATCHWELCOME;
58 } else if (current->avc_3d_extension_flag) {
59 av_log(ctx->log_ctx, AV_LOG_ERROR, "3DAVC not supported.\n");
60 return AVERROR_PATCHWELCOME;
63 av_log(ctx->log_ctx, AV_LOG_ERROR, "MVC not supported.\n");
64 return AVERROR_PATCHWELCOME;
71 static int FUNC(scaling_list)(CodedBitstreamContext *ctx, RWContext *rw,
72 H264RawScalingList *current,
73 int size_of_scaling_list)
78 for (i = 0; i < size_of_scaling_list; i++) {
79 ses(delta_scale[i], -128, +127, 1, i);
80 scale = (scale + current->delta_scale[i] + 256) % 256;
88 static int FUNC(hrd_parameters)(CodedBitstreamContext *ctx, RWContext *rw,
93 ue(cpb_cnt_minus1, 0, 31);
94 ub(4, bit_rate_scale);
95 ub(4, cpb_size_scale);
97 for (i = 0; i <= current->cpb_cnt_minus1; i++) {
98 ues(bit_rate_value_minus1[i], 0, UINT32_MAX - 1, 1, i);
99 ues(cpb_size_value_minus1[i], 0, UINT32_MAX - 1, 1, i);
100 flags(cbr_flag[i], 1, i);
103 ub(5, initial_cpb_removal_delay_length_minus1);
104 ub(5, cpb_removal_delay_length_minus1);
105 ub(5, dpb_output_delay_length_minus1);
106 ub(5, time_offset_length);
111 static int FUNC(vui_parameters)(CodedBitstreamContext *ctx, RWContext *rw,
112 H264RawVUI *current, H264RawSPS *sps)
116 flag(aspect_ratio_info_present_flag);
117 if (current->aspect_ratio_info_present_flag) {
118 ub(8, aspect_ratio_idc);
119 if (current->aspect_ratio_idc == 255) {
124 infer(aspect_ratio_idc, 0);
127 flag(overscan_info_present_flag);
128 if (current->overscan_info_present_flag)
129 flag(overscan_appropriate_flag);
131 flag(video_signal_type_present_flag);
132 if (current->video_signal_type_present_flag) {
134 flag(video_full_range_flag);
135 flag(colour_description_present_flag);
136 if (current->colour_description_present_flag) {
137 ub(8, colour_primaries);
138 ub(8, transfer_characteristics);
139 ub(8, matrix_coefficients);
141 infer(colour_primaries, 2);
142 infer(transfer_characteristics, 2);
143 infer(matrix_coefficients, 2);
146 infer(video_format, 5);
147 infer(video_full_range_flag, 0);
148 infer(colour_primaries, 2);
149 infer(transfer_characteristics, 2);
150 infer(matrix_coefficients, 2);
153 flag(chroma_loc_info_present_flag);
154 if (current->chroma_loc_info_present_flag) {
155 ue(chroma_sample_loc_type_top_field, 0, 5);
156 ue(chroma_sample_loc_type_bottom_field, 0, 5);
158 infer(chroma_sample_loc_type_top_field, 0);
159 infer(chroma_sample_loc_type_bottom_field, 0);
162 flag(timing_info_present_flag);
163 if (current->timing_info_present_flag) {
164 u(32, num_units_in_tick, 1, UINT32_MAX);
165 u(32, time_scale, 1, UINT32_MAX);
166 flag(fixed_frame_rate_flag);
168 infer(fixed_frame_rate_flag, 0);
171 flag(nal_hrd_parameters_present_flag);
172 if (current->nal_hrd_parameters_present_flag)
173 CHECK(FUNC(hrd_parameters)(ctx, rw, ¤t->nal_hrd_parameters));
175 flag(vcl_hrd_parameters_present_flag);
176 if (current->vcl_hrd_parameters_present_flag)
177 CHECK(FUNC(hrd_parameters)(ctx, rw, ¤t->vcl_hrd_parameters));
179 if (current->nal_hrd_parameters_present_flag ||
180 current->vcl_hrd_parameters_present_flag)
181 flag(low_delay_hrd_flag);
183 infer(low_delay_hrd_flag, 1 - current->fixed_frame_rate_flag);
185 flag(pic_struct_present_flag);
187 flag(bitstream_restriction_flag);
188 if (current->bitstream_restriction_flag) {
189 flag(motion_vectors_over_pic_boundaries_flag);
190 ue(max_bytes_per_pic_denom, 0, 16);
191 ue(max_bits_per_mb_denom, 0, 16);
192 // The current version of the standard constrains this to be in
193 // [0,15], but older versions allow 16.
194 ue(log2_max_mv_length_horizontal, 0, 16);
195 ue(log2_max_mv_length_vertical, 0, 16);
196 ue(max_num_reorder_frames, 0, H264_MAX_DPB_FRAMES);
197 ue(max_dec_frame_buffering, 0, H264_MAX_DPB_FRAMES);
199 infer(motion_vectors_over_pic_boundaries_flag, 1);
200 infer(max_bytes_per_pic_denom, 2);
201 infer(max_bits_per_mb_denom, 1);
202 infer(log2_max_mv_length_horizontal, 15);
203 infer(log2_max_mv_length_vertical, 15);
205 if ((sps->profile_idc == 44 || sps->profile_idc == 86 ||
206 sps->profile_idc == 100 || sps->profile_idc == 110 ||
207 sps->profile_idc == 122 || sps->profile_idc == 244) &&
208 sps->constraint_set3_flag) {
209 infer(max_num_reorder_frames, 0);
210 infer(max_dec_frame_buffering, 0);
212 infer(max_num_reorder_frames, H264_MAX_DPB_FRAMES);
213 infer(max_dec_frame_buffering, H264_MAX_DPB_FRAMES);
220 static int FUNC(vui_parameters_default)(CodedBitstreamContext *ctx,
221 RWContext *rw, H264RawVUI *current,
224 infer(aspect_ratio_idc, 0);
226 infer(video_format, 5);
227 infer(video_full_range_flag, 0);
228 infer(colour_primaries, 2);
229 infer(transfer_characteristics, 2);
230 infer(matrix_coefficients, 2);
232 infer(chroma_sample_loc_type_top_field, 0);
233 infer(chroma_sample_loc_type_bottom_field, 0);
235 infer(fixed_frame_rate_flag, 0);
236 infer(low_delay_hrd_flag, 1);
238 infer(pic_struct_present_flag, 0);
240 infer(motion_vectors_over_pic_boundaries_flag, 1);
241 infer(max_bytes_per_pic_denom, 2);
242 infer(max_bits_per_mb_denom, 1);
243 infer(log2_max_mv_length_horizontal, 15);
244 infer(log2_max_mv_length_vertical, 15);
246 if ((sps->profile_idc == 44 || sps->profile_idc == 86 ||
247 sps->profile_idc == 100 || sps->profile_idc == 110 ||
248 sps->profile_idc == 122 || sps->profile_idc == 244) &&
249 sps->constraint_set3_flag) {
250 infer(max_num_reorder_frames, 0);
251 infer(max_dec_frame_buffering, 0);
253 infer(max_num_reorder_frames, H264_MAX_DPB_FRAMES);
254 infer(max_dec_frame_buffering, H264_MAX_DPB_FRAMES);
260 static int FUNC(sps)(CodedBitstreamContext *ctx, RWContext *rw,
265 HEADER("Sequence Parameter Set");
267 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
272 flag(constraint_set0_flag);
273 flag(constraint_set1_flag);
274 flag(constraint_set2_flag);
275 flag(constraint_set3_flag);
276 flag(constraint_set4_flag);
277 flag(constraint_set5_flag);
279 u(2, reserved_zero_2bits, 0, 0);
283 ue(seq_parameter_set_id, 0, 31);
285 if (current->profile_idc == 100 || current->profile_idc == 110 ||
286 current->profile_idc == 122 || current->profile_idc == 244 ||
287 current->profile_idc == 44 || current->profile_idc == 83 ||
288 current->profile_idc == 86 || current->profile_idc == 118 ||
289 current->profile_idc == 128 || current->profile_idc == 138) {
290 ue(chroma_format_idc, 0, 3);
292 if (current->chroma_format_idc == 3)
293 flag(separate_colour_plane_flag);
295 infer(separate_colour_plane_flag, 0);
297 ue(bit_depth_luma_minus8, 0, 6);
298 ue(bit_depth_chroma_minus8, 0, 6);
300 flag(qpprime_y_zero_transform_bypass_flag);
302 flag(seq_scaling_matrix_present_flag);
303 if (current->seq_scaling_matrix_present_flag) {
304 for (i = 0; i < ((current->chroma_format_idc != 3) ? 8 : 12); i++) {
305 flags(seq_scaling_list_present_flag[i], 1, i);
306 if (current->seq_scaling_list_present_flag[i]) {
308 CHECK(FUNC(scaling_list)(ctx, rw,
309 ¤t->scaling_list_4x4[i],
312 CHECK(FUNC(scaling_list)(ctx, rw,
313 ¤t->scaling_list_8x8[i - 6],
319 infer(chroma_format_idc, current->profile_idc == 183 ? 0 : 1);
321 infer(separate_colour_plane_flag, 0);
322 infer(bit_depth_luma_minus8, 0);
323 infer(bit_depth_chroma_minus8, 0);
326 ue(log2_max_frame_num_minus4, 0, 12);
327 ue(pic_order_cnt_type, 0, 2);
329 if (current->pic_order_cnt_type == 0) {
330 ue(log2_max_pic_order_cnt_lsb_minus4, 0, 12);
331 } else if (current->pic_order_cnt_type == 1) {
332 flag(delta_pic_order_always_zero_flag);
333 se(offset_for_non_ref_pic, INT32_MIN + 1, INT32_MAX);
334 se(offset_for_top_to_bottom_field, INT32_MIN + 1, INT32_MAX);
335 ue(num_ref_frames_in_pic_order_cnt_cycle, 0, 255);
337 for (i = 0; i < current->num_ref_frames_in_pic_order_cnt_cycle; i++)
338 ses(offset_for_ref_frame[i], INT32_MIN + 1, INT32_MAX, 1, i);
341 ue(max_num_ref_frames, 0, H264_MAX_DPB_FRAMES);
342 flag(gaps_in_frame_num_allowed_flag);
344 ue(pic_width_in_mbs_minus1, 0, H264_MAX_MB_WIDTH);
345 ue(pic_height_in_map_units_minus1, 0, H264_MAX_MB_HEIGHT);
347 flag(frame_mbs_only_flag);
348 if (!current->frame_mbs_only_flag)
349 flag(mb_adaptive_frame_field_flag);
351 flag(direct_8x8_inference_flag);
353 flag(frame_cropping_flag);
354 if (current->frame_cropping_flag) {
355 ue(frame_crop_left_offset, 0, H264_MAX_WIDTH);
356 ue(frame_crop_right_offset, 0, H264_MAX_WIDTH);
357 ue(frame_crop_top_offset, 0, H264_MAX_HEIGHT);
358 ue(frame_crop_bottom_offset, 0, H264_MAX_HEIGHT);
361 flag(vui_parameters_present_flag);
362 if (current->vui_parameters_present_flag)
363 CHECK(FUNC(vui_parameters)(ctx, rw, ¤t->vui, current));
365 CHECK(FUNC(vui_parameters_default)(ctx, rw, ¤t->vui, current));
367 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
372 static int FUNC(sps_extension)(CodedBitstreamContext *ctx, RWContext *rw,
373 H264RawSPSExtension *current)
377 HEADER("Sequence Parameter Set Extension");
379 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
380 1 << H264_NAL_SPS_EXT));
382 ue(seq_parameter_set_id, 0, 31);
384 ue(aux_format_idc, 0, 3);
386 if (current->aux_format_idc != 0) {
389 ue(bit_depth_aux_minus8, 0, 4);
390 flag(alpha_incr_flag);
392 bits = current->bit_depth_aux_minus8 + 9;
393 ub(bits, alpha_opaque_value);
394 ub(bits, alpha_transparent_value);
397 flag(additional_extension_flag);
399 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
404 static int FUNC(pps)(CodedBitstreamContext *ctx, RWContext *rw,
407 CodedBitstreamH264Context *h264 = ctx->priv_data;
408 const H264RawSPS *sps;
411 HEADER("Picture Parameter Set");
413 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
416 ue(pic_parameter_set_id, 0, 255);
417 ue(seq_parameter_set_id, 0, 31);
419 sps = h264->sps[current->seq_parameter_set_id];
421 av_log(ctx->log_ctx, AV_LOG_ERROR, "SPS id %d not available.\n",
422 current->seq_parameter_set_id);
423 return AVERROR_INVALIDDATA;
426 flag(entropy_coding_mode_flag);
427 flag(bottom_field_pic_order_in_frame_present_flag);
429 ue(num_slice_groups_minus1, 0, 7);
430 if (current->num_slice_groups_minus1 > 0) {
431 unsigned int pic_size;
434 pic_size = (sps->pic_width_in_mbs_minus1 + 1) *
435 (sps->pic_height_in_map_units_minus1 + 1);
437 ue(slice_group_map_type, 0, 6);
439 if (current->slice_group_map_type == 0) {
440 for (iGroup = 0; iGroup <= current->num_slice_groups_minus1; iGroup++)
441 ues(run_length_minus1[iGroup], 0, pic_size - 1, 1, iGroup);
443 } else if (current->slice_group_map_type == 2) {
444 for (iGroup = 0; iGroup < current->num_slice_groups_minus1; iGroup++) {
445 ues(top_left[iGroup], 0, pic_size - 1, 1, iGroup);
446 ues(bottom_right[iGroup],
447 current->top_left[iGroup], pic_size - 1, 1, iGroup);
449 } else if (current->slice_group_map_type == 3 ||
450 current->slice_group_map_type == 4 ||
451 current->slice_group_map_type == 5) {
452 flag(slice_group_change_direction_flag);
453 ue(slice_group_change_rate_minus1, 0, pic_size - 1);
454 } else if (current->slice_group_map_type == 6) {
455 ue(pic_size_in_map_units_minus1, pic_size - 1, pic_size - 1);
457 allocate(current->slice_group_id,
458 current->pic_size_in_map_units_minus1 + 1);
459 for (i = 0; i <= current->pic_size_in_map_units_minus1; i++)
460 us(av_log2(2 * current->num_slice_groups_minus1 + 1),
461 slice_group_id[i], 0, current->num_slice_groups_minus1, 1, i);
465 ue(num_ref_idx_l0_default_active_minus1, 0, 31);
466 ue(num_ref_idx_l1_default_active_minus1, 0, 31);
468 flag(weighted_pred_flag);
469 u(2, weighted_bipred_idc, 0, 2);
471 se(pic_init_qp_minus26, -26 - 6 * sps->bit_depth_luma_minus8, +25);
472 se(pic_init_qs_minus26, -26, +25);
473 se(chroma_qp_index_offset, -12, +12);
475 flag(deblocking_filter_control_present_flag);
476 flag(constrained_intra_pred_flag);
477 flag(redundant_pic_cnt_present_flag);
479 if (more_rbsp_data(current->more_rbsp_data))
481 flag(transform_8x8_mode_flag);
483 flag(pic_scaling_matrix_present_flag);
484 if (current->pic_scaling_matrix_present_flag) {
485 for (i = 0; i < 6 + (((sps->chroma_format_idc != 3) ? 2 : 6) *
486 current->transform_8x8_mode_flag); i++) {
487 flags(pic_scaling_list_present_flag[i], 1, i);
488 if (current->pic_scaling_list_present_flag[i]) {
490 CHECK(FUNC(scaling_list)(ctx, rw,
491 ¤t->scaling_list_4x4[i],
494 CHECK(FUNC(scaling_list)(ctx, rw,
495 ¤t->scaling_list_8x8[i - 6],
501 se(second_chroma_qp_index_offset, -12, +12);
503 infer(transform_8x8_mode_flag, 0);
504 infer(pic_scaling_matrix_present_flag, 0);
505 infer(second_chroma_qp_index_offset, current->chroma_qp_index_offset);
508 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
513 static int FUNC(sei_buffering_period)(CodedBitstreamContext *ctx, RWContext *rw,
514 H264RawSEIBufferingPeriod *current)
516 CodedBitstreamH264Context *h264 = ctx->priv_data;
517 const H264RawSPS *sps;
520 HEADER("Buffering Period");
522 ue(seq_parameter_set_id, 0, 31);
524 sps = h264->sps[current->seq_parameter_set_id];
526 av_log(ctx->log_ctx, AV_LOG_ERROR, "SPS id %d not available.\n",
527 current->seq_parameter_set_id);
528 return AVERROR_INVALIDDATA;
530 h264->active_sps = sps;
532 if (sps->vui.nal_hrd_parameters_present_flag) {
533 for (i = 0; i <= sps->vui.nal_hrd_parameters.cpb_cnt_minus1; i++) {
534 length = sps->vui.nal_hrd_parameters.initial_cpb_removal_delay_length_minus1 + 1;
535 xu(length, initial_cpb_removal_delay[SchedSelIdx],
536 current->nal.initial_cpb_removal_delay[i],
537 1, MAX_UINT_BITS(length), 1, i);
538 xu(length, initial_cpb_removal_delay_offset[SchedSelIdx],
539 current->nal.initial_cpb_removal_delay_offset[i],
540 0, MAX_UINT_BITS(length), 1, i);
544 if (sps->vui.vcl_hrd_parameters_present_flag) {
545 for (i = 0; i <= sps->vui.vcl_hrd_parameters.cpb_cnt_minus1; i++) {
546 length = sps->vui.vcl_hrd_parameters.initial_cpb_removal_delay_length_minus1 + 1;
547 xu(length, initial_cpb_removal_delay[SchedSelIdx],
548 current->vcl.initial_cpb_removal_delay[i],
549 1, MAX_UINT_BITS(length), 1, i);
550 xu(length, initial_cpb_removal_delay_offset[SchedSelIdx],
551 current->vcl.initial_cpb_removal_delay_offset[i],
552 0, MAX_UINT_BITS(length), 1, i);
559 static int FUNC(sei_pic_timestamp)(CodedBitstreamContext *ctx, RWContext *rw,
560 H264RawSEIPicTimestamp *current,
561 const H264RawSPS *sps)
563 uint8_t time_offset_length;
567 flag(nuit_field_based_flag);
568 u(5, counting_type, 0, 6);
569 flag(full_timestamp_flag);
570 flag(discontinuity_flag);
571 flag(cnt_dropped_flag);
573 if (current->full_timestamp_flag) {
574 u(6, seconds_value, 0, 59);
575 u(6, minutes_value, 0, 59);
576 u(5, hours_value, 0, 23);
579 if (current->seconds_flag) {
580 u(6, seconds_value, 0, 59);
582 if (current->minutes_flag) {
583 u(6, minutes_value, 0, 59);
585 if (current->hours_flag)
586 u(5, hours_value, 0, 23);
591 if (sps->vui.nal_hrd_parameters_present_flag)
592 time_offset_length = sps->vui.nal_hrd_parameters.time_offset_length;
593 else if (sps->vui.vcl_hrd_parameters_present_flag)
594 time_offset_length = sps->vui.vcl_hrd_parameters.time_offset_length;
596 time_offset_length = 24;
598 if (time_offset_length > 0)
599 ib(time_offset_length, time_offset);
601 infer(time_offset, 0);
606 static int FUNC(sei_pic_timing)(CodedBitstreamContext *ctx, RWContext *rw,
607 H264RawSEIPicTiming *current)
609 CodedBitstreamH264Context *h264 = ctx->priv_data;
610 const H264RawSPS *sps;
613 HEADER("Picture Timing");
615 sps = h264->active_sps;
617 // If there is exactly one possible SPS but it is not yet active
618 // then just assume that it should be the active one.
620 for (i = 0; i < H264_MAX_SPS_COUNT; i++) {
633 av_log(ctx->log_ctx, AV_LOG_ERROR,
634 "No active SPS for pic_timing.\n");
635 return AVERROR_INVALIDDATA;
638 if (sps->vui.nal_hrd_parameters_present_flag ||
639 sps->vui.vcl_hrd_parameters_present_flag) {
640 const H264RawHRD *hrd;
642 if (sps->vui.nal_hrd_parameters_present_flag)
643 hrd = &sps->vui.nal_hrd_parameters;
644 else if (sps->vui.vcl_hrd_parameters_present_flag)
645 hrd = &sps->vui.vcl_hrd_parameters;
647 av_log(ctx->log_ctx, AV_LOG_ERROR,
648 "No HRD parameters for pic_timing.\n");
649 return AVERROR_INVALIDDATA;
652 ub(hrd->cpb_removal_delay_length_minus1 + 1, cpb_removal_delay);
653 ub(hrd->dpb_output_delay_length_minus1 + 1, dpb_output_delay);
656 if (sps->vui.pic_struct_present_flag) {
657 static const uint8_t num_clock_ts[9] = {
658 1, 1, 1, 2, 2, 3, 3, 2, 3
662 u(4, pic_struct, 0, 8);
663 if (current->pic_struct > 8)
664 return AVERROR_INVALIDDATA;
666 for (i = 0; i < num_clock_ts[current->pic_struct]; i++) {
667 flags(clock_timestamp_flag[i], 1, i);
668 if (current->clock_timestamp_flag[i])
669 CHECK(FUNC(sei_pic_timestamp)(ctx, rw,
670 ¤t->timestamp[i], sps));
677 static int FUNC(sei_pan_scan_rect)(CodedBitstreamContext *ctx, RWContext *rw,
678 H264RawSEIPanScanRect *current)
682 HEADER("Pan-Scan Rectangle");
684 ue(pan_scan_rect_id, 0, UINT32_MAX - 1);
685 flag(pan_scan_rect_cancel_flag);
687 if (!current->pan_scan_rect_cancel_flag) {
688 ue(pan_scan_cnt_minus1, 0, 2);
690 for (i = 0; i <= current->pan_scan_cnt_minus1; i++) {
691 ses(pan_scan_rect_left_offset[i], INT32_MIN + 1, INT32_MAX, 1, i);
692 ses(pan_scan_rect_right_offset[i], INT32_MIN + 1, INT32_MAX, 1, i);
693 ses(pan_scan_rect_top_offset[i], INT32_MIN + 1, INT32_MAX, 1, i);
694 ses(pan_scan_rect_bottom_offset[i], INT32_MIN + 1, INT32_MAX, 1, i);
697 ue(pan_scan_rect_repetition_period, 0, 16384);
703 static int FUNC(sei_user_data_registered)(CodedBitstreamContext *ctx, RWContext *rw,
704 H264RawSEIUserDataRegistered *current,
705 uint32_t *payload_size)
709 HEADER("User Data Registered ITU-T T.35");
711 u(8, itu_t_t35_country_code, 0x00, 0xff);
712 if (current->itu_t_t35_country_code != 0xff)
715 u(8, itu_t_t35_country_code_extension_byte, 0x00, 0xff);
720 if (*payload_size < i) {
721 av_log(ctx->log_ctx, AV_LOG_ERROR,
722 "Invalid SEI user data registered payload.\n");
723 return AVERROR_INVALIDDATA;
725 current->data_length = *payload_size - i;
727 *payload_size = i + current->data_length;
730 allocate(current->data, current->data_length);
731 for (j = 0; j < current->data_length; j++)
732 xu(8, itu_t_t35_payload_byte[i], current->data[j], 0x00, 0xff, 1, i + j);
737 static int FUNC(sei_user_data_unregistered)(CodedBitstreamContext *ctx, RWContext *rw,
738 H264RawSEIUserDataUnregistered *current,
739 uint32_t *payload_size)
743 HEADER("User Data Unregistered");
746 if (*payload_size < 16) {
747 av_log(ctx->log_ctx, AV_LOG_ERROR,
748 "Invalid SEI user data unregistered payload.\n");
749 return AVERROR_INVALIDDATA;
751 current->data_length = *payload_size - 16;
753 *payload_size = 16 + current->data_length;
756 for (i = 0; i < 16; i++)
757 us(8, uuid_iso_iec_11578[i], 0x00, 0xff, 1, i);
759 allocate(current->data, current->data_length);
761 for (i = 0; i < current->data_length; i++)
762 xu(8, user_data_payload_byte[i], current->data[i], 0x00, 0xff, 1, i);
767 static int FUNC(sei_recovery_point)(CodedBitstreamContext *ctx, RWContext *rw,
768 H264RawSEIRecoveryPoint *current)
772 HEADER("Recovery Point");
774 ue(recovery_frame_cnt, 0, 65535);
775 flag(exact_match_flag);
776 flag(broken_link_flag);
777 u(2, changing_slice_group_idc, 0, 2);
782 static int FUNC(sei_display_orientation)(CodedBitstreamContext *ctx, RWContext *rw,
783 H264RawSEIDisplayOrientation *current)
787 HEADER("Display Orientation");
789 flag(display_orientation_cancel_flag);
790 if (!current->display_orientation_cancel_flag) {
793 ub(16, anticlockwise_rotation);
794 ue(display_orientation_repetition_period, 0, 16384);
795 flag(display_orientation_extension_flag);
801 static int FUNC(sei_mastering_display_colour_volume)(CodedBitstreamContext *ctx, RWContext *rw,
802 H264RawSEIMasteringDisplayColourVolume *current)
806 HEADER("Mastering Display Colour Volume");
808 for (c = 0; c < 3; c++) {
809 us(16, display_primaries_x[c], 0, 50000, 1, c);
810 us(16, display_primaries_y[c], 0, 50000, 1, c);
813 u(16, white_point_x, 0, 50000);
814 u(16, white_point_y, 0, 50000);
816 u(32, max_display_mastering_luminance, 1, MAX_UINT_BITS(32));
817 u(32, min_display_mastering_luminance, 0, current->max_display_mastering_luminance - 1);
822 static int FUNC(sei_alternative_transfer_characteristics)(CodedBitstreamContext *ctx,
824 H264RawSEIAlternativeTransferCharacteristics *current)
828 HEADER("Alternative Transfer Characteristics");
830 ub(8, preferred_transfer_characteristics);
835 static int FUNC(sei_payload)(CodedBitstreamContext *ctx, RWContext *rw,
836 H264RawSEIPayload *current)
839 int start_position, end_position;
842 start_position = get_bits_count(rw);
844 start_position = put_bits_count(rw);
847 switch (current->payload_type) {
848 case H264_SEI_TYPE_BUFFERING_PERIOD:
849 CHECK(FUNC(sei_buffering_period)
850 (ctx, rw, ¤t->payload.buffering_period));
852 case H264_SEI_TYPE_PIC_TIMING:
853 CHECK(FUNC(sei_pic_timing)
854 (ctx, rw, ¤t->payload.pic_timing));
856 case H264_SEI_TYPE_PAN_SCAN_RECT:
857 CHECK(FUNC(sei_pan_scan_rect)
858 (ctx, rw, ¤t->payload.pan_scan_rect));
860 case H264_SEI_TYPE_FILLER_PAYLOAD:
862 for (i = 0; i < current->payload_size; i++)
863 fixed(8, ff_byte, 0xff);
866 case H264_SEI_TYPE_USER_DATA_REGISTERED:
867 CHECK(FUNC(sei_user_data_registered)
868 (ctx, rw, ¤t->payload.user_data_registered, ¤t->payload_size));
870 case H264_SEI_TYPE_USER_DATA_UNREGISTERED:
871 CHECK(FUNC(sei_user_data_unregistered)
872 (ctx, rw, ¤t->payload.user_data_unregistered, ¤t->payload_size));
874 case H264_SEI_TYPE_RECOVERY_POINT:
875 CHECK(FUNC(sei_recovery_point)
876 (ctx, rw, ¤t->payload.recovery_point));
878 case H264_SEI_TYPE_DISPLAY_ORIENTATION:
879 CHECK(FUNC(sei_display_orientation)
880 (ctx, rw, ¤t->payload.display_orientation));
882 case H264_SEI_TYPE_MASTERING_DISPLAY_COLOUR_VOLUME:
883 CHECK(FUNC(sei_mastering_display_colour_volume)
884 (ctx, rw, ¤t->payload.mastering_display_colour_volume));
886 case H264_SEI_TYPE_ALTERNATIVE_TRANSFER:
887 CHECK(FUNC(sei_alternative_transfer_characteristics)
888 (ctx, rw, ¤t->payload.alternative_transfer_characteristics));
893 current->payload.other.data_length = current->payload_size;
895 allocate(current->payload.other.data, current->payload.other.data_length);
896 for (i = 0; i < current->payload.other.data_length; i++)
897 xu(8, payload_byte[i], current->payload.other.data[i], 0, 255, 1, i);
901 if (byte_alignment(rw)) {
902 fixed(1, bit_equal_to_one, 1);
903 while (byte_alignment(rw))
904 fixed(1, bit_equal_to_zero, 0);
908 end_position = get_bits_count(rw);
909 if (end_position < start_position + 8 * current->payload_size) {
910 av_log(ctx->log_ctx, AV_LOG_ERROR, "Incorrect SEI payload length: "
911 "header %"PRIu32" bits, actually %d bits.\n",
912 8 * current->payload_size,
913 end_position - start_position);
914 return AVERROR_INVALIDDATA;
917 end_position = put_bits_count(rw);
918 current->payload_size = (end_position - start_position) / 8;
924 static int FUNC(sei)(CodedBitstreamContext *ctx, RWContext *rw,
929 HEADER("Supplemental Enhancement Information");
931 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
935 for (k = 0; k < H264_MAX_SEI_PAYLOADS; k++) {
936 uint32_t payload_type = 0;
937 uint32_t payload_size = 0;
940 while (show_bits(rw, 8) == 0xff) {
941 fixed(8, ff_byte, 0xff);
944 xu(8, last_payload_type_byte, tmp, 0, 254, 0);
947 while (show_bits(rw, 8) == 0xff) {
948 fixed(8, ff_byte, 0xff);
951 xu(8, last_payload_size_byte, tmp, 0, 254, 0);
954 current->payload[k].payload_type = payload_type;
955 current->payload[k].payload_size = payload_size;
957 current->payload_count++;
958 CHECK(FUNC(sei_payload)(ctx, rw, ¤t->payload[k]));
960 if (!cbs_h2645_read_more_rbsp_data(rw))
963 if (k >= H264_MAX_SEI_PAYLOADS) {
964 av_log(ctx->log_ctx, AV_LOG_ERROR, "Too many payloads in "
965 "SEI message: found %d.\n", k);
966 return AVERROR_INVALIDDATA;
969 for (k = 0; k < current->payload_count; k++) {
970 PutBitContext start_state;
974 // Somewhat clumsy: we write the payload twice when
975 // we don't know the size in advance. This will mess
976 // with trace output, but is otherwise harmless.
978 need_size = !current->payload[k].payload_size;
979 for (i = 0; i < 1 + need_size; i++) {
982 tmp = current->payload[k].payload_type;
984 fixed(8, ff_byte, 0xff);
987 xu(8, last_payload_type_byte, tmp, 0, 254, 0);
989 tmp = current->payload[k].payload_size;
991 fixed(8, ff_byte, 0xff);
994 xu(8, last_payload_size_byte, tmp, 0, 254, 0);
996 CHECK(FUNC(sei_payload)(ctx, rw, ¤t->payload[k]));
1001 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
1006 static int FUNC(aud)(CodedBitstreamContext *ctx, RWContext *rw,
1007 H264RawAUD *current)
1011 HEADER("Access Unit Delimiter");
1013 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
1014 1 << H264_NAL_AUD));
1016 ub(3, primary_pic_type);
1018 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
1023 static int FUNC(ref_pic_list_modification)(CodedBitstreamContext *ctx, RWContext *rw,
1024 H264RawSliceHeader *current)
1026 CodedBitstreamH264Context *h264 = ctx->priv_data;
1027 const H264RawSPS *sps = h264->active_sps;
1030 if (current->slice_type % 5 != 2 &&
1031 current->slice_type % 5 != 4) {
1032 flag(ref_pic_list_modification_flag_l0);
1033 if (current->ref_pic_list_modification_flag_l0) {
1034 for (i = 0; i < H264_MAX_RPLM_COUNT; i++) {
1035 xue(modification_of_pic_nums_idc,
1036 current->rplm_l0[i].modification_of_pic_nums_idc, 0, 3, 0);
1038 mopn = current->rplm_l0[i].modification_of_pic_nums_idc;
1042 if (mopn == 0 || mopn == 1)
1043 xue(abs_diff_pic_num_minus1,
1044 current->rplm_l0[i].abs_diff_pic_num_minus1,
1045 0, (1 + current->field_pic_flag) *
1046 (1 << (sps->log2_max_frame_num_minus4 + 4)), 0);
1048 xue(long_term_pic_num,
1049 current->rplm_l0[i].long_term_pic_num,
1050 0, sps->max_num_ref_frames - 1, 0);
1055 if (current->slice_type % 5 == 1) {
1056 flag(ref_pic_list_modification_flag_l1);
1057 if (current->ref_pic_list_modification_flag_l1) {
1058 for (i = 0; i < H264_MAX_RPLM_COUNT; i++) {
1059 xue(modification_of_pic_nums_idc,
1060 current->rplm_l1[i].modification_of_pic_nums_idc, 0, 3, 0);
1062 mopn = current->rplm_l1[i].modification_of_pic_nums_idc;
1066 if (mopn == 0 || mopn == 1)
1067 xue(abs_diff_pic_num_minus1,
1068 current->rplm_l1[i].abs_diff_pic_num_minus1,
1069 0, (1 + current->field_pic_flag) *
1070 (1 << (sps->log2_max_frame_num_minus4 + 4)), 0);
1072 xue(long_term_pic_num,
1073 current->rplm_l1[i].long_term_pic_num,
1074 0, sps->max_num_ref_frames - 1, 0);
1082 static int FUNC(pred_weight_table)(CodedBitstreamContext *ctx, RWContext *rw,
1083 H264RawSliceHeader *current)
1085 CodedBitstreamH264Context *h264 = ctx->priv_data;
1086 const H264RawSPS *sps = h264->active_sps;
1090 ue(luma_log2_weight_denom, 0, 7);
1092 chroma = !sps->separate_colour_plane_flag && sps->chroma_format_idc != 0;
1094 ue(chroma_log2_weight_denom, 0, 7);
1096 for (i = 0; i <= current->num_ref_idx_l0_active_minus1; i++) {
1097 flags(luma_weight_l0_flag[i], 1, i);
1098 if (current->luma_weight_l0_flag[i]) {
1099 ses(luma_weight_l0[i], -128, +127, 1, i);
1100 ses(luma_offset_l0[i], -128, +127, 1, i);
1103 flags(chroma_weight_l0_flag[i], 1, i);
1104 if (current->chroma_weight_l0_flag[i]) {
1105 for (j = 0; j < 2; j++) {
1106 ses(chroma_weight_l0[i][j], -128, +127, 2, i, j);
1107 ses(chroma_offset_l0[i][j], -128, +127, 2, i, j);
1113 if (current->slice_type % 5 == 1) {
1114 for (i = 0; i <= current->num_ref_idx_l1_active_minus1; i++) {
1115 flags(luma_weight_l1_flag[i], 1, i);
1116 if (current->luma_weight_l1_flag[i]) {
1117 ses(luma_weight_l1[i], -128, +127, 1, i);
1118 ses(luma_offset_l1[i], -128, +127, 1, i);
1121 flags(chroma_weight_l1_flag[i], 1, i);
1122 if (current->chroma_weight_l1_flag[i]) {
1123 for (j = 0; j < 2; j++) {
1124 ses(chroma_weight_l1[i][j], -128, +127, 2, i, j);
1125 ses(chroma_offset_l1[i][j], -128, +127, 2, i, j);
1135 static int FUNC(dec_ref_pic_marking)(CodedBitstreamContext *ctx, RWContext *rw,
1136 H264RawSliceHeader *current, int idr_pic_flag)
1138 CodedBitstreamH264Context *h264 = ctx->priv_data;
1139 const H264RawSPS *sps = h264->active_sps;
1144 flag(no_output_of_prior_pics_flag);
1145 flag(long_term_reference_flag);
1147 flag(adaptive_ref_pic_marking_mode_flag);
1148 if (current->adaptive_ref_pic_marking_mode_flag) {
1149 for (i = 0; i < H264_MAX_MMCO_COUNT; i++) {
1150 xue(memory_management_control_operation,
1151 current->mmco[i].memory_management_control_operation,
1154 mmco = current->mmco[i].memory_management_control_operation;
1158 if (mmco == 1 || mmco == 3)
1159 xue(difference_of_pic_nums_minus1,
1160 current->mmco[i].difference_of_pic_nums_minus1,
1163 xue(long_term_pic_num,
1164 current->mmco[i].long_term_pic_num,
1165 0, sps->max_num_ref_frames - 1, 0);
1166 if (mmco == 3 || mmco == 6)
1167 xue(long_term_frame_idx,
1168 current->mmco[i].long_term_frame_idx,
1169 0, sps->max_num_ref_frames - 1, 0);
1171 xue(max_long_term_frame_idx_plus1,
1172 current->mmco[i].max_long_term_frame_idx_plus1,
1173 0, sps->max_num_ref_frames, 0);
1175 if (i == H264_MAX_MMCO_COUNT) {
1176 av_log(ctx->log_ctx, AV_LOG_ERROR, "Too many "
1177 "memory management control operations.\n");
1178 return AVERROR_INVALIDDATA;
1186 static int FUNC(slice_header)(CodedBitstreamContext *ctx, RWContext *rw,
1187 H264RawSliceHeader *current)
1189 CodedBitstreamH264Context *h264 = ctx->priv_data;
1190 const H264RawSPS *sps;
1191 const H264RawPPS *pps;
1194 int slice_type_i, slice_type_p, slice_type_b;
1195 int slice_type_si, slice_type_sp;
1197 HEADER("Slice Header");
1199 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
1200 1 << H264_NAL_SLICE |
1201 1 << H264_NAL_IDR_SLICE |
1202 1 << H264_NAL_AUXILIARY_SLICE));
1204 if (current->nal_unit_header.nal_unit_type == H264_NAL_AUXILIARY_SLICE) {
1205 if (!h264->last_slice_nal_unit_type) {
1206 av_log(ctx->log_ctx, AV_LOG_ERROR, "Auxiliary slice "
1207 "is not decodable without the main picture "
1208 "in the same access unit.\n");
1209 return AVERROR_INVALIDDATA;
1211 idr_pic_flag = h264->last_slice_nal_unit_type == H264_NAL_IDR_SLICE;
1213 idr_pic_flag = current->nal_unit_header.nal_unit_type == H264_NAL_IDR_SLICE;
1216 ue(first_mb_in_slice, 0, H264_MAX_MB_PIC_SIZE - 1);
1217 ue(slice_type, 0, 9);
1219 slice_type_i = current->slice_type % 5 == 2;
1220 slice_type_p = current->slice_type % 5 == 0;
1221 slice_type_b = current->slice_type % 5 == 1;
1222 slice_type_si = current->slice_type % 5 == 4;
1223 slice_type_sp = current->slice_type % 5 == 3;
1225 if (idr_pic_flag && !(slice_type_i || slice_type_si)) {
1226 av_log(ctx->log_ctx, AV_LOG_ERROR, "Invalid slice type %d "
1227 "for IDR picture.\n", current->slice_type);
1228 return AVERROR_INVALIDDATA;
1231 ue(pic_parameter_set_id, 0, 255);
1233 pps = h264->pps[current->pic_parameter_set_id];
1235 av_log(ctx->log_ctx, AV_LOG_ERROR, "PPS id %d not available.\n",
1236 current->pic_parameter_set_id);
1237 return AVERROR_INVALIDDATA;
1239 h264->active_pps = pps;
1241 sps = h264->sps[pps->seq_parameter_set_id];
1243 av_log(ctx->log_ctx, AV_LOG_ERROR, "SPS id %d not available.\n",
1244 pps->seq_parameter_set_id);
1245 return AVERROR_INVALIDDATA;
1247 h264->active_sps = sps;
1249 if (sps->separate_colour_plane_flag)
1250 u(2, colour_plane_id, 0, 2);
1252 ub(sps->log2_max_frame_num_minus4 + 4, frame_num);
1254 if (!sps->frame_mbs_only_flag) {
1255 flag(field_pic_flag);
1256 if (current->field_pic_flag)
1257 flag(bottom_field_flag);
1259 infer(bottom_field_flag, 0);
1261 infer(field_pic_flag, 0);
1262 infer(bottom_field_flag, 0);
1266 ue(idr_pic_id, 0, 65535);
1268 if (sps->pic_order_cnt_type == 0) {
1269 ub(sps->log2_max_pic_order_cnt_lsb_minus4 + 4, pic_order_cnt_lsb);
1270 if (pps->bottom_field_pic_order_in_frame_present_flag &&
1271 !current->field_pic_flag)
1272 se(delta_pic_order_cnt_bottom, INT32_MIN + 1, INT32_MAX);
1274 } else if (sps->pic_order_cnt_type == 1) {
1275 if (!sps->delta_pic_order_always_zero_flag) {
1276 se(delta_pic_order_cnt[0], INT32_MIN + 1, INT32_MAX);
1277 if (pps->bottom_field_pic_order_in_frame_present_flag &&
1278 !current->field_pic_flag)
1279 se(delta_pic_order_cnt[1], INT32_MIN + 1, INT32_MAX);
1281 infer(delta_pic_order_cnt[1], 0);
1283 infer(delta_pic_order_cnt[0], 0);
1284 infer(delta_pic_order_cnt[1], 0);
1288 if (pps->redundant_pic_cnt_present_flag)
1289 ue(redundant_pic_cnt, 0, 127);
1291 infer(redundant_pic_cnt, 0);
1293 if (current->nal_unit_header.nal_unit_type != H264_NAL_AUXILIARY_SLICE
1294 && !current->redundant_pic_cnt)
1295 h264->last_slice_nal_unit_type =
1296 current->nal_unit_header.nal_unit_type;
1299 flag(direct_spatial_mv_pred_flag);
1301 if (slice_type_p || slice_type_sp || slice_type_b) {
1302 flag(num_ref_idx_active_override_flag);
1303 if (current->num_ref_idx_active_override_flag) {
1304 ue(num_ref_idx_l0_active_minus1, 0, 31);
1306 ue(num_ref_idx_l1_active_minus1, 0, 31);
1308 infer(num_ref_idx_l0_active_minus1,
1309 pps->num_ref_idx_l0_default_active_minus1);
1310 infer(num_ref_idx_l1_active_minus1,
1311 pps->num_ref_idx_l1_default_active_minus1);
1315 if (current->nal_unit_header.nal_unit_type == 20 ||
1316 current->nal_unit_header.nal_unit_type == 21) {
1317 av_log(ctx->log_ctx, AV_LOG_ERROR, "MVC / 3DAVC not supported.\n");
1318 return AVERROR_PATCHWELCOME;
1320 CHECK(FUNC(ref_pic_list_modification)(ctx, rw, current));
1323 if ((pps->weighted_pred_flag && (slice_type_p || slice_type_sp)) ||
1324 (pps->weighted_bipred_idc == 1 && slice_type_b)) {
1325 CHECK(FUNC(pred_weight_table)(ctx, rw, current));
1328 if (current->nal_unit_header.nal_ref_idc != 0) {
1329 CHECK(FUNC(dec_ref_pic_marking)(ctx, rw, current, idr_pic_flag));
1332 if (pps->entropy_coding_mode_flag &&
1333 !slice_type_i && !slice_type_si) {
1334 ue(cabac_init_idc, 0, 2);
1337 se(slice_qp_delta, - 51 - 6 * sps->bit_depth_luma_minus8,
1338 + 51 + 6 * sps->bit_depth_luma_minus8);
1339 if (slice_type_sp || slice_type_si) {
1341 flag(sp_for_switch_flag);
1342 se(slice_qs_delta, -51, +51);
1345 if (pps->deblocking_filter_control_present_flag) {
1346 ue(disable_deblocking_filter_idc, 0, 2);
1347 if (current->disable_deblocking_filter_idc != 1) {
1348 se(slice_alpha_c0_offset_div2, -6, +6);
1349 se(slice_beta_offset_div2, -6, +6);
1351 infer(slice_alpha_c0_offset_div2, 0);
1352 infer(slice_beta_offset_div2, 0);
1355 infer(disable_deblocking_filter_idc, 0);
1356 infer(slice_alpha_c0_offset_div2, 0);
1357 infer(slice_beta_offset_div2, 0);
1360 if (pps->num_slice_groups_minus1 > 0 &&
1361 pps->slice_group_map_type >= 3 &&
1362 pps->slice_group_map_type <= 5) {
1363 unsigned int pic_size, max, bits;
1365 pic_size = (sps->pic_width_in_mbs_minus1 + 1) *
1366 (sps->pic_height_in_map_units_minus1 + 1);
1367 max = (pic_size + pps->slice_group_change_rate_minus1) /
1368 (pps->slice_group_change_rate_minus1 + 1);
1369 bits = av_ceil_log2(max + 1);
1371 u(bits, slice_group_change_cycle, 0, max);
1374 if (pps->entropy_coding_mode_flag) {
1375 while (byte_alignment(rw))
1376 fixed(1, cabac_alignment_one_bit, 1);
1382 static int FUNC(filler)(CodedBitstreamContext *ctx, RWContext *rw,
1383 H264RawFiller *current)
1387 HEADER("Filler Data");
1389 CHECK(FUNC(nal_unit_header)(ctx, rw, ¤t->nal_unit_header,
1390 1 << H264_NAL_FILLER_DATA));
1393 while (show_bits(rw, 8) == 0xff) {
1394 fixed(8, ff_byte, 0xff);
1395 ++current->filler_size;
1400 for (i = 0; i < current->filler_size; i++)
1401 fixed(8, ff_byte, 0xff);
1405 CHECK(FUNC(rbsp_trailing_bits)(ctx, rw));
1410 static int FUNC(end_of_sequence)(CodedBitstreamContext *ctx, RWContext *rw,
1411 H264RawNALUnitHeader *current)
1413 HEADER("End of Sequence");
1415 return FUNC(nal_unit_header)(ctx, rw, current,
1416 1 << H264_NAL_END_SEQUENCE);
1419 static int FUNC(end_of_stream)(CodedBitstreamContext *ctx, RWContext *rw,
1420 H264RawNALUnitHeader *current)
1422 HEADER("End of Stream");
1424 return FUNC(nal_unit_header)(ctx, rw, current,
1425 1 << H264_NAL_END_STREAM);