/**
* @file
- * H.264 / AVC / MPEG4 part10 parser.
+ * H.264 / AVC / MPEG-4 part10 parser.
* @author Michael Niedermayer <michaelni@gmx.at>
*/
#include "get_bits.h"
#include "golomb.h"
#include "h264.h"
+#include "h264dec.h"
+#include "h264_sei.h"
+#include "h264_ps.h"
#include "h264data.h"
#include "internal.h"
#include "mpegutils.h"
#include "parser.h"
typedef struct H264ParseContext {
- H264Context h;
ParseContext pc;
+ H264ParamSets ps;
+ H264DSPContext h264dsp;
+ H264POCContext poc;
+ H264SEIContext sei;
+ int is_avc;
+ int nal_length_size;
int got_first;
+ int picture_structure;
} H264ParseContext;
static int h264_find_frame_end(H264ParseContext *p, const uint8_t *buf,
int buf_size)
{
- H264Context *h = &p->h;
int i;
uint32_t state;
ParseContext *pc = &p->pc;
for (i = 0; i < buf_size; i++) {
if (state == 7) {
- i += h->h264dsp.startcode_find_candidate(buf + i, buf_size - i);
+ i += p->h264dsp.startcode_find_candidate(buf + i, buf_size - i);
if (i < buf_size)
state = 2;
} else if (state <= 2) {
return i - (state & 5);
}
-static int scan_mmco_reset(AVCodecParserContext *s)
+static int scan_mmco_reset(AVCodecParserContext *s, GetBitContext *gb,
+ AVCodecContext *avctx)
{
+ H264PredWeightTable pwt;
+ int slice_type_nos = s->pict_type & 3;
H264ParseContext *p = s->priv_data;
- H264Context *h = &p->h;
- H264SliceContext *sl = &h->slice_ctx[0];
+ int list_count, ref_count[2];
- sl->slice_type_nos = s->pict_type & 3;
- if (h->pps.redundant_pic_cnt_present)
- get_ue_golomb(&sl->gb); // redundant_pic_count
+ if (p->ps.pps->redundant_pic_cnt_present)
+ get_ue_golomb(gb); // redundant_pic_count
- if (ff_set_ref_count(h, sl) < 0)
+ if (slice_type_nos == AV_PICTURE_TYPE_B)
+ get_bits1(gb); // direct_spatial_mv_pred
+
+ if (ff_h264_parse_ref_count(&list_count, ref_count, gb, p->ps.pps,
+ slice_type_nos, p->picture_structure) < 0)
return AVERROR_INVALIDDATA;
- if (sl->slice_type_nos != AV_PICTURE_TYPE_I) {
+ if (slice_type_nos != AV_PICTURE_TYPE_I) {
int list;
- for (list = 0; list < sl->list_count; list++) {
- if (get_bits1(&sl->gb)) {
+ for (list = 0; list < list_count; list++) {
+ if (get_bits1(gb)) {
int index;
for (index = 0; ; index++) {
- unsigned int reordering_of_pic_nums_idc = get_ue_golomb_31(&sl->gb);
+ unsigned int reordering_of_pic_nums_idc = get_ue_golomb_31(gb);
if (reordering_of_pic_nums_idc < 3)
- get_ue_golomb(&sl->gb);
+ get_ue_golomb(gb);
else if (reordering_of_pic_nums_idc > 3) {
- av_log(h->avctx, AV_LOG_ERROR,
+ av_log(avctx, AV_LOG_ERROR,
"illegal reordering_of_pic_nums_idc %d\n",
reordering_of_pic_nums_idc);
return AVERROR_INVALIDDATA;
} else
break;
- if (index >= sl->ref_count[list]) {
- av_log(h->avctx, AV_LOG_ERROR,
+ if (index >= ref_count[list]) {
+ av_log(avctx, AV_LOG_ERROR,
"reference count %d overflow\n", index);
return AVERROR_INVALIDDATA;
}
}
}
- if ((h->pps.weighted_pred && sl->slice_type_nos == AV_PICTURE_TYPE_P) ||
- (h->pps.weighted_bipred_idc == 1 && sl->slice_type_nos == AV_PICTURE_TYPE_B))
- ff_pred_weight_table(h, sl);
+ if ((p->ps.pps->weighted_pred && slice_type_nos == AV_PICTURE_TYPE_P) ||
+ (p->ps.pps->weighted_bipred_idc == 1 && slice_type_nos == AV_PICTURE_TYPE_B))
+ ff_h264_pred_weight_table(gb, p->ps.sps, ref_count, slice_type_nos,
+ &pwt);
- if (get_bits1(&sl->gb)) { // adaptive_ref_pic_marking_mode_flag
+ if (get_bits1(gb)) { // adaptive_ref_pic_marking_mode_flag
int i;
for (i = 0; i < MAX_MMCO_COUNT; i++) {
- MMCOOpcode opcode = get_ue_golomb_31(&sl->gb);
+ MMCOOpcode opcode = get_ue_golomb_31(gb);
if (opcode > (unsigned) MMCO_LONG) {
- av_log(h->avctx, AV_LOG_ERROR,
+ av_log(avctx, AV_LOG_ERROR,
"illegal memory management control operation %d\n",
opcode);
return AVERROR_INVALIDDATA;
return 1;
if (opcode == MMCO_SHORT2UNUSED || opcode == MMCO_SHORT2LONG)
- get_ue_golomb(&sl->gb);
+ get_ue_golomb(gb);
if (opcode == MMCO_SHORT2LONG || opcode == MMCO_LONG2UNUSED ||
opcode == MMCO_LONG || opcode == MMCO_SET_MAX_LONG)
- get_ue_golomb_31(&sl->gb);
+ get_ue_golomb_31(gb);
}
}
const uint8_t *buf, int buf_size)
{
H264ParseContext *p = s->priv_data;
- H264Context *h = &p->h;
- H264SliceContext *sl = &h->slice_ctx[0];
const uint8_t *buf_end = buf + buf_size;
+
+ H2645NAL nal = { NULL };
+
unsigned int pps_id;
unsigned int slice_type;
int state = -1, got_reset = 0;
- const uint8_t *ptr;
int field_poc[2];
+ int ret;
/* set some sane default values */
s->pict_type = AV_PICTURE_TYPE_I;
s->key_frame = 0;
s->picture_structure = AV_PICTURE_STRUCTURE_UNKNOWN;
- h->avctx = avctx;
- ff_h264_reset_sei(h);
+ ff_h264_sei_uninit(&p->sei);
if (!buf_size)
return 0;
for (;;) {
- int src_length, dst_length, consumed;
+ const SPS *sps;
+ int src_length, consumed;
buf = avpriv_find_start_code(buf, buf_end, &state);
if (buf >= buf_end)
break;
}
break;
}
- ptr = ff_h264_decode_nal(h, sl, buf, &dst_length, &consumed, src_length);
- if (!ptr || dst_length < 0)
+
+ consumed = ff_h2645_extract_rbsp(buf, src_length, &nal);
+ if (consumed < 0)
break;
- init_get_bits(&h->gb, ptr, 8 * dst_length);
- switch (h->nal_unit_type) {
+ ret = init_get_bits(&nal.gb, nal.data, nal.size * 8);
+ if (ret < 0)
+ goto fail;
+ get_bits1(&nal.gb);
+ nal.ref_idc = get_bits(&nal.gb, 2);
+ nal.type = get_bits(&nal.gb, 5);
+
+ switch (nal.type) {
case NAL_SPS:
- ff_h264_decode_seq_parameter_set(h);
+ ff_h264_decode_seq_parameter_set(&nal.gb, avctx, &p->ps);
break;
case NAL_PPS:
- ff_h264_decode_picture_parameter_set(h, h->gb.size_in_bits);
+ ff_h264_decode_picture_parameter_set(&nal.gb, avctx, &p->ps,
+ nal.size_bits);
break;
case NAL_SEI:
- ff_h264_decode_sei(h);
+ ff_h264_sei_decode(&p->sei, &nal.gb, &p->ps, avctx);
break;
case NAL_IDR_SLICE:
s->key_frame = 1;
- h->prev_frame_num = 0;
- h->prev_frame_num_offset = 0;
- h->prev_poc_msb =
- h->prev_poc_lsb = 0;
+ p->poc.prev_frame_num = 0;
+ p->poc.prev_frame_num_offset = 0;
+ p->poc.prev_poc_msb =
+ p->poc.prev_poc_lsb = 0;
/* fall through */
case NAL_SLICE:
- init_get_bits(&sl->gb, ptr, 8 * dst_length);
- get_ue_golomb(&sl->gb); // skip first_mb_in_slice
- slice_type = get_ue_golomb_31(&sl->gb);
+ get_ue_golomb(&nal.gb); // skip first_mb_in_slice
+ slice_type = get_ue_golomb_31(&nal.gb);
s->pict_type = ff_h264_golomb_to_pict_type[slice_type % 5];
- if (h->sei_recovery_frame_cnt >= 0) {
+ if (p->sei.recovery_point.recovery_frame_cnt >= 0) {
/* key frame, since recovery_frame_cnt is set */
s->key_frame = 1;
}
- pps_id = get_ue_golomb(&sl->gb);
+ pps_id = get_ue_golomb(&nal.gb);
if (pps_id >= MAX_PPS_COUNT) {
- av_log(h->avctx, AV_LOG_ERROR,
+ av_log(avctx, AV_LOG_ERROR,
"pps_id %u out of range\n", pps_id);
- return -1;
+ goto fail;
}
- if (!h->pps_buffers[pps_id]) {
- av_log(h->avctx, AV_LOG_ERROR,
+ if (!p->ps.pps_list[pps_id]) {
+ av_log(avctx, AV_LOG_ERROR,
"non-existing PPS %u referenced\n", pps_id);
- return -1;
+ goto fail;
}
- h->pps = *h->pps_buffers[pps_id];
- if (!h->sps_buffers[h->pps.sps_id]) {
- av_log(h->avctx, AV_LOG_ERROR,
- "non-existing SPS %u referenced\n", h->pps.sps_id);
- return -1;
+ p->ps.pps = (const PPS*)p->ps.pps_list[pps_id]->data;
+ if (!p->ps.sps_list[p->ps.pps->sps_id]) {
+ av_log(avctx, AV_LOG_ERROR,
+ "non-existing SPS %u referenced\n", p->ps.pps->sps_id);
+ goto fail;
}
- h->sps = *h->sps_buffers[h->pps.sps_id];
- h->frame_num = get_bits(&sl->gb, h->sps.log2_max_frame_num);
+ p->ps.sps = (SPS*)p->ps.sps_list[p->ps.pps->sps_id]->data;
+
+ sps = p->ps.sps;
+
+ p->poc.frame_num = get_bits(&nal.gb, sps->log2_max_frame_num);
- s->coded_width = 16 * h->sps.mb_width;
- s->coded_height = 16 * h->sps.mb_height;
- s->width = s->coded_width - (h->sps.crop_right + h->sps.crop_left);
- s->height = s->coded_height - (h->sps.crop_top + h->sps.crop_bottom);
+ s->coded_width = 16 * sps->mb_width;
+ s->coded_height = 16 * sps->mb_height;
+ s->width = s->coded_width - (sps->crop_right + sps->crop_left);
+ s->height = s->coded_height - (sps->crop_top + sps->crop_bottom);
if (s->width <= 0 || s->height <= 0) {
s->width = s->coded_width;
s->height = s->coded_height;
}
- switch (h->sps.bit_depth_luma) {
+ switch (sps->bit_depth_luma) {
case 9:
- if (CHROMA444(h)) s->format = AV_PIX_FMT_YUV444P9;
- else if (CHROMA422(h)) s->format = AV_PIX_FMT_YUV422P9;
- else s->format = AV_PIX_FMT_YUV420P9;
+ if (sps->chroma_format_idc == 3) s->format = AV_PIX_FMT_YUV444P9;
+ else if (sps->chroma_format_idc == 2) s->format = AV_PIX_FMT_YUV422P9;
+ else s->format = AV_PIX_FMT_YUV420P9;
break;
case 10:
- if (CHROMA444(h)) s->format = AV_PIX_FMT_YUV444P10;
- else if (CHROMA422(h)) s->format = AV_PIX_FMT_YUV422P10;
- else s->format = AV_PIX_FMT_YUV420P10;
+ if (sps->chroma_format_idc == 3) s->format = AV_PIX_FMT_YUV444P10;
+ else if (sps->chroma_format_idc == 2) s->format = AV_PIX_FMT_YUV422P10;
+ else s->format = AV_PIX_FMT_YUV420P10;
break;
case 8:
- if (CHROMA444(h)) s->format = AV_PIX_FMT_YUV444P;
- else if (CHROMA422(h)) s->format = AV_PIX_FMT_YUV422P;
- else s->format = AV_PIX_FMT_YUV420P;
+ if (sps->chroma_format_idc == 3) s->format = AV_PIX_FMT_YUV444P;
+ else if (sps->chroma_format_idc == 2) s->format = AV_PIX_FMT_YUV422P;
+ else s->format = AV_PIX_FMT_YUV420P;
break;
default:
s->format = AV_PIX_FMT_NONE;
}
- avctx->profile = ff_h264_get_profile(&h->sps);
- avctx->level = h->sps.level_idc;
+ avctx->profile = ff_h264_get_profile(sps);
+ avctx->level = sps->level_idc;
- if (h->sps.frame_mbs_only_flag) {
- h->picture_structure = PICT_FRAME;
+ if (sps->frame_mbs_only_flag) {
+ p->picture_structure = PICT_FRAME;
} else {
- if (get_bits1(&sl->gb)) { // field_pic_flag
- h->picture_structure = PICT_TOP_FIELD + get_bits1(&sl->gb); // bottom_field_flag
+ if (get_bits1(&nal.gb)) { // field_pic_flag
+ p->picture_structure = PICT_TOP_FIELD + get_bits1(&nal.gb); // bottom_field_flag
} else {
- h->picture_structure = PICT_FRAME;
+ p->picture_structure = PICT_FRAME;
}
}
- if (h->nal_unit_type == NAL_IDR_SLICE)
- get_ue_golomb(&sl->gb); /* idr_pic_id */
- if (h->sps.poc_type == 0) {
- h->poc_lsb = get_bits(&sl->gb, h->sps.log2_max_poc_lsb);
+ if (nal.type == NAL_IDR_SLICE)
+ get_ue_golomb(&nal.gb); /* idr_pic_id */
+ if (sps->poc_type == 0) {
+ p->poc.poc_lsb = get_bits(&nal.gb, sps->log2_max_poc_lsb);
- if (h->pps.pic_order_present == 1 &&
- h->picture_structure == PICT_FRAME)
- h->delta_poc_bottom = get_se_golomb(&sl->gb);
+ if (p->ps.pps->pic_order_present == 1 &&
+ p->picture_structure == PICT_FRAME)
+ p->poc.delta_poc_bottom = get_se_golomb(&nal.gb);
}
- if (h->sps.poc_type == 1 &&
- !h->sps.delta_pic_order_always_zero_flag) {
- h->delta_poc[0] = get_se_golomb(&sl->gb);
+ if (sps->poc_type == 1 &&
+ !sps->delta_pic_order_always_zero_flag) {
+ p->poc.delta_poc[0] = get_se_golomb(&nal.gb);
- if (h->pps.pic_order_present == 1 &&
- h->picture_structure == PICT_FRAME)
- h->delta_poc[1] = get_se_golomb(&sl->gb);
+ if (p->ps.pps->pic_order_present == 1 &&
+ p->picture_structure == PICT_FRAME)
+ p->poc.delta_poc[1] = get_se_golomb(&nal.gb);
}
/* Decode POC of this picture.
* The prev_ values needed for decoding POC of the next picture are not set here. */
field_poc[0] = field_poc[1] = INT_MAX;
- ff_init_poc(h, field_poc, &s->output_picture_number);
+ ff_h264_init_poc(field_poc, &s->output_picture_number, sps,
+ &p->poc, p->picture_structure, nal.ref_idc);
/* Continue parsing to check if MMCO_RESET is present.
* FIXME: MMCO_RESET could appear in non-first slice.
* Maybe, we should parse all undisposable non-IDR slice of this
* picture until encountering MMCO_RESET in a slice of it. */
- if (h->nal_ref_idc && h->nal_unit_type != NAL_IDR_SLICE) {
- got_reset = scan_mmco_reset(s);
+ if (nal.ref_idc && nal.type != NAL_IDR_SLICE) {
+ got_reset = scan_mmco_reset(s, &nal.gb, avctx);
if (got_reset < 0)
- return got_reset;
+ goto fail;
}
/* Set up the prev_ values for decoding POC of the next picture. */
- h->prev_frame_num = got_reset ? 0 : h->frame_num;
- h->prev_frame_num_offset = got_reset ? 0 : h->frame_num_offset;
- if (h->nal_ref_idc != 0) {
+ p->poc.prev_frame_num = got_reset ? 0 : p->poc.frame_num;
+ p->poc.prev_frame_num_offset = got_reset ? 0 : p->poc.frame_num_offset;
+ if (nal.ref_idc != 0) {
if (!got_reset) {
- h->prev_poc_msb = h->poc_msb;
- h->prev_poc_lsb = h->poc_lsb;
+ p->poc.prev_poc_msb = p->poc.poc_msb;
+ p->poc.prev_poc_lsb = p->poc.poc_lsb;
} else {
- h->prev_poc_msb = 0;
- h->prev_poc_lsb =
- h->picture_structure == PICT_BOTTOM_FIELD ? 0 : field_poc[0];
+ p->poc.prev_poc_msb = 0;
+ p->poc.prev_poc_lsb =
+ p->picture_structure == PICT_BOTTOM_FIELD ? 0 : field_poc[0];
}
}
- if (h->sps.pic_struct_present_flag) {
- switch (h->sei_pic_struct) {
+ if (sps->pic_struct_present_flag) {
+ switch (p->sei.picture_timing.pic_struct) {
case SEI_PIC_STRUCT_TOP_FIELD:
case SEI_PIC_STRUCT_BOTTOM_FIELD:
s->repeat_pict = 0;
s->repeat_pict = 5;
break;
default:
- s->repeat_pict = h->picture_structure == PICT_FRAME ? 1 : 0;
+ s->repeat_pict = p->picture_structure == PICT_FRAME ? 1 : 0;
break;
}
} else {
- s->repeat_pict = h->picture_structure == PICT_FRAME ? 1 : 0;
+ s->repeat_pict = p->picture_structure == PICT_FRAME ? 1 : 0;
}
- if (h->picture_structure == PICT_FRAME) {
+ if (p->picture_structure == PICT_FRAME) {
s->picture_structure = AV_PICTURE_STRUCTURE_FRAME;
- if (h->sps.pic_struct_present_flag) {
- switch (h->sei_pic_struct) {
+ if (sps->pic_struct_present_flag) {
+ switch (p->sei.picture_timing.pic_struct) {
case SEI_PIC_STRUCT_TOP_BOTTOM:
case SEI_PIC_STRUCT_TOP_BOTTOM_TOP:
s->field_order = AV_FIELD_TT;
s->field_order = AV_FIELD_PROGRESSIVE;
}
} else {
- if (h->picture_structure == PICT_TOP_FIELD)
+ if (p->picture_structure == PICT_TOP_FIELD)
s->picture_structure = AV_PICTURE_STRUCTURE_TOP_FIELD;
else
s->picture_structure = AV_PICTURE_STRUCTURE_BOTTOM_FIELD;
s->field_order = AV_FIELD_UNKNOWN;
}
+ av_freep(&nal.rbsp_buffer);
return 0; /* no need to evaluate the rest */
}
buf += consumed;
}
/* didn't find a picture! */
- av_log(h->avctx, AV_LOG_ERROR, "missing picture in access unit\n");
+ av_log(avctx, AV_LOG_ERROR, "missing picture in access unit\n");
+fail:
+ av_freep(&nal.rbsp_buffer);
return -1;
}
const uint8_t *buf, int buf_size)
{
H264ParseContext *p = s->priv_data;
- H264Context *h = &p->h;
ParseContext *pc = &p->pc;
int next;
if (!p->got_first) {
p->got_first = 1;
if (avctx->extradata_size) {
- h->avctx = avctx;
- // must be done like in the decoder.
- // otherwise opening the parser, creating extradata,
- // and then closing and opening again
- // will cause has_b_frames to be always set.
- // NB: estimate_timings_from_pts behaves exactly like this.
- if (!avctx->has_b_frames)
- h->low_delay = 1;
- ff_h264_decode_extradata(h);
+ ff_h264_decode_extradata(avctx->extradata, avctx->extradata_size,
+ &p->ps, &p->is_avc, &p->nal_length_size,
+ avctx->err_recognition, avctx);
}
}
parse_nal_units(s, avctx, buf, buf_size);
- if (h->sei_cpb_removal_delay >= 0) {
- s->dts_sync_point = h->sei_buffering_period_present;
- s->dts_ref_dts_delta = h->sei_cpb_removal_delay;
- s->pts_dts_delta = h->sei_dpb_output_delay;
+ if (p->sei.picture_timing.cpb_removal_delay >= 0) {
+ s->dts_sync_point = p->sei.buffering_period.present;
+ s->dts_ref_dts_delta = p->sei.picture_timing.cpb_removal_delay;
+ s->pts_dts_delta = p->sei.picture_timing.dpb_output_delay;
} else {
s->dts_sync_point = INT_MIN;
s->dts_ref_dts_delta = INT_MIN;
static void h264_close(AVCodecParserContext *s)
{
H264ParseContext *p = s->priv_data;
- H264Context *h = &p->h;
ParseContext *pc = &p->pc;
+ int i;
av_free(pc->buffer);
- ff_h264_free_context(h);
+
+ ff_h264_sei_uninit(&p->sei);
+
+ for (i = 0; i < FF_ARRAY_ELEMS(p->ps.sps_list); i++)
+ av_buffer_unref(&p->ps.sps_list[i]);
+
+ for (i = 0; i < FF_ARRAY_ELEMS(p->ps.pps_list); i++)
+ av_buffer_unref(&p->ps.pps_list[i]);
}
static av_cold int init(AVCodecParserContext *s)
{
H264ParseContext *p = s->priv_data;
- H264Context *h = &p->h;
-
- h->slice_ctx = av_mallocz(sizeof(*h->slice_ctx));
- if (!h->slice_ctx)
- return 0;
- h->nb_slice_ctx = 1;
- h->slice_context_count = 1;
- ff_h264dsp_init(&h->h264dsp, 8, 1);
+ ff_h264dsp_init(&p->h264dsp, 8, 1);
return 0;
}