#define STREAM_DURATION 5.0
#define STREAM_FRAME_RATE 25 /* 25 images/s */
#define STREAM_NB_FRAMES ((int)(STREAM_DURATION * STREAM_FRAME_RATE))
+#define STREAM_PIX_FMT PIX_FMT_YUV420P /* default pix_fmt */
/**************************************************************/
/* audio output */
exit(1);
}
- c = &st->codec;
+ c = st->codec;
c->codec_id = codec_id;
c->codec_type = CODEC_TYPE_AUDIO;
AVCodecContext *c;
AVCodec *codec;
- c = &st->codec;
+ c = st->codec;
/* find the audio encoder */
codec = avcodec_find_encoder(c->codec_id);
support to compute the input frame size in samples */
if (c->frame_size <= 1) {
audio_input_frame_size = audio_outbuf_size / c->channels;
- switch(st->codec.codec_id) {
+ switch(st->codec->codec_id) {
case CODEC_ID_PCM_S16LE:
case CODEC_ID_PCM_S16BE:
case CODEC_ID_PCM_U16LE:
void write_audio_frame(AVFormatContext *oc, AVStream *st)
{
- int out_size;
AVCodecContext *c;
-
-
- c = &st->codec;
+ AVPacket pkt;
+ av_init_packet(&pkt);
+
+ c = st->codec;
get_audio_frame(samples, audio_input_frame_size, c->channels);
- out_size = avcodec_encode_audio(c, audio_outbuf, audio_outbuf_size, samples);
+ pkt.size= avcodec_encode_audio(c, audio_outbuf, audio_outbuf_size, samples);
+
+ pkt.pts= av_rescale_q(c->coded_frame->pts, c->time_base, st->time_base);
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= st->index;
+ pkt.data= audio_outbuf;
/* write the compressed frame in the media file */
- if (av_write_frame(oc, st->index, audio_outbuf, out_size) != 0) {
+ if (av_write_frame(oc, &pkt) != 0) {
fprintf(stderr, "Error while writing audio frame\n");
exit(1);
}
void close_audio(AVFormatContext *oc, AVStream *st)
{
- avcodec_close(&st->codec);
+ avcodec_close(st->codec);
av_free(samples);
av_free(audio_outbuf);
exit(1);
}
- c = &st->codec;
+ c = st->codec;
c->codec_id = codec_id;
c->codec_type = CODEC_TYPE_VIDEO;
/* resolution must be a multiple of two */
c->width = 352;
c->height = 288;
- /* frames per second */
- c->frame_rate = STREAM_FRAME_RATE;
- c->frame_rate_base = 1;
+ /* time base: this is the fundamental unit of time (in seconds) in terms
+ of which frame timestamps are represented. for fixed-fps content,
+ timebase should be 1/framerate and timestamp increments should be
+ identically 1. */
+ c->time_base.den = STREAM_FRAME_RATE;
+ c->time_base.num = 1;
c->gop_size = 12; /* emit one intra frame every twelve frames at most */
+ c->pix_fmt = STREAM_PIX_FMT;
if (c->codec_id == CODEC_ID_MPEG2VIDEO) {
/* just for testing, we also add B frames */
c->max_b_frames = 2;
AVCodec *codec;
AVCodecContext *c;
- c = &st->codec;
+ c = st->codec;
/* find the video encoder */
codec = avcodec_find_encoder(c->codec_id);
{
int out_size, ret;
AVCodecContext *c;
- AVFrame *picture_ptr;
- c = &st->codec;
+ c = st->codec;
if (frame_count >= STREAM_NB_FRAMES) {
/* no more frame to compress. The codec has a latency of a few
frames if using B frames, so we get the last frames by
- passing a NULL picture */
- picture_ptr = NULL;
+ passing the same picture again */
} else {
if (c->pix_fmt != PIX_FMT_YUV420P) {
/* as we only generate a YUV420P picture, we must convert it
} else {
fill_yuv_image(picture, frame_count, c->width, c->height);
}
- picture_ptr = picture;
}
if (oc->oformat->flags & AVFMT_RAWPICTURE) {
/* raw video case. The API will change slightly in the near
futur for that */
- ret = av_write_frame(oc, st->index,
- (uint8_t *)picture_ptr, sizeof(AVPicture));
+ AVPacket pkt;
+ av_init_packet(&pkt);
+
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= st->index;
+ pkt.data= (uint8_t *)picture;
+ pkt.size= sizeof(AVPicture);
+
+ ret = av_write_frame(oc, &pkt);
} else {
/* encode the image */
- out_size = avcodec_encode_video(c, video_outbuf, video_outbuf_size, picture_ptr);
+ out_size = avcodec_encode_video(c, video_outbuf, video_outbuf_size, picture);
/* if zero size, it means the image was buffered */
- if (out_size != 0) {
+ if (out_size > 0) {
+ AVPacket pkt;
+ av_init_packet(&pkt);
+
+ pkt.pts= av_rescale_q(c->coded_frame->pts, c->time_base, st->time_base);
+ if(c->coded_frame->key_frame)
+ pkt.flags |= PKT_FLAG_KEY;
+ pkt.stream_index= st->index;
+ pkt.data= video_outbuf;
+ pkt.size= out_size;
+
/* write the compressed frame in the media file */
- /* XXX: in case of B frames, the pts is not yet valid */
- ret = av_write_frame(oc, st->index, video_outbuf, out_size);
+ ret = av_write_frame(oc, &pkt);
} else {
ret = 0;
}
void close_video(AVFormatContext *oc, AVStream *st)
{
- avcodec_close(&st->codec);
+ avcodec_close(st->codec);
av_free(picture->data[0]);
av_free(picture);
if (tmp_picture) {
for(;;) {
/* compute current audio and video time */
if (audio_st)
- audio_pts = (double)audio_st->pts.val * oc->pts_num / oc->pts_den;
+ audio_pts = (double)audio_st->pts.val * audio_st->time_base.num / audio_st->time_base.den;
else
audio_pts = 0.0;
if (video_st)
- video_pts = (double)video_st->pts.val * oc->pts_num / oc->pts_den;
+ video_pts = (double)video_st->pts.val * video_st->time_base.num / video_st->time_base.den;
else
video_pts = 0.0;