Mercurial > libavformat.hg
changeset 4677:96c24e93b7d8 libavformat
correctly interleave audio and video according to specs
author | bcoudurier |
---|---|
date | Mon, 09 Mar 2009 01:10:28 +0000 |
parents | 0e45c20bbb9e |
children | 65036104b950 |
files | gxfenc.c |
diffstat | 1 files changed, 42 insertions(+), 9 deletions(-) [+] |
line wrap: on
line diff
--- a/gxfenc.c Mon Mar 09 00:14:55 2009 +0000 +++ b/gxfenc.c Mon Mar 09 01:10:28 2009 +0000 @@ -43,6 +43,7 @@ int p_per_gop; int b_per_i_or_p; ///< number of B frames per I frame or P frame int first_gop_closed; + unsigned order; ///< interleaving order } GXFStreamContext; typedef struct GXFContext { @@ -56,7 +57,7 @@ uint32_t umf_length; uint16_t umf_track_size; uint16_t umf_media_size; - int sample_rate; + AVRational time_base; int flags; GXFStreamContext timecode_track; } GXFContext; @@ -370,13 +371,14 @@ { GXFContext *gxf = s->priv_data; ByteIOContext *pb = s->pb; + int timecode_base = gxf->time_base.den == 60000 ? 60 : 50; // XXX drop frame uint32_t timecode = - gxf->nb_fields / (gxf->sample_rate * 3600) % 24 << 24 | // hours - gxf->nb_fields / (gxf->sample_rate * 60) % 60 << 16 | // minutes - gxf->nb_fields / gxf->sample_rate % 60 << 8 | // seconds - gxf->nb_fields % gxf->sample_rate; // fields + gxf->nb_fields / (timecode_base * 3600) % 24 << 24 | // hours + gxf->nb_fields / (timecode_base * 60) % 60 << 16 | // minutes + gxf->nb_fields / timecode_base % 60 << 8 | // seconds + gxf->nb_fields % timecode_base; // fields put_le32(pb, gxf->flags); put_le32(pb, gxf->nb_fields); /* length of the longest track */ @@ -647,19 +649,24 @@ gxf->flags |= 0x04000000; /* audio is 16 bit pcm */ media_info = 'A'; } else if (st->codec->codec_type == CODEC_TYPE_VIDEO) { + if (i != 0) { + av_log(s, AV_LOG_ERROR, "video stream must be the first track\n"); + return -1; + } /* FIXME check from time_base ? */ if (st->codec->height == 480 || st->codec->height == 512) { /* NTSC or NTSC+VBI */ sc->frame_rate_index = 5; sc->sample_rate = 60; gxf->flags |= 0x00000080; + gxf->time_base = (AVRational){ 1001, 60000 }; } else { /* assume PAL */ sc->frame_rate_index = 6; sc->media_type++; sc->sample_rate = 50; gxf->flags |= 0x00000040; + gxf->time_base = (AVRational){ 1, 50 }; } - gxf->sample_rate = sc->sample_rate; - av_set_pts_info(st, 64, 1, st->codec->time_base.den); + av_set_pts_info(st, 64, gxf->time_base.num, gxf->time_base.den); if (gxf_find_lines_index(st) < 0) sc->lines_index = -1; sc->sample_size = st->codec->bit_rate; @@ -704,6 +711,7 @@ } /* FIXME first 10 audio tracks are 0 to 9 next 22 are A to V */ sc->media_info = media_info<<8 | ('0'+tracks[media_info]++); + sc->order = s->nb_streams - st->index; } if (ff_audio_interleave_init(s, GXF_samples_per_frame, (AVRational){ 1, 48000 }) < 0) @@ -770,7 +778,8 @@ if (st->codec->codec_type == CODEC_TYPE_VIDEO) { field_nb = gxf->nb_fields; } else { - field_nb = av_rescale_rnd(pkt->dts, gxf->sample_rate, st->codec->time_base.den, AV_ROUND_UP); + field_nb = av_rescale_rnd(pkt->dts, gxf->time_base.den, + (int64_t)48000*gxf->time_base.num, AV_ROUND_UP); } put_byte(pb, sc->media_type); @@ -828,10 +837,34 @@ return updatePacketSize(pb, pos); } +static int gxf_compare_field_nb(AVFormatContext *s, AVPacket *next, AVPacket *cur) +{ + GXFContext *gxf = s->priv_data; + AVPacket *pkt[2] = { cur, next }; + int i, field_nb[2]; + GXFStreamContext *sc[2]; + + for (i = 0; i < 2; i++) { + AVStream *st = s->streams[pkt[i]->stream_index]; + sc[i] = st->priv_data; + if (st->codec->codec_type == CODEC_TYPE_AUDIO) { + field_nb[i] = av_rescale_rnd(pkt[i]->dts, gxf->time_base.den, + (int64_t)48000*gxf->time_base.num, AV_ROUND_UP); + field_nb[i] &= ~1; // compare against even field number because audio must be before video + } else + field_nb[i] = pkt[i]->dts; // dts are field based + } + + return field_nb[1] > field_nb[0] || + (field_nb[1] == field_nb[0] && sc[1]->order > sc[0]->order); +} + static int gxf_interleave_packet(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush) { + if (pkt && s->streams[pkt->stream_index]->codec->codec_type == CODEC_TYPE_VIDEO) + pkt->duration = 2; // enforce 2 fields return ff_audio_rechunk_interleave(s, out, pkt, flush, - av_interleave_packet_per_dts, ff_interleave_compare_dts); + av_interleave_packet_per_dts, gxf_compare_field_nb); } AVOutputFormat gxf_muxer = {