And rename it to retimeinterleave, use the pcm_rechunk bitstream filter for rechunking. By seperating the two functions we hopefully get cleaner code. Signed-off-by: Marton Balint <cus@passwd.hu>tags/n4.3
@@ -2722,6 +2722,7 @@ fraps_decoder_select="bswapdsp huffman" | |||||
g2m_decoder_deps="zlib" | g2m_decoder_deps="zlib" | ||||
g2m_decoder_select="blockdsp idctdsp jpegtables" | g2m_decoder_select="blockdsp idctdsp jpegtables" | ||||
g729_decoder_select="audiodsp" | g729_decoder_select="audiodsp" | ||||
gxf_encoder_select="pcm_rechunk_bsf" | |||||
h261_decoder_select="mpegvideo" | h261_decoder_select="mpegvideo" | ||||
h261_encoder_select="mpegvideoenc" | h261_encoder_select="mpegvideoenc" | ||||
h263_decoder_select="h263_parser h263dsp mpegvideo qpeldsp" | h263_decoder_select="h263_parser h263dsp mpegvideo qpeldsp" | ||||
@@ -2794,6 +2795,7 @@ mv30_decoder_select="aandcttables blockdsp" | |||||
mvha_decoder_deps="zlib" | mvha_decoder_deps="zlib" | ||||
mvha_decoder_select="llviddsp" | mvha_decoder_select="llviddsp" | ||||
mwsc_decoder_deps="zlib" | mwsc_decoder_deps="zlib" | ||||
mxf_encoder_select="pcm_rechunk_bsf" | |||||
mxpeg_decoder_select="mjpeg_decoder" | mxpeg_decoder_select="mjpeg_decoder" | ||||
nellymoser_decoder_select="mdct sinewin" | nellymoser_decoder_select="mdct sinewin" | ||||
nellymoser_encoder_select="audio_frame_queue mdct sinewin" | nellymoser_encoder_select="audio_frame_queue mdct sinewin" | ||||
@@ -205,7 +205,7 @@ OBJS-$(CONFIG_GIF_DEMUXER) += gifdec.o | |||||
OBJS-$(CONFIG_GSM_DEMUXER) += gsmdec.o | OBJS-$(CONFIG_GSM_DEMUXER) += gsmdec.o | ||||
OBJS-$(CONFIG_GSM_MUXER) += rawenc.o | OBJS-$(CONFIG_GSM_MUXER) += rawenc.o | ||||
OBJS-$(CONFIG_GXF_DEMUXER) += gxf.o | OBJS-$(CONFIG_GXF_DEMUXER) += gxf.o | ||||
OBJS-$(CONFIG_GXF_MUXER) += gxfenc.o audiointerleave.o | |||||
OBJS-$(CONFIG_GXF_MUXER) += gxfenc.o retimeinterleave.o | |||||
OBJS-$(CONFIG_G722_DEMUXER) += g722.o rawdec.o | OBJS-$(CONFIG_G722_DEMUXER) += g722.o rawdec.o | ||||
OBJS-$(CONFIG_G722_MUXER) += rawenc.o | OBJS-$(CONFIG_G722_MUXER) += rawenc.o | ||||
OBJS-$(CONFIG_G723_1_DEMUXER) += g723_1.o | OBJS-$(CONFIG_G723_1_DEMUXER) += g723_1.o | ||||
@@ -347,7 +347,7 @@ OBJS-$(CONFIG_MUSX_DEMUXER) += musx.o | |||||
OBJS-$(CONFIG_MV_DEMUXER) += mvdec.o | OBJS-$(CONFIG_MV_DEMUXER) += mvdec.o | ||||
OBJS-$(CONFIG_MVI_DEMUXER) += mvi.o | OBJS-$(CONFIG_MVI_DEMUXER) += mvi.o | ||||
OBJS-$(CONFIG_MXF_DEMUXER) += mxfdec.o mxf.o | OBJS-$(CONFIG_MXF_DEMUXER) += mxfdec.o mxf.o | ||||
OBJS-$(CONFIG_MXF_MUXER) += mxfenc.o mxf.o audiointerleave.o avc.o | |||||
OBJS-$(CONFIG_MXF_MUXER) += mxfenc.o mxf.o retimeinterleave.o avc.o | |||||
OBJS-$(CONFIG_MXG_DEMUXER) += mxg.o | OBJS-$(CONFIG_MXG_DEMUXER) += mxg.o | ||||
OBJS-$(CONFIG_NC_DEMUXER) += ncdec.o | OBJS-$(CONFIG_NC_DEMUXER) += ncdec.o | ||||
OBJS-$(CONFIG_NISTSPHERE_DEMUXER) += nistspheredec.o pcm.o | OBJS-$(CONFIG_NISTSPHERE_DEMUXER) += nistspheredec.o pcm.o | ||||
@@ -1,148 +0,0 @@ | |||||
/* | |||||
* Audio Interleaving functions | |||||
* | |||||
* Copyright (c) 2009 Baptiste Coudurier <baptiste dot coudurier at gmail dot com> | |||||
* | |||||
* This file is part of FFmpeg. | |||||
* | |||||
* FFmpeg is free software; you can redistribute it and/or | |||||
* modify it under the terms of the GNU Lesser General Public | |||||
* License as published by the Free Software Foundation; either | |||||
* version 2.1 of the License, or (at your option) any later version. | |||||
* | |||||
* FFmpeg is distributed in the hope that it will be useful, | |||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of | |||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |||||
* Lesser General Public License for more details. | |||||
* | |||||
* You should have received a copy of the GNU Lesser General Public | |||||
* License along with FFmpeg; if not, write to the Free Software | |||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | |||||
*/ | |||||
#include "libavutil/fifo.h" | |||||
#include "libavutil/mathematics.h" | |||||
#include "avformat.h" | |||||
#include "audiointerleave.h" | |||||
#include "internal.h" | |||||
void ff_audio_interleave_close(AVFormatContext *s) | |||||
{ | |||||
int i; | |||||
for (i = 0; i < s->nb_streams; i++) { | |||||
AVStream *st = s->streams[i]; | |||||
AudioInterleaveContext *aic = st->priv_data; | |||||
if (aic && st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) | |||||
av_fifo_freep(&aic->fifo); | |||||
} | |||||
} | |||||
int ff_audio_interleave_init(AVFormatContext *s, | |||||
const int samples_per_frame, | |||||
AVRational time_base) | |||||
{ | |||||
int i; | |||||
if (!time_base.num) { | |||||
av_log(s, AV_LOG_ERROR, "timebase not set for audio interleave\n"); | |||||
return AVERROR(EINVAL); | |||||
} | |||||
for (i = 0; i < s->nb_streams; i++) { | |||||
AVStream *st = s->streams[i]; | |||||
AudioInterleaveContext *aic = st->priv_data; | |||||
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { | |||||
int max_samples = samples_per_frame ? samples_per_frame : | |||||
av_rescale_rnd(st->codecpar->sample_rate, time_base.num, time_base.den, AV_ROUND_UP); | |||||
aic->sample_size = (st->codecpar->channels * | |||||
av_get_bits_per_sample(st->codecpar->codec_id)) / 8; | |||||
if (!aic->sample_size) { | |||||
av_log(s, AV_LOG_ERROR, "could not compute sample size\n"); | |||||
return AVERROR(EINVAL); | |||||
} | |||||
aic->samples_per_frame = samples_per_frame; | |||||
aic->time_base = time_base; | |||||
if (!(aic->fifo = av_fifo_alloc_array(100, max_samples))) | |||||
return AVERROR(ENOMEM); | |||||
aic->fifo_size = 100 * max_samples; | |||||
} | |||||
} | |||||
return 0; | |||||
} | |||||
static int interleave_new_audio_packet(AVFormatContext *s, AVPacket *pkt, | |||||
int stream_index, int flush) | |||||
{ | |||||
AVStream *st = s->streams[stream_index]; | |||||
AudioInterleaveContext *aic = st->priv_data; | |||||
int ret; | |||||
int nb_samples = aic->samples_per_frame ? aic->samples_per_frame : | |||||
(av_rescale_q(aic->n + 1, av_make_q(st->codecpar->sample_rate, 1), av_inv_q(aic->time_base)) - aic->nb_samples); | |||||
int frame_size = nb_samples * aic->sample_size; | |||||
int size = FFMIN(av_fifo_size(aic->fifo), frame_size); | |||||
if (!size || (!flush && size == av_fifo_size(aic->fifo))) | |||||
return 0; | |||||
ret = av_new_packet(pkt, frame_size); | |||||
if (ret < 0) | |||||
return ret; | |||||
av_fifo_generic_read(aic->fifo, pkt->data, size, NULL); | |||||
if (size < pkt->size) | |||||
memset(pkt->data + size, 0, pkt->size - size); | |||||
pkt->dts = pkt->pts = aic->dts; | |||||
pkt->duration = av_rescale_q(nb_samples, st->time_base, aic->time_base); | |||||
pkt->stream_index = stream_index; | |||||
aic->dts += pkt->duration; | |||||
aic->nb_samples += nb_samples; | |||||
aic->n++; | |||||
return pkt->size; | |||||
} | |||||
int ff_audio_rechunk_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush, | |||||
int (*get_packet)(AVFormatContext *, AVPacket *, AVPacket *, int), | |||||
int (*compare_ts)(AVFormatContext *, const AVPacket *, const AVPacket *)) | |||||
{ | |||||
int i, ret; | |||||
if (pkt) { | |||||
AVStream *st = s->streams[pkt->stream_index]; | |||||
AudioInterleaveContext *aic = st->priv_data; | |||||
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { | |||||
unsigned new_size = av_fifo_size(aic->fifo) + pkt->size; | |||||
if (new_size > aic->fifo_size) { | |||||
if (av_fifo_realloc2(aic->fifo, new_size) < 0) | |||||
return AVERROR(ENOMEM); | |||||
aic->fifo_size = new_size; | |||||
} | |||||
av_fifo_generic_write(aic->fifo, pkt->data, pkt->size, NULL); | |||||
} else { | |||||
// rewrite pts and dts to be decoded time line position | |||||
pkt->pts = pkt->dts = aic->dts; | |||||
aic->dts += pkt->duration; | |||||
if ((ret = ff_interleave_add_packet(s, pkt, compare_ts)) < 0) | |||||
return ret; | |||||
} | |||||
pkt = NULL; | |||||
} | |||||
for (i = 0; i < s->nb_streams; i++) { | |||||
AVStream *st = s->streams[i]; | |||||
if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { | |||||
AVPacket new_pkt; | |||||
while ((ret = interleave_new_audio_packet(s, &new_pkt, i, flush)) > 0) { | |||||
if ((ret = ff_interleave_add_packet(s, &new_pkt, compare_ts)) < 0) | |||||
return ret; | |||||
} | |||||
if (ret < 0) | |||||
return ret; | |||||
} | |||||
} | |||||
return get_packet(s, out, NULL, flush); | |||||
} |
@@ -27,8 +27,9 @@ | |||||
#include "avformat.h" | #include "avformat.h" | ||||
#include "internal.h" | #include "internal.h" | ||||
#include "gxf.h" | #include "gxf.h" | ||||
#include "audiointerleave.h" | |||||
#include "retimeinterleave.h" | |||||
#define GXF_SAMPLES_PER_FRAME 32768 | |||||
#define GXF_AUDIO_PACKET_SIZE 65536 | #define GXF_AUDIO_PACKET_SIZE 65536 | ||||
#define GXF_TIMECODE(c, d, h, m, s, f) \ | #define GXF_TIMECODE(c, d, h, m, s, f) \ | ||||
@@ -44,7 +45,7 @@ typedef struct GXFTimecode{ | |||||
} GXFTimecode; | } GXFTimecode; | ||||
typedef struct GXFStreamContext { | typedef struct GXFStreamContext { | ||||
AudioInterleaveContext aic; | |||||
RetimeInterleaveContext aic; | |||||
uint32_t track_type; | uint32_t track_type; | ||||
uint32_t sample_size; | uint32_t sample_size; | ||||
uint32_t sample_rate; | uint32_t sample_rate; | ||||
@@ -663,8 +664,6 @@ static int gxf_write_umf_packet(AVFormatContext *s) | |||||
return updatePacketSize(pb, pos); | return updatePacketSize(pb, pos); | ||||
} | } | ||||
static const int GXF_samples_per_frame = 32768; | |||||
static void gxf_init_timecode_track(GXFStreamContext *sc, GXFStreamContext *vsc) | static void gxf_init_timecode_track(GXFStreamContext *sc, GXFStreamContext *vsc) | ||||
{ | { | ||||
if (!vsc) | if (!vsc) | ||||
@@ -736,6 +735,9 @@ static int gxf_write_header(AVFormatContext *s) | |||||
av_log(s, AV_LOG_ERROR, "only mono tracks are allowed\n"); | av_log(s, AV_LOG_ERROR, "only mono tracks are allowed\n"); | ||||
return -1; | return -1; | ||||
} | } | ||||
ret = ff_stream_add_bitstream_filter(st, "pcm_rechunk", "n="AV_STRINGIFY(GXF_SAMPLES_PER_FRAME)); | |||||
if (ret < 0) | |||||
return ret; | |||||
sc->track_type = 2; | sc->track_type = 2; | ||||
sc->sample_rate = st->codecpar->sample_rate; | sc->sample_rate = st->codecpar->sample_rate; | ||||
avpriv_set_pts_info(st, 64, 1, sc->sample_rate); | avpriv_set_pts_info(st, 64, 1, sc->sample_rate); | ||||
@@ -813,14 +815,12 @@ static int gxf_write_header(AVFormatContext *s) | |||||
return -1; | return -1; | ||||
} | } | ||||
} | } | ||||
ff_retime_interleave_init(&sc->aic, st->time_base); | |||||
/* FIXME first 10 audio tracks are 0 to 9 next 22 are A to V */ | /* FIXME first 10 audio tracks are 0 to 9 next 22 are A to V */ | ||||
sc->media_info = media_info<<8 | ('0'+tracks[media_info]++); | sc->media_info = media_info<<8 | ('0'+tracks[media_info]++); | ||||
sc->order = s->nb_streams - st->index; | sc->order = s->nb_streams - st->index; | ||||
} | } | ||||
if (ff_audio_interleave_init(s, GXF_samples_per_frame, (AVRational){ 1, 48000 }) < 0) | |||||
return -1; | |||||
if (tcr && vsc) | if (tcr && vsc) | ||||
gxf_init_timecode(s, &gxf->tc, tcr->value, vsc->fields); | gxf_init_timecode(s, &gxf->tc, tcr->value, vsc->fields); | ||||
@@ -877,8 +877,6 @@ static void gxf_deinit(AVFormatContext *s) | |||||
{ | { | ||||
GXFContext *gxf = s->priv_data; | GXFContext *gxf = s->priv_data; | ||||
ff_audio_interleave_close(s); | |||||
av_freep(&gxf->flt_entries); | av_freep(&gxf->flt_entries); | ||||
av_freep(&gxf->map_offsets); | av_freep(&gxf->map_offsets); | ||||
} | } | ||||
@@ -1016,8 +1014,8 @@ static int gxf_interleave_packet(AVFormatContext *s, AVPacket *out, AVPacket *pk | |||||
{ | { | ||||
if (pkt && s->streams[pkt->stream_index]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) | if (pkt && s->streams[pkt->stream_index]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO) | ||||
pkt->duration = 2; // enforce 2 fields | pkt->duration = 2; // enforce 2 fields | ||||
return ff_audio_rechunk_interleave(s, out, pkt, flush, | |||||
ff_interleave_packet_per_dts, gxf_compare_field_nb); | |||||
return ff_retime_interleave(s, out, pkt, flush, | |||||
ff_interleave_packet_per_dts, gxf_compare_field_nb); | |||||
} | } | ||||
AVOutputFormat ff_gxf_muxer = { | AVOutputFormat ff_gxf_muxer = { | ||||
@@ -52,7 +52,7 @@ | |||||
#include "libavcodec/h264_ps.h" | #include "libavcodec/h264_ps.h" | ||||
#include "libavcodec/golomb.h" | #include "libavcodec/golomb.h" | ||||
#include "libavcodec/internal.h" | #include "libavcodec/internal.h" | ||||
#include "audiointerleave.h" | |||||
#include "retimeinterleave.h" | |||||
#include "avformat.h" | #include "avformat.h" | ||||
#include "avio_internal.h" | #include "avio_internal.h" | ||||
#include "internal.h" | #include "internal.h" | ||||
@@ -79,7 +79,7 @@ typedef struct MXFIndexEntry { | |||||
} MXFIndexEntry; | } MXFIndexEntry; | ||||
typedef struct MXFStreamContext { | typedef struct MXFStreamContext { | ||||
AudioInterleaveContext aic; | |||||
RetimeInterleaveContext aic; | |||||
UID track_essence_element_key; | UID track_essence_element_key; | ||||
int index; ///< index in mxf_essence_container_uls table | int index; ///< index in mxf_essence_container_uls table | ||||
const UID *codec_ul; | const UID *codec_ul; | ||||
@@ -2538,6 +2538,7 @@ static int mxf_write_header(AVFormatContext *s) | |||||
if (mxf->signal_standard >= 0) | if (mxf->signal_standard >= 0) | ||||
sc->signal_standard = mxf->signal_standard; | sc->signal_standard = mxf->signal_standard; | ||||
} else if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { | } else if (st->codecpar->codec_type == AVMEDIA_TYPE_AUDIO) { | ||||
char bsf_arg[32]; | |||||
if (st->codecpar->sample_rate != 48000) { | if (st->codecpar->sample_rate != 48000) { | ||||
av_log(s, AV_LOG_ERROR, "only 48khz is implemented\n"); | av_log(s, AV_LOG_ERROR, "only 48khz is implemented\n"); | ||||
return -1; | return -1; | ||||
@@ -2580,6 +2581,10 @@ static int mxf_write_header(AVFormatContext *s) | |||||
av_rescale_rnd(st->codecpar->sample_rate, mxf->time_base.num, mxf->time_base.den, AV_ROUND_UP) * | av_rescale_rnd(st->codecpar->sample_rate, mxf->time_base.num, mxf->time_base.den, AV_ROUND_UP) * | ||||
av_get_bits_per_sample(st->codecpar->codec_id) / 8; | av_get_bits_per_sample(st->codecpar->codec_id) / 8; | ||||
} | } | ||||
snprintf(bsf_arg, sizeof(bsf_arg), "r=%d/%d", mxf->tc.rate.num, mxf->tc.rate.den); | |||||
ret = ff_stream_add_bitstream_filter(st, "pcm_rechunk", bsf_arg); | |||||
if (ret < 0) | |||||
return ret; | |||||
} else if (st->codecpar->codec_type == AVMEDIA_TYPE_DATA) { | } else if (st->codecpar->codec_type == AVMEDIA_TYPE_DATA) { | ||||
AVDictionaryEntry *e = av_dict_get(st->metadata, "data_type", NULL, 0); | AVDictionaryEntry *e = av_dict_get(st->metadata, "data_type", NULL, 0); | ||||
if (e && !strcmp(e->value, "vbi_vanc_smpte_436M")) { | if (e && !strcmp(e->value, "vbi_vanc_smpte_436M")) { | ||||
@@ -2593,6 +2598,7 @@ static int mxf_write_header(AVFormatContext *s) | |||||
return -1; | return -1; | ||||
} | } | ||||
} | } | ||||
ff_retime_interleave_init(&sc->aic, av_inv_q(mxf->tc.rate)); | |||||
if (sc->index == -1) { | if (sc->index == -1) { | ||||
sc->index = mxf_get_essence_container_ul_index(st->codecpar->codec_id); | sc->index = mxf_get_essence_container_ul_index(st->codecpar->codec_id); | ||||
@@ -2646,9 +2652,6 @@ static int mxf_write_header(AVFormatContext *s) | |||||
return AVERROR(ENOMEM); | return AVERROR(ENOMEM); | ||||
mxf->timecode_track->index = -1; | mxf->timecode_track->index = -1; | ||||
if (ff_audio_interleave_init(s, 0, av_inv_q(mxf->tc.rate)) < 0) | |||||
return -1; | |||||
return 0; | return 0; | ||||
} | } | ||||
@@ -3010,8 +3013,6 @@ static void mxf_deinit(AVFormatContext *s) | |||||
{ | { | ||||
MXFContext *mxf = s->priv_data; | MXFContext *mxf = s->priv_data; | ||||
ff_audio_interleave_close(s); | |||||
av_freep(&mxf->index_entries); | av_freep(&mxf->index_entries); | ||||
av_freep(&mxf->body_partition_offset); | av_freep(&mxf->body_partition_offset); | ||||
if (mxf->timecode_track) { | if (mxf->timecode_track) { | ||||
@@ -3086,8 +3087,8 @@ static int mxf_compare_timestamps(AVFormatContext *s, const AVPacket *next, | |||||
static int mxf_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush) | static int mxf_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush) | ||||
{ | { | ||||
return ff_audio_rechunk_interleave(s, out, pkt, flush, | |||||
mxf_interleave_get_packet, mxf_compare_timestamps); | |||||
return ff_retime_interleave(s, out, pkt, flush, | |||||
mxf_interleave_get_packet, mxf_compare_timestamps); | |||||
} | } | ||||
#define MXF_COMMON_OPTIONS \ | #define MXF_COMMON_OPTIONS \ | ||||
@@ -0,0 +1,51 @@ | |||||
/* | |||||
* Retime Interleaving functions | |||||
* | |||||
* Copyright (c) 2009 Baptiste Coudurier <baptiste dot coudurier at gmail dot com> | |||||
* | |||||
* This file is part of FFmpeg. | |||||
* | |||||
* FFmpeg is free software; you can redistribute it and/or | |||||
* modify it under the terms of the GNU Lesser General Public | |||||
* License as published by the Free Software Foundation; either | |||||
* version 2.1 of the License, or (at your option) any later version. | |||||
* | |||||
* FFmpeg is distributed in the hope that it will be useful, | |||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of | |||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | |||||
* Lesser General Public License for more details. | |||||
* | |||||
* You should have received a copy of the GNU Lesser General Public | |||||
* License along with FFmpeg; if not, write to the Free Software | |||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | |||||
*/ | |||||
#include "libavutil/mathematics.h" | |||||
#include "avformat.h" | |||||
#include "retimeinterleave.h" | |||||
#include "internal.h" | |||||
void ff_retime_interleave_init(RetimeInterleaveContext *aic, AVRational time_base) | |||||
{ | |||||
aic->time_base = time_base; | |||||
} | |||||
int ff_retime_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush, | |||||
int (*get_packet)(AVFormatContext *, AVPacket *, AVPacket *, int), | |||||
int (*compare_ts)(AVFormatContext *, const AVPacket *, const AVPacket *)) | |||||
{ | |||||
int ret; | |||||
if (pkt) { | |||||
AVStream *st = s->streams[pkt->stream_index]; | |||||
RetimeInterleaveContext *aic = st->priv_data; | |||||
pkt->duration = av_rescale_q(pkt->duration, st->time_base, aic->time_base); | |||||
// rewrite pts and dts to be decoded time line position | |||||
pkt->pts = pkt->dts = aic->dts; | |||||
aic->dts += pkt->duration; | |||||
if ((ret = ff_interleave_add_packet(s, pkt, compare_ts)) < 0) | |||||
return ret; | |||||
} | |||||
return get_packet(s, out, NULL, flush); | |||||
} |
@@ -20,36 +20,31 @@ | |||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA | ||||
*/ | */ | ||||
#ifndef AVFORMAT_AUDIOINTERLEAVE_H | |||||
#define AVFORMAT_AUDIOINTERLEAVE_H | |||||
#ifndef AVFORMAT_RETIMEINTERLEAVE_H | |||||
#define AVFORMAT_RETIMEINTERLEAVE_H | |||||
#include "libavutil/fifo.h" | |||||
#include "avformat.h" | #include "avformat.h" | ||||
typedef struct AudioInterleaveContext { | |||||
AVFifoBuffer *fifo; | |||||
unsigned fifo_size; ///< size of currently allocated FIFO | |||||
int64_t n; ///< number of generated packets | |||||
int64_t nb_samples; ///< number of generated samples | |||||
typedef struct RetimeInterleaveContext { | |||||
uint64_t dts; ///< current dts | uint64_t dts; ///< current dts | ||||
int sample_size; ///< size of one sample all channels included | |||||
int samples_per_frame; ///< samples per frame if fixed, 0 otherwise | |||||
AVRational time_base; ///< time base of output audio packets | |||||
} AudioInterleaveContext; | |||||
AVRational time_base; ///< time base of output packets | |||||
} RetimeInterleaveContext; | |||||
int ff_audio_interleave_init(AVFormatContext *s, const int samples_per_frame, AVRational time_base); | |||||
void ff_audio_interleave_close(AVFormatContext *s); | |||||
/** | |||||
* Init the retime interleave context | |||||
*/ | |||||
void ff_retime_interleave_init(RetimeInterleaveContext *aic, AVRational time_base); | |||||
/** | /** | ||||
* Rechunk audio PCM packets per AudioInterleaveContext->samples_per_frame | |||||
* and interleave them correctly. | |||||
* The first element of AVStream->priv_data must be AudioInterleaveContext | |||||
* Retime packets per RetimeInterleaveContext->time_base and interleave them | |||||
* correctly. | |||||
* The first element of AVStream->priv_data must be RetimeInterleaveContext | |||||
* when using this function. | * when using this function. | ||||
* | * | ||||
* @param get_packet function will output a packet when streams are correctly interleaved. | * @param get_packet function will output a packet when streams are correctly interleaved. | ||||
* @param compare_ts function will compare AVPackets and decide interleaving order. | * @param compare_ts function will compare AVPackets and decide interleaving order. | ||||
*/ | */ | ||||
int ff_audio_rechunk_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush, | |||||
int ff_retime_interleave(AVFormatContext *s, AVPacket *out, AVPacket *pkt, int flush, | |||||
int (*get_packet)(AVFormatContext *, AVPacket *, AVPacket *, int), | int (*get_packet)(AVFormatContext *, AVPacket *, AVPacket *, int), | ||||
int (*compare_ts)(AVFormatContext *, const AVPacket *, const AVPacket *)); | int (*compare_ts)(AVFormatContext *, const AVPacket *, const AVPacket *)); | ||||