* commit 'e3db34291f4401a16f6ac92721617a9f33cd4c31': amrnb: decode directly to the user-provided AVFrame als: decode directly to the user-provided AVFrame alac: decode directly to the user-provided AVFrame adxenc: alloc/free coded_frame instead of keeping it in the ADXContext adx: decode directly to the user-provided AVFrame Conflicts: libavcodec/alsdec.c libavcodec/amrnbdec.c Merged-by: Michael Niedermayer <michaelni@gmx.at>tags/n1.2
| @@ -40,7 +40,6 @@ typedef struct ADXChannelState { | |||
| } ADXChannelState; | |||
| typedef struct ADXContext { | |||
| AVFrame frame; | |||
| int channels; | |||
| ADXChannelState prev[2]; | |||
| int header_parsed; | |||
| @@ -52,9 +52,6 @@ static av_cold int adx_decode_init(AVCodecContext *avctx) | |||
| avctx->sample_fmt = AV_SAMPLE_FMT_S16P; | |||
| avcodec_get_frame_defaults(&c->frame); | |||
| avctx->coded_frame = &c->frame; | |||
| return 0; | |||
| } | |||
| @@ -98,6 +95,7 @@ static int adx_decode(ADXContext *c, int16_t *out, int offset, | |||
| static int adx_decode_frame(AVCodecContext *avctx, void *data, | |||
| int *got_frame_ptr, AVPacket *avpkt) | |||
| { | |||
| AVFrame *frame = data; | |||
| int buf_size = avpkt->size; | |||
| ADXContext *c = avctx->priv_data; | |||
| int16_t **samples; | |||
| @@ -143,12 +141,12 @@ static int adx_decode_frame(AVCodecContext *avctx, void *data, | |||
| } | |||
| /* get output buffer */ | |||
| c->frame.nb_samples = num_blocks * BLOCK_SAMPLES; | |||
| if ((ret = ff_get_buffer(avctx, &c->frame)) < 0) { | |||
| frame->nb_samples = num_blocks * BLOCK_SAMPLES; | |||
| if ((ret = ff_get_buffer(avctx, frame)) < 0) { | |||
| av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); | |||
| return ret; | |||
| } | |||
| samples = (int16_t **)c->frame.extended_data; | |||
| samples = (int16_t **)frame->extended_data; | |||
| samples_offset = 0; | |||
| while (num_blocks--) { | |||
| @@ -164,8 +162,7 @@ static int adx_decode_frame(AVCodecContext *avctx, void *data, | |||
| samples_offset += BLOCK_SAMPLES; | |||
| } | |||
| *got_frame_ptr = 1; | |||
| *(AVFrame *)data = c->frame; | |||
| *got_frame_ptr = 1; | |||
| return buf - avpkt->data; | |||
| } | |||
| @@ -107,6 +107,14 @@ static int adx_encode_header(AVCodecContext *avctx, uint8_t *buf, int bufsize) | |||
| return HEADER_SIZE; | |||
| } | |||
| #if FF_API_OLD_ENCODE_AUDIO | |||
| static av_cold int adx_encode_close(AVCodecContext *avctx) | |||
| { | |||
| av_freep(&avctx->coded_frame); | |||
| return 0; | |||
| } | |||
| #endif | |||
| static av_cold int adx_encode_init(AVCodecContext *avctx) | |||
| { | |||
| ADXContext *c = avctx->priv_data; | |||
| @@ -118,8 +126,8 @@ static av_cold int adx_encode_init(AVCodecContext *avctx) | |||
| avctx->frame_size = BLOCK_SAMPLES; | |||
| #if FF_API_OLD_ENCODE_AUDIO | |||
| avcodec_get_frame_defaults(&c->frame); | |||
| avctx->coded_frame = &c->frame; | |||
| if (!(avctx->coded_frame = avcodec_alloc_frame())) | |||
| return AVERROR(ENOMEM); | |||
| #endif | |||
| /* the cutoff can be adjusted, but this seems to work pretty well */ | |||
| @@ -167,6 +175,9 @@ AVCodec ff_adpcm_adx_encoder = { | |||
| .id = AV_CODEC_ID_ADPCM_ADX, | |||
| .priv_data_size = sizeof(ADXContext), | |||
| .init = adx_encode_init, | |||
| #if FF_API_OLD_ENCODE_AUDIO | |||
| .close = adx_encode_close, | |||
| #endif | |||
| .encode2 = adx_encode_frame, | |||
| .sample_fmts = (const enum AVSampleFormat[]) { AV_SAMPLE_FMT_S16, | |||
| AV_SAMPLE_FMT_NONE }, | |||
| @@ -58,7 +58,6 @@ | |||
| typedef struct { | |||
| AVCodecContext *avctx; | |||
| AVFrame frame; | |||
| GetBitContext gb; | |||
| int channels; | |||
| @@ -254,7 +253,7 @@ static void append_extra_bits(int32_t *buffer[2], int32_t *extra_bits_buffer[2], | |||
| buffer[ch][i] = (buffer[ch][i] << extra_bits) | extra_bits_buffer[ch][i]; | |||
| } | |||
| static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| static int decode_element(AVCodecContext *avctx, AVFrame *frame, int ch_index, | |||
| int channels) | |||
| { | |||
| ALACContext *alac = avctx->priv_data; | |||
| @@ -289,8 +288,8 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| } | |||
| if (!alac->nb_samples) { | |||
| /* get output buffer */ | |||
| alac->frame.nb_samples = output_samples; | |||
| if ((ret = ff_get_buffer(avctx, &alac->frame)) < 0) { | |||
| frame->nb_samples = output_samples; | |||
| if ((ret = ff_get_buffer(avctx, frame)) < 0) { | |||
| av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); | |||
| return ret; | |||
| } | |||
| @@ -302,7 +301,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| alac->nb_samples = output_samples; | |||
| if (alac->direct_output) { | |||
| for (ch = 0; ch < channels; ch++) | |||
| alac->output_samples_buffer[ch] = (int32_t *)alac->frame.extended_data[ch_index + ch]; | |||
| alac->output_samples_buffer[ch] = (int32_t *)frame->extended_data[ch_index + ch]; | |||
| } | |||
| if (is_compressed) { | |||
| @@ -390,7 +389,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| switch(alac->sample_size) { | |||
| case 16: { | |||
| for (ch = 0; ch < channels; ch++) { | |||
| int16_t *outbuffer = (int16_t *)alac->frame.extended_data[ch_index + ch]; | |||
| int16_t *outbuffer = (int16_t *)frame->extended_data[ch_index + ch]; | |||
| for (i = 0; i < alac->nb_samples; i++) | |||
| *outbuffer++ = alac->output_samples_buffer[ch][i]; | |||
| }} | |||
| @@ -405,7 +404,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| }else{ | |||
| switch(alac->sample_size) { | |||
| case 16: { | |||
| int16_t *outbuffer = ((int16_t *)alac->frame.extended_data[0]) + ch_index; | |||
| int16_t *outbuffer = ((int16_t *)frame->extended_data[0]) + ch_index; | |||
| for (i = 0; i < alac->nb_samples; i++) { | |||
| for (ch = 0; ch < channels; ch++) | |||
| *outbuffer++ = alac->output_samples_buffer[ch][i]; | |||
| @@ -414,7 +413,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| } | |||
| break; | |||
| case 24: { | |||
| int32_t *outbuffer = ((int32_t *)alac->frame.extended_data[0]) + ch_index; | |||
| int32_t *outbuffer = ((int32_t *)frame->extended_data[0]) + ch_index; | |||
| for (i = 0; i < alac->nb_samples; i++) { | |||
| for (ch = 0; ch < channels; ch++) | |||
| *outbuffer++ = alac->output_samples_buffer[ch][i] << 8; | |||
| @@ -423,7 +422,7 @@ static int decode_element(AVCodecContext *avctx, void *data, int ch_index, | |||
| } | |||
| break; | |||
| case 32: { | |||
| int32_t *outbuffer = ((int32_t *)alac->frame.extended_data[0]) + ch_index; | |||
| int32_t *outbuffer = ((int32_t *)frame->extended_data[0]) + ch_index; | |||
| for (i = 0; i < alac->nb_samples; i++) { | |||
| for (ch = 0; ch < channels; ch++) | |||
| *outbuffer++ = alac->output_samples_buffer[ch][i]; | |||
| @@ -441,6 +440,7 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data, | |||
| int *got_frame_ptr, AVPacket *avpkt) | |||
| { | |||
| ALACContext *alac = avctx->priv_data; | |||
| AVFrame *frame = data; | |||
| enum AlacRawDataBlockType element; | |||
| int channels; | |||
| int ch, ret, got_end; | |||
| @@ -469,7 +469,7 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data, | |||
| return AVERROR_INVALIDDATA; | |||
| } | |||
| ret = decode_element(avctx, data, | |||
| ret = decode_element(avctx, frame, | |||
| ff_alac_channel_layout_offsets[alac->channels - 1][ch], | |||
| channels); | |||
| if (ret < 0 && get_bits_left(&alac->gb)) | |||
| @@ -487,8 +487,7 @@ static int alac_decode_frame(AVCodecContext *avctx, void *data, | |||
| avpkt->size * 8 - get_bits_count(&alac->gb)); | |||
| } | |||
| *got_frame_ptr = 1; | |||
| *(AVFrame *)data = alac->frame; | |||
| *got_frame_ptr = 1; | |||
| return avpkt->size; | |||
| } | |||
| @@ -616,9 +615,6 @@ static av_cold int alac_decode_init(AVCodecContext * avctx) | |||
| return ret; | |||
| } | |||
| avcodec_get_frame_defaults(&alac->frame); | |||
| avctx->coded_frame = &alac->frame; | |||
| return 0; | |||
| } | |||
| @@ -192,7 +192,6 @@ typedef struct { | |||
| typedef struct { | |||
| AVCodecContext *avctx; | |||
| AVFrame frame; | |||
| ALSSpecificConfig sconf; | |||
| GetBitContext gb; | |||
| DSPContext dsp; | |||
| @@ -1450,6 +1449,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| AVPacket *avpkt) | |||
| { | |||
| ALSDecContext *ctx = avctx->priv_data; | |||
| AVFrame *frame = data; | |||
| ALSSpecificConfig *sconf = &ctx->sconf; | |||
| const uint8_t *buffer = avpkt->data; | |||
| int buffer_size = avpkt->size; | |||
| @@ -1479,8 +1479,8 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| ctx->frame_id++; | |||
| /* get output buffer */ | |||
| ctx->frame.nb_samples = ctx->cur_frame_length; | |||
| if ((ret = ff_get_buffer(avctx, &ctx->frame)) < 0) { | |||
| frame->nb_samples = ctx->cur_frame_length; | |||
| if ((ret = ff_get_buffer(avctx, frame)) < 0) { | |||
| av_log(avctx, AV_LOG_ERROR, "get_buffer() failed.\n"); | |||
| return ret; | |||
| } | |||
| @@ -1488,7 +1488,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| // transform decoded frame into output format | |||
| #define INTERLEAVE_OUTPUT(bps) \ | |||
| { \ | |||
| int##bps##_t *dest = (int##bps##_t*)ctx->frame.data[0]; \ | |||
| int##bps##_t *dest = (int##bps##_t*)frame->data[0]; \ | |||
| shift = bps - ctx->avctx->bits_per_raw_sample; \ | |||
| if (!ctx->cs_switch) { \ | |||
| for (sample = 0; sample < ctx->cur_frame_length; sample++) \ | |||
| @@ -1512,7 +1512,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| int swap = HAVE_BIGENDIAN != sconf->msb_first; | |||
| if (ctx->avctx->bits_per_raw_sample == 24) { | |||
| int32_t *src = (int32_t *)ctx->frame.data[0]; | |||
| int32_t *src = (int32_t *)frame->data[0]; | |||
| for (sample = 0; | |||
| sample < ctx->cur_frame_length * avctx->channels; | |||
| @@ -1533,7 +1533,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| if (swap) { | |||
| if (ctx->avctx->bits_per_raw_sample <= 16) { | |||
| int16_t *src = (int16_t*) ctx->frame.data[0]; | |||
| int16_t *src = (int16_t*) frame->data[0]; | |||
| int16_t *dest = (int16_t*) ctx->crc_buffer; | |||
| for (sample = 0; | |||
| sample < ctx->cur_frame_length * avctx->channels; | |||
| @@ -1541,12 +1541,12 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| *dest++ = av_bswap16(src[sample]); | |||
| } else { | |||
| ctx->dsp.bswap_buf((uint32_t*)ctx->crc_buffer, | |||
| (uint32_t *)ctx->frame.data[0], | |||
| (uint32_t *)frame->data[0], | |||
| ctx->cur_frame_length * avctx->channels); | |||
| } | |||
| crc_source = ctx->crc_buffer; | |||
| } else { | |||
| crc_source = ctx->frame.data[0]; | |||
| crc_source = frame->data[0]; | |||
| } | |||
| ctx->crc = av_crc(ctx->crc_table, ctx->crc, crc_source, | |||
| @@ -1562,9 +1562,7 @@ static int decode_frame(AVCodecContext *avctx, void *data, int *got_frame_ptr, | |||
| } | |||
| } | |||
| *got_frame_ptr = 1; | |||
| *(AVFrame *)data = ctx->frame; | |||
| *got_frame_ptr = 1; | |||
| bytes_read = invalid_frame ? buffer_size : | |||
| (get_bits_count(&ctx->gb) + 7) >> 3; | |||
| @@ -1761,9 +1759,6 @@ static av_cold int decode_init(AVCodecContext *avctx) | |||
| ff_dsputil_init(&ctx->dsp, avctx); | |||
| avcodec_get_frame_defaults(&ctx->frame); | |||
| avctx->coded_frame = &ctx->frame; | |||
| return 0; | |||
| } | |||
| @@ -98,7 +98,6 @@ | |||
| #define AMR_AGC_ALPHA 0.9 | |||
| typedef struct AMRContext { | |||
| AVFrame avframe; ///< AVFrame for decoded samples | |||
| AMRNBFrame frame; ///< decoded AMR parameters (lsf coefficients, codebook indexes, etc) | |||
| uint8_t bad_frame_indicator; ///< bad frame ? 1 : 0 | |||
| enum Mode cur_frame_mode; | |||
| @@ -185,9 +184,6 @@ static av_cold int amrnb_decode_init(AVCodecContext *avctx) | |||
| for (i = 0; i < 4; i++) | |||
| p->prediction_error[i] = MIN_ENERGY; | |||
| avcodec_get_frame_defaults(&p->avframe); | |||
| avctx->coded_frame = &p->avframe; | |||
| ff_acelp_filter_init(&p->acelpf_ctx); | |||
| ff_acelp_vectors_init(&p->acelpv_ctx); | |||
| ff_celp_filter_init(&p->celpf_ctx); | |||
| @@ -954,6 +950,7 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data, | |||
| { | |||
| AMRContext *p = avctx->priv_data; // pointer to private data | |||
| AVFrame *frame = data; | |||
| const uint8_t *buf = avpkt->data; | |||
| int buf_size = avpkt->size; | |||
| float *buf_out; // pointer to the output data buffer | |||
| @@ -965,12 +962,12 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data, | |||
| const float *synth_fixed_vector; // pointer to the fixed vector that synthesis should use | |||
| /* get output buffer */ | |||
| p->avframe.nb_samples = AMR_BLOCK_SIZE; | |||
| if ((ret = ff_get_buffer(avctx, &p->avframe)) < 0) { | |||
| frame->nb_samples = AMR_BLOCK_SIZE; | |||
| if ((ret = ff_get_buffer(avctx, frame)) < 0) { | |||
| av_log(avctx, AV_LOG_ERROR, "get_buffer() failed\n"); | |||
| return ret; | |||
| } | |||
| buf_out = (float *)p->avframe.data[0]; | |||
| buf_out = (float *)frame->data[0]; | |||
| p->cur_frame_mode = unpack_bitstream(p, buf, buf_size); | |||
| if (p->cur_frame_mode == NO_DATA) { | |||
| @@ -1078,8 +1075,7 @@ static int amrnb_decode_frame(AVCodecContext *avctx, void *data, | |||
| p->acelpv_ctx.weighted_vector_sumf(p->lsf_avg, p->lsf_avg, p->lsf_q[3], | |||
| 0.84, 0.16, LP_FILTER_ORDER); | |||
| *got_frame_ptr = 1; | |||
| *(AVFrame *)data = p->avframe; | |||
| *got_frame_ptr = 1; | |||
| /* return the amount of bytes consumed if everything was OK */ | |||
| return frame_sizes_nb[p->cur_frame_mode] + 1; // +7 for rounding and +8 for TOC | |||