Browse Source

qsvdec: move qsv_process_data() from qsvdec_h264 to the common code

It will be shared with the upcoming mpeg2 and hevc decoders.
tags/n2.8
Anton Khirnov 10 years ago
parent
commit
96dca089b1
3 changed files with 118 additions and 110 deletions
  1. +106
    -3
      libavcodec/qsvdec.c
  2. +9
    -3
      libavcodec/qsvdec.h
  3. +3
    -104
      libavcodec/qsvdec_h264.c

+ 106
- 3
libavcodec/qsvdec.c View File

@@ -34,6 +34,7 @@

#include "avcodec.h"
#include "internal.h"
#include "qsv.h"
#include "qsv_internal.h"
#include "qsvdec.h"

@@ -209,9 +210,9 @@ static QSVFrame *find_frame(QSVContext *q, mfxFrameSurface1 *surf)
return NULL;
}

int ff_qsv_decode(AVCodecContext *avctx, QSVContext *q,
AVFrame *frame, int *got_frame,
AVPacket *avpkt)
static int qsv_decode(AVCodecContext *avctx, QSVContext *q,
AVFrame *frame, int *got_frame,
AVPacket *avpkt)
{
QSVFrame *out_frame;
mfxFrameSurface1 *insurf;
@@ -317,8 +318,110 @@ int ff_qsv_decode_close(QSVContext *q)
av_fifo_free(q->async_fifo);
q->async_fifo = NULL;

av_parser_close(q->parser);
avcodec_free_context(&q->avctx_internal);

if (q->internal_session)
MFXClose(q->internal_session);

return 0;
}

int ff_qsv_process_data(AVCodecContext *avctx, QSVContext *q,
AVFrame *frame, int *got_frame, AVPacket *pkt)
{
uint8_t *dummy_data;
int dummy_size;
int ret;

if (!q->avctx_internal) {
q->avctx_internal = avcodec_alloc_context3(NULL);
if (!q->avctx_internal)
return AVERROR(ENOMEM);

if (avctx->extradata) {
q->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
if (!q->avctx_internal->extradata)
return AVERROR(ENOMEM);

memcpy(q->avctx_internal->extradata, avctx->extradata,
avctx->extradata_size);
q->avctx_internal->extradata_size = avctx->extradata_size;
}

q->parser = av_parser_init(avctx->codec_id);
if (!q->parser)
return AVERROR(ENOMEM);

q->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES;
q->orig_pix_fmt = AV_PIX_FMT_NONE;
}

if (!pkt->size)
return qsv_decode(avctx, q, frame, got_frame, pkt);

/* we assume the packets are already split properly and want
* just the codec parameters here */
av_parser_parse2(q->parser, q->avctx_internal,
&dummy_data, &dummy_size,
pkt->data, pkt->size, pkt->pts, pkt->dts,
pkt->pos);

/* TODO: flush delayed frames on reinit */
if (q->parser->format != q->orig_pix_fmt ||
q->parser->coded_width != avctx->coded_width ||
q->parser->coded_height != avctx->coded_height) {
mfxSession session = NULL;

enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV,
AV_PIX_FMT_NONE,
AV_PIX_FMT_NONE };
enum AVPixelFormat qsv_format;

qsv_format = ff_qsv_map_pixfmt(q->parser->format);
if (qsv_format < 0) {
av_log(avctx, AV_LOG_ERROR,
"Only 8-bit YUV420 streams are supported.\n");
ret = AVERROR(ENOSYS);
goto reinit_fail;
}

q->orig_pix_fmt = q->parser->format;
avctx->pix_fmt = pix_fmts[1] = qsv_format;
avctx->width = q->parser->width;
avctx->height = q->parser->height;
avctx->coded_width = q->parser->coded_width;
avctx->coded_height = q->parser->coded_height;
avctx->level = q->avctx_internal->level;
avctx->profile = q->avctx_internal->profile;

ret = ff_get_format(avctx, pix_fmts);
if (ret < 0)
goto reinit_fail;

avctx->pix_fmt = ret;

if (avctx->hwaccel_context) {
AVQSVContext *user_ctx = avctx->hwaccel_context;
session = user_ctx->session;
q->iopattern = user_ctx->iopattern;
q->ext_buffers = user_ctx->ext_buffers;
q->nb_ext_buffers = user_ctx->nb_ext_buffers;
}

ret = ff_qsv_decode_init(avctx, q, session);
if (ret < 0)
goto reinit_fail;
}

return qsv_decode(avctx, q, frame, got_frame, pkt);

reinit_fail:
q->orig_pix_fmt = q->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE;
return ret;
}

void ff_qsv_decode_flush(AVCodecContext *avctx, QSVContext *q)
{
q->orig_pix_fmt = AV_PIX_FMT_NONE;
}

+ 9
- 3
libavcodec/qsvdec.h View File

@@ -50,6 +50,11 @@ typedef struct QSVContext {

AVFifoBuffer *async_fifo;

// the internal parser and codec context for parsing the data
AVCodecParserContext *parser;
AVCodecContext *avctx_internal;
enum AVPixelFormat orig_pix_fmt;

// options set by the caller
int async_depth;
int iopattern;
@@ -62,9 +67,10 @@ int ff_qsv_map_pixfmt(enum AVPixelFormat format);

int ff_qsv_decode_init(AVCodecContext *s, QSVContext *q, mfxSession session);

int ff_qsv_decode(AVCodecContext *s, QSVContext *q,
AVFrame *frame, int *got_frame,
AVPacket *avpkt);
int ff_qsv_process_data(AVCodecContext *avctx, QSVContext *q,
AVFrame *frame, int *got_frame, AVPacket *pkt);

void ff_qsv_decode_flush(AVCodecContext *avctx, QSVContext *q);

int ff_qsv_decode_close(QSVContext *q);



+ 3
- 104
libavcodec/qsvdec_h264.c View File

@@ -41,11 +41,6 @@ typedef struct QSVH264Context {
AVClass *class;
QSVContext qsv;

// the internal parser and codec context for parsing the data
AVCodecParserContext *parser;
AVCodecContext *avctx_internal;
enum AVPixelFormat orig_pix_fmt;

// the filter for converting to Annex B
AVBitStreamFilterContext *bsf;

@@ -81,8 +76,6 @@ static av_cold int qsv_decode_close(AVCodecContext *avctx)
av_fifo_free(s->packet_fifo);

av_bitstream_filter_close(s->bsf);
av_parser_close(s->parser);
avcodec_free_context(&s->avctx_internal);

return 0;
}
@@ -92,7 +85,6 @@ static av_cold int qsv_decode_init(AVCodecContext *avctx)
QSVH264Context *s = avctx->priv_data;
int ret;

s->orig_pix_fmt = AV_PIX_FMT_NONE;

s->packet_fifo = av_fifo_alloc(sizeof(AVPacket));
if (!s->packet_fifo) {
@@ -106,30 +98,6 @@ static av_cold int qsv_decode_init(AVCodecContext *avctx)
goto fail;
}

s->avctx_internal = avcodec_alloc_context3(NULL);
if (!s->avctx_internal) {
ret = AVERROR(ENOMEM);
goto fail;
}

if (avctx->extradata) {
s->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE);
if (!s->avctx_internal->extradata) {
ret = AVERROR(ENOMEM);
goto fail;
}
memcpy(s->avctx_internal->extradata, avctx->extradata,
avctx->extradata_size);
s->avctx_internal->extradata_size = avctx->extradata_size;
}

s->parser = av_parser_init(AV_CODEC_ID_H264);
if (!s->parser) {
ret = AVERROR(ENOMEM);
goto fail;
}
s->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES;

s->qsv.iopattern = MFX_IOPATTERN_OUT_SYSTEM_MEMORY;

return 0;
@@ -138,75 +106,6 @@ fail:
return ret;
}

static int qsv_process_data(AVCodecContext *avctx, AVFrame *frame,
int *got_frame, AVPacket *pkt)
{
QSVH264Context *s = avctx->priv_data;
uint8_t *dummy_data;
int dummy_size;
int ret;

/* we assume the packets are already split properly and want
* just the codec parameters here */
av_parser_parse2(s->parser, s->avctx_internal,
&dummy_data, &dummy_size,
pkt->data, pkt->size, pkt->pts, pkt->dts,
pkt->pos);

/* TODO: flush delayed frames on reinit */
if (s->parser->format != s->orig_pix_fmt ||
s->parser->coded_width != avctx->coded_width ||
s->parser->coded_height != avctx->coded_height) {
mfxSession session = NULL;

enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV,
AV_PIX_FMT_NONE,
AV_PIX_FMT_NONE };
enum AVPixelFormat qsv_format;

qsv_format = ff_qsv_map_pixfmt(s->parser->format);
if (qsv_format < 0) {
av_log(avctx, AV_LOG_ERROR,
"Only 8-bit YUV420 streams are supported.\n");
ret = AVERROR(ENOSYS);
goto reinit_fail;
}

s->orig_pix_fmt = s->parser->format;
avctx->pix_fmt = pix_fmts[1] = qsv_format;
avctx->width = s->parser->width;
avctx->height = s->parser->height;
avctx->coded_width = s->parser->coded_width;
avctx->coded_height = s->parser->coded_height;
avctx->level = s->avctx_internal->level;
avctx->profile = s->avctx_internal->profile;

ret = ff_get_format(avctx, pix_fmts);
if (ret < 0)
goto reinit_fail;

avctx->pix_fmt = ret;

if (avctx->hwaccel_context) {
AVQSVContext *user_ctx = avctx->hwaccel_context;
session = user_ctx->session;
s->qsv.iopattern = user_ctx->iopattern;
s->qsv.ext_buffers = user_ctx->ext_buffers;
s->qsv.nb_ext_buffers = user_ctx->nb_ext_buffers;
}

ret = ff_qsv_decode_init(avctx, &s->qsv, session);
if (ret < 0)
goto reinit_fail;
}

return ff_qsv_decode(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered);

reinit_fail:
s->orig_pix_fmt = s->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE;
return ret;
}

static int qsv_decode_frame(AVCodecContext *avctx, void *data,
int *got_frame, AVPacket *avpkt)
{
@@ -239,7 +138,7 @@ static int qsv_decode_frame(AVCodecContext *avctx, void *data,

/* no more data */
if (av_fifo_size(s->packet_fifo) < sizeof(AVPacket))
return avpkt->size ? avpkt->size : ff_qsv_decode(avctx, &s->qsv, frame, got_frame, avpkt);
return avpkt->size ? avpkt->size : ff_qsv_process_data(avctx, &s->qsv, frame, got_frame, avpkt);

if (s->filtered_data != s->input_ref.data)
av_freep(&s->filtered_data);
@@ -259,7 +158,7 @@ static int qsv_decode_frame(AVCodecContext *avctx, void *data,
s->pkt_filtered.size = size;
}

ret = qsv_process_data(avctx, frame, got_frame, &s->pkt_filtered);
ret = ff_qsv_process_data(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered);
if (ret < 0)
return ret;

@@ -275,7 +174,7 @@ static void qsv_decode_flush(AVCodecContext *avctx)
QSVH264Context *s = avctx->priv_data;

qsv_clear_buffers(s);
s->orig_pix_fmt = AV_PIX_FMT_NONE;
ff_qsv_decode_flush(avctx, &s->qsv);
}

AVHWAccel ff_h264_qsv_hwaccel = {


Loading…
Cancel
Save