From 4e08c821106fc1d6d358864abf0d8488b12f38c6 Mon Sep 17 00:00:00 2001 From: Anton Khirnov Date: Tue, 10 Feb 2015 10:40:59 +0100 Subject: lavc: add an Intel libmfx-based H.264 decoder Based on the code by Luca Barbato and Yukinori Yamazoe . --- libavcodec/qsv_h264.c | 313 ++++++++++++++++++++++++++++++++++++++++++++++++++ 1 file changed, 313 insertions(+) create mode 100644 libavcodec/qsv_h264.c (limited to 'libavcodec/qsv_h264.c') diff --git a/libavcodec/qsv_h264.c b/libavcodec/qsv_h264.c new file mode 100644 index 0000000000..a482e9c326 --- /dev/null +++ b/libavcodec/qsv_h264.c @@ -0,0 +1,313 @@ +/* + * Intel MediaSDK QSV based H.264 decoder + * + * copyright (c) 2013 Luca Barbato + * copyright (c) 2015 Anton Khirnov + * + * This file is part of Libav. + * + * Libav is free software; you can redistribute it and/or + * modify it under the terms of the GNU Lesser General Public + * License as published by the Free Software Foundation; either + * version 2.1 of the License, or (at your option) any later version. + * + * Libav is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU + * Lesser General Public License for more details. + * + * You should have received a copy of the GNU Lesser General Public + * License along with Libav; if not, write to the Free Software + * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA + */ + + +#include +#include + +#include + +#include "libavutil/common.h" +#include "libavutil/fifo.h" +#include "libavutil/opt.h" + +#include "avcodec.h" +#include "internal.h" +#include "qsv_internal.h" +#include "qsv.h" + +typedef struct QSVH264Context { + AVClass *class; + QSVContext qsv; + + // the internal parser and codec context for parsing the data + AVCodecParserContext *parser; + AVCodecContext *avctx_internal; + enum AVPixelFormat orig_pix_fmt; + + // the filter for converting to Annex B + AVBitStreamFilterContext *bsf; + + AVFifoBuffer *packet_fifo; + + AVPacket input_ref; + AVPacket pkt_filtered; + uint8_t *filtered_data; +} QSVH264Context; + +static void qsv_clear_buffers(QSVH264Context *s) +{ + AVPacket pkt; + while (av_fifo_size(s->packet_fifo) >= sizeof(pkt)) { + av_fifo_generic_read(s->packet_fifo, &pkt, sizeof(pkt), NULL); + av_packet_unref(&pkt); + } + + if (s->filtered_data != s->input_ref.data) + av_freep(&s->filtered_data); + s->filtered_data = NULL; + av_packet_unref(&s->input_ref); +} + +static av_cold int qsv_decode_close(AVCodecContext *avctx) +{ + QSVH264Context *s = avctx->priv_data; + + ff_qsv_close(&s->qsv); + + qsv_clear_buffers(s); + + av_fifo_free(s->packet_fifo); + + av_bitstream_filter_close(s->bsf); + av_parser_close(s->parser); + avcodec_free_context(&s->avctx_internal); + + return 0; +} + +static av_cold int qsv_decode_init(AVCodecContext *avctx) +{ + QSVH264Context *s = avctx->priv_data; + int ret; + + s->orig_pix_fmt = AV_PIX_FMT_NONE; + + s->packet_fifo = av_fifo_alloc(sizeof(AVPacket)); + if (!s->packet_fifo) { + ret = AVERROR(ENOMEM); + goto fail; + } + + s->bsf = av_bitstream_filter_init("h264_mp4toannexb"); + if (!s->bsf) { + ret = AVERROR(ENOMEM); + goto fail; + } + + s->avctx_internal = avcodec_alloc_context3(NULL); + if (!s->avctx_internal) { + ret = AVERROR(ENOMEM); + goto fail; + } + + if (avctx->extradata) { + s->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE); + if (!s->avctx_internal->extradata) { + ret = AVERROR(ENOMEM); + goto fail; + } + memcpy(s->avctx_internal->extradata, avctx->extradata, + avctx->extradata_size); + s->avctx_internal->extradata_size = avctx->extradata_size; + } + + s->parser = av_parser_init(AV_CODEC_ID_H264); + if (!s->parser) { + ret = AVERROR(ENOMEM); + goto fail; + } + s->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES; + + s->qsv.iopattern = MFX_IOPATTERN_OUT_SYSTEM_MEMORY; + + return 0; +fail: + qsv_decode_close(avctx); + return ret; +} + +static int qsv_process_data(AVCodecContext *avctx, AVFrame *frame, + int *got_frame, AVPacket *pkt) +{ + QSVH264Context *s = avctx->priv_data; + uint8_t *dummy_data; + int dummy_size; + int ret; + + /* we assume the packets are already split properly and want + * just the codec parameters here */ + av_parser_parse2(s->parser, s->avctx_internal, + &dummy_data, &dummy_size, + pkt->data, pkt->size, pkt->pts, pkt->dts, + pkt->pos); + + /* TODO: flush delayed frames on reinit */ + if (s->parser->format != s->orig_pix_fmt || + s->parser->coded_width != avctx->coded_width || + s->parser->coded_height != avctx->coded_height) { + mfxSession session = NULL; + + enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV, + AV_PIX_FMT_NONE, + AV_PIX_FMT_NONE }; + enum AVPixelFormat qsv_format; + + qsv_format = ff_qsv_map_pixfmt(s->parser->format); + if (qsv_format < 0) { + av_log(avctx, AV_LOG_ERROR, + "Only 8-bit YUV420 streams are supported.\n"); + ret = AVERROR(ENOSYS); + goto reinit_fail; + } + + s->orig_pix_fmt = s->parser->format; + avctx->pix_fmt = pix_fmts[1] = qsv_format; + avctx->width = s->parser->width; + avctx->height = s->parser->height; + avctx->coded_width = s->parser->coded_width; + avctx->coded_height = s->parser->coded_height; + avctx->level = s->avctx_internal->level; + avctx->profile = s->avctx_internal->profile; + + ret = ff_get_format(avctx, pix_fmts); + if (ret < 0) + goto reinit_fail; + + avctx->pix_fmt = ret; + + if (avctx->hwaccel_context) { + AVQSVContext *user_ctx = avctx->hwaccel_context; + session = user_ctx->session; + s->qsv.iopattern = user_ctx->iopattern; + s->qsv.ext_buffers = user_ctx->ext_buffers; + s->qsv.nb_ext_buffers = user_ctx->nb_ext_buffers; + } + + ret = ff_qsv_init(avctx, &s->qsv, session); + if (ret < 0) + goto reinit_fail; + } + + return ff_qsv_decode(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered); + +reinit_fail: + s->orig_pix_fmt = s->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE; + return ret; +} + +static int qsv_decode_frame(AVCodecContext *avctx, void *data, + int *got_frame, AVPacket *avpkt) +{ + QSVH264Context *s = avctx->priv_data; + AVFrame *frame = data; + int ret; + + /* buffer the input packet */ + if (avpkt->size) { + AVPacket input_ref = { 0 }; + + if (av_fifo_space(s->packet_fifo) < sizeof(input_ref)) { + ret = av_fifo_realloc2(s->packet_fifo, + av_fifo_size(s->packet_fifo) + sizeof(input_ref)); + if (ret < 0) + return ret; + } + + ret = av_packet_ref(&input_ref, avpkt); + if (ret < 0) + return ret; + av_fifo_generic_write(s->packet_fifo, &input_ref, sizeof(input_ref), NULL); + } + + /* process buffered data */ + while (!*got_frame) { + /* prepare the input data -- convert to Annex B if needed */ + if (s->pkt_filtered.size <= 0) { + int size; + + /* no more data */ + if (av_fifo_size(s->packet_fifo) < sizeof(AVPacket)) + return avpkt->size ? avpkt->size : ff_qsv_decode(avctx, &s->qsv, frame, got_frame, avpkt); + + if (s->filtered_data != s->input_ref.data) + av_freep(&s->filtered_data); + s->filtered_data = NULL; + av_packet_unref(&s->input_ref); + + av_fifo_generic_read(s->packet_fifo, &s->input_ref, sizeof(s->input_ref), NULL); + ret = av_bitstream_filter_filter(s->bsf, avctx, NULL, + &s->filtered_data, &size, + s->input_ref.data, s->input_ref.size, 0); + if (ret < 0) { + s->filtered_data = s->input_ref.data; + size = s->input_ref.size; + } + s->pkt_filtered = s->input_ref; + s->pkt_filtered.data = s->filtered_data; + s->pkt_filtered.size = size; + } + + ret = qsv_process_data(avctx, frame, got_frame, &s->pkt_filtered); + if (ret < 0) + return ret; + + s->pkt_filtered.size -= ret; + s->pkt_filtered.data += ret; + } + + return avpkt->size; +} + +static void qsv_decode_flush(AVCodecContext *avctx) +{ + QSVH264Context *s = avctx->priv_data; + + qsv_clear_buffers(s); + s->orig_pix_fmt = AV_PIX_FMT_NONE; +} + +AVHWAccel ff_h264_qsv_hwaccel = { + .name = "h264_qsv", + .type = AVMEDIA_TYPE_VIDEO, + .id = AV_CODEC_ID_H264, + .pix_fmt = AV_PIX_FMT_QSV, +}; + +#define OFFSET(x) offsetof(QSVH264Context, x) +#define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM +static const AVOption options[] = { + { "async_depth", "Internal parallelization depth, the higher the value the higher the latency.", OFFSET(qsv.async_depth), AV_OPT_TYPE_INT, { .i64 = ASYNC_DEPTH_DEFAULT }, 0, INT_MAX, VD }, + { NULL }, +}; + +static const AVClass class = { + .class_name = "h264_qsv", + .item_name = av_default_item_name, + .option = options, + .version = LIBAVUTIL_VERSION_INT, +}; + +AVCodec ff_h264_qsv_decoder = { + .name = "h264_qsv", + .long_name = NULL_IF_CONFIG_SMALL("H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 (Intel Quick Sync Video acceleration)"), + .priv_data_size = sizeof(QSVH264Context), + .type = AVMEDIA_TYPE_VIDEO, + .id = AV_CODEC_ID_H264, + .init = qsv_decode_init, + .decode = qsv_decode_frame, + .flush = qsv_decode_flush, + .close = qsv_decode_close, + .capabilities = CODEC_CAP_DELAY, + .priv_class = &class, +}; -- cgit v1.2.3