/* * Intel MediaSDK QSV based H.264 decoder * * copyright (c) 2013 Luca Barbato * copyright (c) 2015 Anton Khirnov * * This file is part of Libav. * * Libav is free software; you can redistribute it and/or * modify it under the terms of the GNU Lesser General Public * License as published by the Free Software Foundation; either * version 2.1 of the License, or (at your option) any later version. * * Libav is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU * Lesser General Public License for more details. * * You should have received a copy of the GNU Lesser General Public * License along with Libav; if not, write to the Free Software * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include #include "libavutil/common.h" #include "libavutil/fifo.h" #include "libavutil/opt.h" #include "avcodec.h" #include "internal.h" #include "qsv_internal.h" #include "qsvdec.h" #include "qsv.h" typedef struct QSVH264Context { AVClass *class; QSVContext qsv; // the internal parser and codec context for parsing the data AVCodecParserContext *parser; AVCodecContext *avctx_internal; enum AVPixelFormat orig_pix_fmt; // the filter for converting to Annex B AVBitStreamFilterContext *bsf; AVFifoBuffer *packet_fifo; AVPacket input_ref; AVPacket pkt_filtered; uint8_t *filtered_data; } QSVH264Context; static void qsv_clear_buffers(QSVH264Context *s) { AVPacket pkt; while (av_fifo_size(s->packet_fifo) >= sizeof(pkt)) { av_fifo_generic_read(s->packet_fifo, &pkt, sizeof(pkt), NULL); av_packet_unref(&pkt); } if (s->filtered_data != s->input_ref.data) av_freep(&s->filtered_data); s->filtered_data = NULL; av_packet_unref(&s->input_ref); } static av_cold int qsv_decode_close(AVCodecContext *avctx) { QSVH264Context *s = avctx->priv_data; ff_qsv_decode_close(&s->qsv); qsv_clear_buffers(s); av_fifo_free(s->packet_fifo); av_bitstream_filter_close(s->bsf); av_parser_close(s->parser); avcodec_free_context(&s->avctx_internal); return 0; } static av_cold int qsv_decode_init(AVCodecContext *avctx) { QSVH264Context *s = avctx->priv_data; int ret; s->orig_pix_fmt = AV_PIX_FMT_NONE; s->packet_fifo = av_fifo_alloc(sizeof(AVPacket)); if (!s->packet_fifo) { ret = AVERROR(ENOMEM); goto fail; } s->bsf = av_bitstream_filter_init("h264_mp4toannexb"); if (!s->bsf) { ret = AVERROR(ENOMEM); goto fail; } s->avctx_internal = avcodec_alloc_context3(NULL); if (!s->avctx_internal) { ret = AVERROR(ENOMEM); goto fail; } if (avctx->extradata) { s->avctx_internal->extradata = av_mallocz(avctx->extradata_size + FF_INPUT_BUFFER_PADDING_SIZE); if (!s->avctx_internal->extradata) { ret = AVERROR(ENOMEM); goto fail; } memcpy(s->avctx_internal->extradata, avctx->extradata, avctx->extradata_size); s->avctx_internal->extradata_size = avctx->extradata_size; } s->parser = av_parser_init(AV_CODEC_ID_H264); if (!s->parser) { ret = AVERROR(ENOMEM); goto fail; } s->parser->flags |= PARSER_FLAG_COMPLETE_FRAMES; s->qsv.iopattern = MFX_IOPATTERN_OUT_SYSTEM_MEMORY; return 0; fail: qsv_decode_close(avctx); return ret; } static int qsv_process_data(AVCodecContext *avctx, AVFrame *frame, int *got_frame, AVPacket *pkt) { QSVH264Context *s = avctx->priv_data; uint8_t *dummy_data; int dummy_size; int ret; /* we assume the packets are already split properly and want * just the codec parameters here */ av_parser_parse2(s->parser, s->avctx_internal, &dummy_data, &dummy_size, pkt->data, pkt->size, pkt->pts, pkt->dts, pkt->pos); /* TODO: flush delayed frames on reinit */ if (s->parser->format != s->orig_pix_fmt || s->parser->coded_width != avctx->coded_width || s->parser->coded_height != avctx->coded_height) { mfxSession session = NULL; enum AVPixelFormat pix_fmts[3] = { AV_PIX_FMT_QSV, AV_PIX_FMT_NONE, AV_PIX_FMT_NONE }; enum AVPixelFormat qsv_format; qsv_format = ff_qsv_map_pixfmt(s->parser->format); if (qsv_format < 0) { av_log(avctx, AV_LOG_ERROR, "Only 8-bit YUV420 streams are supported.\n"); ret = AVERROR(ENOSYS); goto reinit_fail; } s->orig_pix_fmt = s->parser->format; avctx->pix_fmt = pix_fmts[1] = qsv_format; avctx->width = s->parser->width; avctx->height = s->parser->height; avctx->coded_width = s->parser->coded_width; avctx->coded_height = s->parser->coded_height; avctx->level = s->avctx_internal->level; avctx->profile = s->avctx_internal->profile; ret = ff_get_format(avctx, pix_fmts); if (ret < 0) goto reinit_fail; avctx->pix_fmt = ret; if (avctx->hwaccel_context) { AVQSVContext *user_ctx = avctx->hwaccel_context; session = user_ctx->session; s->qsv.iopattern = user_ctx->iopattern; s->qsv.ext_buffers = user_ctx->ext_buffers; s->qsv.nb_ext_buffers = user_ctx->nb_ext_buffers; } ret = ff_qsv_decode_init(avctx, &s->qsv, session); if (ret < 0) goto reinit_fail; } return ff_qsv_decode(avctx, &s->qsv, frame, got_frame, &s->pkt_filtered); reinit_fail: s->orig_pix_fmt = s->parser->format = avctx->pix_fmt = AV_PIX_FMT_NONE; return ret; } static int qsv_decode_frame(AVCodecContext *avctx, void *data, int *got_frame, AVPacket *avpkt) { QSVH264Context *s = avctx->priv_data; AVFrame *frame = data; int ret; /* buffer the input packet */ if (avpkt->size) { AVPacket input_ref = { 0 }; if (av_fifo_space(s->packet_fifo) < sizeof(input_ref)) { ret = av_fifo_realloc2(s->packet_fifo, av_fifo_size(s->packet_fifo) + sizeof(input_ref)); if (ret < 0) return ret; } ret = av_packet_ref(&input_ref, avpkt); if (ret < 0) return ret; av_fifo_generic_write(s->packet_fifo, &input_ref, sizeof(input_ref), NULL); } /* process buffered data */ while (!*got_frame) { /* prepare the input data -- convert to Annex B if needed */ if (s->pkt_filtered.size <= 0) { int size; /* no more data */ if (av_fifo_size(s->packet_fifo) < sizeof(AVPacket)) return avpkt->size ? avpkt->size : ff_qsv_decode(avctx, &s->qsv, frame, got_frame, avpkt); if (s->filtered_data != s->input_ref.data) av_freep(&s->filtered_data); s->filtered_data = NULL; av_packet_unref(&s->input_ref); av_fifo_generic_read(s->packet_fifo, &s->input_ref, sizeof(s->input_ref), NULL); ret = av_bitstream_filter_filter(s->bsf, avctx, NULL, &s->filtered_data, &size, s->input_ref.data, s->input_ref.size, 0); if (ret < 0) { s->filtered_data = s->input_ref.data; size = s->input_ref.size; } s->pkt_filtered = s->input_ref; s->pkt_filtered.data = s->filtered_data; s->pkt_filtered.size = size; } ret = qsv_process_data(avctx, frame, got_frame, &s->pkt_filtered); if (ret < 0) return ret; s->pkt_filtered.size -= ret; s->pkt_filtered.data += ret; } return avpkt->size; } static void qsv_decode_flush(AVCodecContext *avctx) { QSVH264Context *s = avctx->priv_data; qsv_clear_buffers(s); s->orig_pix_fmt = AV_PIX_FMT_NONE; } AVHWAccel ff_h264_qsv_hwaccel = { .name = "h264_qsv", .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H264, .pix_fmt = AV_PIX_FMT_QSV, }; #define OFFSET(x) offsetof(QSVH264Context, x) #define VD AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_DECODING_PARAM static const AVOption options[] = { { "async_depth", "Internal parallelization depth, the higher the value the higher the latency.", OFFSET(qsv.async_depth), AV_OPT_TYPE_INT, { .i64 = ASYNC_DEPTH_DEFAULT }, 0, INT_MAX, VD }, { NULL }, }; static const AVClass class = { .class_name = "h264_qsv", .item_name = av_default_item_name, .option = options, .version = LIBAVUTIL_VERSION_INT, }; AVCodec ff_h264_qsv_decoder = { .name = "h264_qsv", .long_name = NULL_IF_CONFIG_SMALL("H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 (Intel Quick Sync Video acceleration)"), .priv_data_size = sizeof(QSVH264Context), .type = AVMEDIA_TYPE_VIDEO, .id = AV_CODEC_ID_H264, .init = qsv_decode_init, .decode = qsv_decode_frame, .flush = qsv_decode_flush, .close = qsv_decode_close, .capabilities = CODEC_CAP_DELAY, .priv_class = &class, };