summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMark Thompson <sw@jkqxz.net>2016-03-23 23:49:02 +0000
committerAnton Khirnov <anton@khirnov.net>2016-03-30 09:05:28 +0200
commit104c804bcaac24b52eb51ed17df2fb311e6ae73e (patch)
tree75a1e1ad753b37d95e7fef5abbce785ab7f2d603
parent5d273d3efac340ef8de445c955ff44c7abed4e8f (diff)
lavc: VAAPI encode common infrastructure
Signed-off-by: Anton Khirnov <anton@khirnov.net>
-rwxr-xr-xconfigure2
-rw-r--r--libavcodec/Makefile1
-rw-r--r--libavcodec/vaapi_encode.c1096
-rw-r--r--libavcodec/vaapi_encode.h217
4 files changed, 1316 insertions, 0 deletions
diff --git a/configure b/configure
index f91a5d410b..f71dbe5bd1 100755
--- a/configure
+++ b/configure
@@ -1708,6 +1708,7 @@ CONFIG_EXTRA="
texturedsp
texturedspenc
tpeldsp
+ vaapi_encode
vc1dsp
videodsp
vp3dsp
@@ -1877,6 +1878,7 @@ nvenc_deps_any="dlopen LoadLibrary"
nvenc_extralibs='$ldl'
qsvdec_select="qsv"
qsvenc_select="qsv"
+vaapi_encode_deps="vaapi"
vc1dsp_select="h264chroma qpeldsp startcode"
# decoders / encoders
diff --git a/libavcodec/Makefile b/libavcodec/Makefile
index 0026b1fb7c..292066e52e 100644
--- a/libavcodec/Makefile
+++ b/libavcodec/Makefile
@@ -107,6 +107,7 @@ OBJS-$(CONFIG_STARTCODE) += startcode.o
OBJS-$(CONFIG_TEXTUREDSP) += texturedsp.o
OBJS-$(CONFIG_TEXTUREDSPENC) += texturedspenc.o
OBJS-$(CONFIG_TPELDSP) += tpeldsp.o
+OBJS-$(CONFIG_VAAPI_ENCODE) += vaapi_encode.o
OBJS-$(CONFIG_VC1DSP) += vc1dsp.o
OBJS-$(CONFIG_VIDEODSP) += videodsp.o
OBJS-$(CONFIG_VP3DSP) += vp3dsp.o
diff --git a/libavcodec/vaapi_encode.c b/libavcodec/vaapi_encode.c
new file mode 100644
index 0000000000..9f2083d1f7
--- /dev/null
+++ b/libavcodec/vaapi_encode.c
@@ -0,0 +1,1096 @@
+/*
+ * This file is part of Libav.
+ *
+ * Libav is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * Libav is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with Libav; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#include <inttypes.h>
+#include <string.h>
+
+#include "libavutil/avassert.h"
+#include "libavutil/log.h"
+#include "libavutil/pixdesc.h"
+
+#include "vaapi_encode.h"
+#include "avcodec.h"
+
+static const char *picture_type_name[] = { "IDR", "I", "P", "B" };
+
+static int vaapi_encode_make_packed_header(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ int type, char *data, size_t bit_len)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAStatus vas;
+ VABufferID param_buffer, data_buffer;
+ VAEncPackedHeaderParameterBuffer params = {
+ .type = type,
+ .bit_length = bit_len,
+ .has_emulation_bytes = 1,
+ };
+
+ av_assert0(pic->nb_param_buffers + 2 <= MAX_PARAM_BUFFERS);
+
+ vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+ VAEncPackedHeaderParameterBufferType,
+ sizeof(params), 1, &params, &param_buffer);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer "
+ "for packed header (type %d): %d (%s).\n",
+ type, vas, vaErrorStr(vas));
+ return AVERROR(EIO);
+ }
+ pic->param_buffers[pic->nb_param_buffers++] = param_buffer;
+
+ vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+ VAEncPackedHeaderDataBufferType,
+ (bit_len + 7) / 8, 1, data, &data_buffer);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create data buffer "
+ "for packed header (type %d): %d (%s).\n",
+ type, vas, vaErrorStr(vas));
+ return AVERROR(EIO);
+ }
+ pic->param_buffers[pic->nb_param_buffers++] = data_buffer;
+
+ av_log(avctx, AV_LOG_DEBUG, "Packed header buffer (%d) is %#x/%#x "
+ "(%zu bits).\n", type, param_buffer, data_buffer, bit_len);
+ return 0;
+}
+
+static int vaapi_encode_make_param_buffer(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ int type, char *data, size_t len)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAStatus vas;
+ VABufferID buffer;
+
+ av_assert0(pic->nb_param_buffers + 1 <= MAX_PARAM_BUFFERS);
+
+ vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+ type, len, 1, data, &buffer);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer "
+ "(type %d): %d (%s).\n", type, vas, vaErrorStr(vas));
+ return AVERROR(EIO);
+ }
+ pic->param_buffers[pic->nb_param_buffers++] = buffer;
+
+ av_log(avctx, AV_LOG_DEBUG, "Param buffer (%d) is %#x.\n",
+ type, buffer);
+ return 0;
+}
+
+static int vaapi_encode_wait(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAStatus vas;
+
+ av_assert0(pic->encode_issued);
+
+ if (pic->encode_complete) {
+ // Already waited for this picture.
+ return 0;
+ }
+
+ av_log(avctx, AV_LOG_DEBUG, "Sync to pic %"PRId64"/%"PRId64" "
+ "(recon surface %#x).\n", pic->display_order,
+ pic->encode_order, pic->recon_surface);
+
+ vas = vaSyncSurface(ctx->hwctx->display, pic->recon_surface);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to sync to picture completion: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ return AVERROR(EIO);
+ }
+
+ // Input is definitely finished with now.
+ av_frame_free(&pic->input_image);
+
+ pic->encode_complete = 1;
+ return 0;
+}
+
+static int vaapi_encode_issue(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodeSlice *slice;
+ VAStatus vas;
+ int err, i;
+ char data[MAX_PARAM_BUFFER_SIZE];
+ size_t bit_len;
+
+ av_log(avctx, AV_LOG_DEBUG, "Issuing encode for pic %"PRId64"/%"PRId64" "
+ "as type %s.\n", pic->display_order, pic->encode_order,
+ picture_type_name[pic->type]);
+ if (pic->nb_refs == 0) {
+ av_log(avctx, AV_LOG_DEBUG, "No reference pictures.\n");
+ } else {
+ av_log(avctx, AV_LOG_DEBUG, "Refers to:");
+ for (i = 0; i < pic->nb_refs; i++) {
+ av_log(avctx, AV_LOG_DEBUG, " %"PRId64"/%"PRId64,
+ pic->refs[i]->display_order, pic->refs[i]->encode_order);
+ }
+ av_log(avctx, AV_LOG_DEBUG, ".\n");
+ }
+
+ av_assert0(pic->input_available && !pic->encode_issued);
+ for (i = 0; i < pic->nb_refs; i++) {
+ av_assert0(pic->refs[i]);
+ // If we are serialised then the references must have already
+ // completed. If not, they must have been issued but need not
+ // have completed yet.
+ if (ctx->issue_mode == ISSUE_MODE_SERIALISE_EVERYTHING)
+ av_assert0(pic->refs[i]->encode_complete);
+ else
+ av_assert0(pic->refs[i]->encode_issued);
+ }
+
+ av_log(avctx, AV_LOG_DEBUG, "Input surface is %#x.\n", pic->input_surface);
+
+ pic->recon_image = av_frame_alloc();
+ if (!pic->recon_image) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+
+ err = av_hwframe_get_buffer(ctx->recon_frames_ref, pic->recon_image, 0);
+ if (err < 0) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ pic->recon_surface = (VASurfaceID)(uintptr_t)pic->recon_image->data[3];
+ av_log(avctx, AV_LOG_DEBUG, "Recon surface is %#x.\n", pic->recon_surface);
+
+ vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context,
+ VAEncCodedBufferType,
+ MAX_OUTPUT_BUFFER_SIZE, 1, 0,
+ &pic->output_buffer);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create bitstream "
+ "output buffer: %d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ av_log(avctx, AV_LOG_DEBUG, "Output buffer is %#x.\n",
+ pic->output_buffer);
+
+ if (ctx->codec->picture_params_size > 0) {
+ pic->codec_picture_params = av_malloc(ctx->codec->picture_params_size);
+ if (!pic->codec_picture_params)
+ goto fail;
+ memcpy(pic->codec_picture_params, ctx->codec_picture_params,
+ ctx->codec->picture_params_size);
+ } else {
+ av_assert0(!ctx->codec_picture_params);
+ }
+
+ pic->nb_param_buffers = 0;
+
+ if (pic->type == PICTURE_TYPE_IDR && ctx->codec->init_sequence_params) {
+ err = vaapi_encode_make_param_buffer(avctx, pic,
+ VAEncSequenceParameterBufferType,
+ ctx->codec_sequence_params,
+ ctx->codec->sequence_params_size);
+ if (err < 0)
+ goto fail;
+ }
+
+ if (ctx->codec->init_picture_params) {
+ err = ctx->codec->init_picture_params(avctx, pic);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to initialise picture "
+ "parameters: %d.\n", err);
+ goto fail;
+ }
+ err = vaapi_encode_make_param_buffer(avctx, pic,
+ VAEncPictureParameterBufferType,
+ pic->codec_picture_params,
+ ctx->codec->picture_params_size);
+ if (err < 0)
+ goto fail;
+ }
+
+ if (pic->type == PICTURE_TYPE_IDR) {
+ if (ctx->codec->write_sequence_header) {
+ bit_len = 8 * sizeof(data);
+ err = ctx->codec->write_sequence_header(avctx, data, &bit_len);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to write per-sequence "
+ "header: %d.\n", err);
+ goto fail;
+ }
+ err = vaapi_encode_make_packed_header(avctx, pic,
+ ctx->codec->sequence_header_type,
+ data, bit_len);
+ if (err < 0)
+ goto fail;
+ }
+ }
+
+ if (ctx->codec->write_picture_header) {
+ bit_len = 8 * sizeof(data);
+ err = ctx->codec->write_picture_header(avctx, pic, data, &bit_len);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to write per-picture "
+ "header: %d.\n", err);
+ goto fail;
+ }
+ err = vaapi_encode_make_packed_header(avctx, pic,
+ ctx->codec->picture_header_type,
+ data, bit_len);
+ if (err < 0)
+ goto fail;
+ }
+
+ if (ctx->codec->write_extra_buffer) {
+ for (i = 0;; i++) {
+ size_t len = sizeof(data);
+ int type;
+ err = ctx->codec->write_extra_buffer(avctx, pic, i, &type,
+ data, &len);
+ if (err == AVERROR_EOF)
+ break;
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to write extra "
+ "buffer %d: %d.\n", i, err);
+ goto fail;
+ }
+
+ err = vaapi_encode_make_param_buffer(avctx, pic, type,
+ data, len);
+ if (err < 0)
+ goto fail;
+ }
+ }
+
+ av_assert0(pic->nb_slices <= MAX_PICTURE_SLICES);
+ for (i = 0; i < pic->nb_slices; i++) {
+ slice = av_mallocz(sizeof(*slice));
+ if (!slice) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ pic->slices[i] = slice;
+
+ if (ctx->codec->slice_params_size > 0) {
+ slice->codec_slice_params = av_mallocz(ctx->codec->slice_params_size);
+ if (!slice->codec_slice_params) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ }
+
+ if (ctx->codec->init_slice_params) {
+ err = ctx->codec->init_slice_params(avctx, pic, slice);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to initalise slice "
+ "parameters: %d.\n", err);
+ goto fail;
+ }
+ }
+
+ if (ctx->codec->write_slice_header) {
+ bit_len = 8 * sizeof(data);
+ err = ctx->codec->write_slice_header(avctx, pic, slice,
+ data, &bit_len);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to write per-slice "
+ "header: %d.\n", err);
+ goto fail;
+ }
+ err = vaapi_encode_make_packed_header(avctx, pic,
+ ctx->codec->slice_header_type,
+ data, bit_len);
+ if (err < 0)
+ goto fail;
+ }
+
+ if (ctx->codec->init_slice_params) {
+ err = vaapi_encode_make_param_buffer(avctx, pic,
+ VAEncSliceParameterBufferType,
+ slice->codec_slice_params,
+ ctx->codec->slice_params_size);
+ if (err < 0)
+ goto fail;
+ }
+ }
+
+ vas = vaBeginPicture(ctx->hwctx->display, ctx->va_context,
+ pic->input_surface);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to begin picture encode issue: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail_with_picture;
+ }
+
+ vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context,
+ pic->param_buffers, pic->nb_param_buffers);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to upload encode parameters: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail_with_picture;
+ }
+
+ vas = vaEndPicture(ctx->hwctx->display, ctx->va_context);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to end picture encode issue: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail_at_end;
+ }
+
+ pic->encode_issued = 1;
+
+ if (ctx->issue_mode == ISSUE_MODE_SERIALISE_EVERYTHING)
+ return vaapi_encode_wait(avctx, pic);
+ else
+ return 0;
+
+fail_with_picture:
+ vaEndPicture(ctx->hwctx->display, ctx->va_context);
+fail:
+ for(i = 0; i < pic->nb_param_buffers; i++)
+ vaDestroyBuffer(ctx->hwctx->display, pic->param_buffers[i]);
+fail_at_end:
+ av_freep(&pic->codec_picture_params);
+ av_frame_free(&pic->recon_image);
+ return err;
+}
+
+static int vaapi_encode_output(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic, AVPacket *pkt)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VACodedBufferSegment *buf_list, *buf;
+ VAStatus vas;
+ int err;
+
+ err = vaapi_encode_wait(avctx, pic);
+ if (err < 0)
+ return err;
+
+ buf_list = NULL;
+ vas = vaMapBuffer(ctx->hwctx->display, pic->output_buffer,
+ (void**)&buf_list);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to map output buffers: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail;
+ }
+
+ for (buf = buf_list; buf; buf = buf->next) {
+ av_log(avctx, AV_LOG_DEBUG, "Output buffer: %u bytes "
+ "(status %08x).\n", buf->size, buf->status);
+
+ err = av_new_packet(pkt, buf->size);
+ if (err < 0)
+ goto fail;
+
+ memcpy(pkt->data, buf->buf, buf->size);
+ }
+
+ if (pic->type == PICTURE_TYPE_IDR)
+ pkt->flags |= AV_PKT_FLAG_KEY;
+
+ pkt->pts = pic->pts;
+
+ vas = vaUnmapBuffer(ctx->hwctx->display, pic->output_buffer);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to unmap output buffers: "
+ "%d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail;
+ }
+
+ vaDestroyBuffer(ctx->hwctx->display, pic->output_buffer);
+ pic->output_buffer = VA_INVALID_ID;
+
+ av_log(avctx, AV_LOG_DEBUG, "Output read for pic %"PRId64"/%"PRId64".\n",
+ pic->display_order, pic->encode_order);
+ return 0;
+
+fail:
+ if (pic->output_buffer != VA_INVALID_ID) {
+ vaUnmapBuffer(ctx->hwctx->display, pic->output_buffer);
+ vaDestroyBuffer(ctx->hwctx->display, pic->output_buffer);
+ pic->output_buffer = VA_INVALID_ID;
+ }
+ return err;
+}
+
+static int vaapi_encode_discard(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+
+ vaapi_encode_wait(avctx, pic);
+
+ if (pic->output_buffer != VA_INVALID_ID) {
+ av_log(avctx, AV_LOG_DEBUG, "Discard output for pic "
+ "%"PRId64"/%"PRId64".\n",
+ pic->display_order, pic->encode_order);
+
+ vaDestroyBuffer(ctx->hwctx->display, pic->output_buffer);
+ pic->output_buffer = VA_INVALID_ID;
+ }
+
+ return 0;
+}
+
+static VAAPIEncodePicture *vaapi_encode_alloc(void)
+{
+ VAAPIEncodePicture *pic;
+
+ pic = av_mallocz(sizeof(*pic));
+ if (!pic)
+ return NULL;
+
+ pic->input_surface = VA_INVALID_ID;
+ pic->recon_surface = VA_INVALID_ID;
+ pic->output_buffer = VA_INVALID_ID;
+
+ return pic;
+}
+
+static int vaapi_encode_free(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic)
+{
+ int i;
+
+ if (pic->encode_issued)
+ vaapi_encode_discard(avctx, pic);
+
+ for (i = 0; i < pic->nb_slices; i++) {
+ av_freep(&pic->slices[i]->priv_data);
+ av_freep(&pic->slices[i]->codec_slice_params);
+ av_freep(&pic->slices[i]);
+ }
+ av_freep(&pic->codec_picture_params);
+
+ av_frame_free(&pic->input_image);
+ av_frame_free(&pic->recon_image);
+
+ // Output buffer should already be destroyed.
+ av_assert0(pic->output_buffer == VA_INVALID_ID);
+
+ av_freep(&pic->priv_data);
+ av_freep(&pic->codec_picture_params);
+
+ av_free(pic);
+
+ return 0;
+}
+
+static int vaapi_encode_step(AVCodecContext *avctx,
+ VAAPIEncodePicture *target)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *pic;
+ int i, err;
+
+ if (ctx->issue_mode == ISSUE_MODE_SERIALISE_EVERYTHING ||
+ ctx->issue_mode == ISSUE_MODE_MINIMISE_LATENCY) {
+ // These two modes are equivalent, except that we wait for
+ // immediate completion on each operation if serialised.
+
+ if (!target) {
+ // No target, nothing to do yet.
+ return 0;
+ }
+
+ if (target->encode_complete) {
+ // Already done.
+ return 0;
+ }
+
+ pic = target;
+ for (i = 0; i < pic->nb_refs; i++) {
+ if (!pic->refs[i]->encode_complete) {
+ err = vaapi_encode_step(avctx, pic->refs[i]);
+ if (err < 0)
+ return err;
+ }
+ }
+
+ err = vaapi_encode_issue(avctx, pic);
+ if (err < 0)
+ return err;
+
+ } else if (ctx->issue_mode == ISSUE_MODE_MAXIMISE_THROUGHPUT) {
+ int activity;
+
+ do {
+ activity = 0;
+ for (pic = ctx->pic_start; pic; pic = pic->next) {
+ if (!pic->input_available || pic->encode_issued)
+ continue;
+ for (i = 0; i < pic->nb_refs; i++) {
+ if (!pic->refs[i]->encode_issued)
+ break;
+ }
+ if (i < pic->nb_refs)
+ continue;
+ err = vaapi_encode_issue(avctx, pic);
+ if (err < 0)
+ return err;
+ activity = 1;
+ }
+ } while(activity);
+
+ if (target) {
+ av_assert0(target->encode_issued && "broken dependencies?");
+ }
+
+ } else {
+ av_assert0(0);
+ }
+
+ return 0;
+}
+
+static int vaapi_encode_get_next(AVCodecContext *avctx,
+ VAAPIEncodePicture **pic_out)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *start, *end, *pic;
+ int i;
+
+ for (pic = ctx->pic_start; pic; pic = pic->next) {
+ if (pic->next)
+ av_assert0(pic->display_order + 1 == pic->next->display_order);
+ if (pic->display_order == ctx->input_order) {
+ *pic_out = pic;
+ return 0;
+ }
+ }
+
+ if (ctx->input_order == 0) {
+ // First frame is always an IDR frame.
+ av_assert0(!ctx->pic_start && !ctx->pic_end);
+
+ pic = vaapi_encode_alloc();
+ if (!pic)
+ return AVERROR(ENOMEM);
+
+ pic->type = PICTURE_TYPE_IDR;
+ pic->display_order = 0;
+ pic->encode_order = 0;
+
+ ctx->pic_start = ctx->pic_end = pic;
+
+ *pic_out = pic;
+ return 0;
+ }
+
+ pic = vaapi_encode_alloc();
+ if (!pic)
+ return AVERROR(ENOMEM);
+
+ if (ctx->p_per_i == 0 || ctx->p_counter == ctx->p_per_i) {
+ if (ctx->i_per_idr == 0 || ctx->i_counter == ctx->i_per_idr) {
+ pic->type = PICTURE_TYPE_IDR;
+ ctx->i_counter = 0;
+ } else {
+ pic->type = PICTURE_TYPE_I;
+ ++ctx->i_counter;
+ }
+ ctx->p_counter = 0;
+ } else {
+ pic->type = PICTURE_TYPE_P;
+ pic->refs[0] = ctx->pic_end;
+ pic->nb_refs = 1;
+ ++ctx->p_counter;
+ }
+ start = end = pic;
+
+ if (pic->type != PICTURE_TYPE_IDR) {
+ // If that was not an IDR frame, add B frames display-before and
+ // encode-after it.
+
+ for (i = 0; i < ctx->b_per_p; i++) {
+ pic = vaapi_encode_alloc();
+ if (!pic)
+ goto fail;
+
+ pic->type = PICTURE_TYPE_B;
+ pic->refs[0] = ctx->pic_end;
+ pic->refs[1] = end;
+ pic->nb_refs = 2;
+
+ pic->next = start;
+ pic->display_order = ctx->input_order + ctx->b_per_p - i - 1;
+ pic->encode_order = pic->display_order + 1;
+ start = pic;
+ }
+ }
+
+ for (i = 0, pic = start; pic; i++, pic = pic->next) {
+ pic->display_order = ctx->input_order + i;
+ if (end->type == PICTURE_TYPE_IDR)
+ pic->encode_order = ctx->input_order + i;
+ else if (pic == end)
+ pic->encode_order = ctx->input_order;
+ else
+ pic->encode_order = ctx->input_order + i + 1;
+ }
+
+ av_assert0(ctx->pic_end);
+ ctx->pic_end->next = start;
+ ctx->pic_end = end;
+
+ *pic_out = start;
+
+ av_log(avctx, AV_LOG_DEBUG, "Pictures:");
+ for (pic = ctx->pic_start; pic; pic = pic->next) {
+ av_log(avctx, AV_LOG_DEBUG, " %s (%"PRId64"/%"PRId64")",
+ picture_type_name[pic->type],
+ pic->display_order, pic->encode_order);
+ }
+ av_log(avctx, AV_LOG_DEBUG, "\n");
+
+ return 0;
+
+fail:
+ while (start) {
+ pic = start->next;
+ vaapi_encode_free(avctx, start);
+ start = pic;
+ }
+ return AVERROR(ENOMEM);
+}
+
+static int vaapi_encode_mangle_end(AVCodecContext *avctx)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *pic, *last_pic, *next;
+
+ // Find the last picture we actually have input for.
+ for (pic = ctx->pic_start; pic; pic = pic->next) {
+ if (!pic->input_available)
+ break;
+ last_pic = pic;
+ }
+
+ if (pic) {
+ av_assert0(last_pic);
+
+ if (last_pic->type == PICTURE_TYPE_B) {
+ // Some fixing up is required. Change the type of this
+ // picture to P, then modify preceeding B references which
+ // point beyond it to point at it instead.
+
+ last_pic->type = PICTURE_TYPE_P;
+ last_pic->encode_order = last_pic->refs[1]->encode_order;
+
+ for (pic = ctx->pic_start; pic != last_pic; pic = pic->next) {
+ if (pic->type == PICTURE_TYPE_B &&
+ pic->refs[1] == last_pic->refs[1])
+ pic->refs[1] = last_pic;
+ }
+
+ last_pic->nb_refs = 1;
+ last_pic->refs[1] = NULL;
+ } else {
+ // We can use the current structure (no references point
+ // beyond the end), but there are unused pics to discard.
+ }
+
+ // Discard all following pics, they will never be used.
+ for (pic = last_pic->next; pic; pic = next) {
+ next = pic->next;
+ vaapi_encode_free(avctx, pic);
+ }
+
+ last_pic->next = NULL;
+ ctx->pic_end = last_pic;
+
+ } else {
+ // Input is available for all pictures, so we don't need to
+ // mangle anything.
+ }
+
+ av_log(avctx, AV_LOG_DEBUG, "Pictures at end of stream:");
+ for (pic = ctx->pic_start; pic; pic = pic->next) {
+ av_log(avctx, AV_LOG_DEBUG, " %s (%"PRId64"/%"PRId64")",
+ picture_type_name[pic->type],
+ pic->display_order, pic->encode_order);
+ }
+ av_log(avctx, AV_LOG_DEBUG, "\n");
+
+ return 0;
+}
+
+static int vaapi_encode_clear_old(AVCodecContext *avctx)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *pic, *old;
+ int i;
+
+ while (ctx->pic_start != ctx->pic_end) {
+ old = ctx->pic_start;
+ if (old->encode_order > ctx->output_order)
+ break;
+
+ for (pic = old->next; pic; pic = pic->next) {
+ if (pic->encode_complete)
+ continue;
+ for (i = 0; i < pic->nb_refs; i++) {
+ if (pic->refs[i] == old) {
+ // We still need this picture because it's referred to
+ // directly by a later one, so it and all following
+ // pictures have to stay.
+ return 0;
+ }
+ }
+ }
+
+ pic = ctx->pic_start;
+ ctx->pic_start = pic->next;
+ vaapi_encode_free(avctx, pic);
+ }
+
+ return 0;
+}
+
+int ff_vaapi_encode2(AVCodecContext *avctx, AVPacket *pkt,
+ const AVFrame *input_image, int *got_packet)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *pic;
+ int err;
+
+ if (input_image) {
+ av_log(avctx, AV_LOG_DEBUG, "Encode frame: %ux%u (%"PRId64").\n",
+ input_image->width, input_image->height, input_image->pts);
+
+ err = vaapi_encode_get_next(avctx, &pic);
+ if (err) {
+ av_log(avctx, AV_LOG_ERROR, "Input setup failed: %d.\n", err);
+ return err;
+ }
+
+ pic->input_image = av_frame_alloc();
+ if (!pic->input_image) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ err = av_frame_ref(pic->input_image, input_image);
+ if (err < 0)
+ goto fail;
+ pic->input_surface = (VASurfaceID)(uintptr_t)input_image->data[3];
+ pic->pts = input_image->pts;
+
+ if (ctx->input_order == 0)
+ ctx->first_pts = pic->pts;
+ if (ctx->input_order == ctx->decode_delay)
+ ctx->dts_pts_diff = pic->pts - ctx->first_pts;
+ if (ctx->output_delay > 0)
+ ctx->ts_ring[ctx->input_order % (3 * ctx->output_delay)] = pic->pts;
+
+ pic->input_available = 1;
+
+ } else {
+ if (!ctx->end_of_stream) {
+ err = vaapi_encode_mangle_end(avctx);
+ if (err < 0)
+ goto fail;
+ ctx->end_of_stream = 1;
+ }
+ }
+
+ ++ctx->input_order;
+ ++ctx->output_order;
+ av_assert0(ctx->output_order + ctx->output_delay + 1 == ctx->input_order);
+
+ for (pic = ctx->pic_start; pic; pic = pic->next)
+ if (pic->encode_order == ctx->output_order)
+ break;
+
+ // pic can be null here if we don't have a specific target in this
+ // iteration. We might still issue encodes if things can be overlapped,
+ // even though we don't intend to output anything.
+
+ err = vaapi_encode_step(avctx, pic);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Encode failed: %d.\n", err);
+ goto fail;
+ }
+
+ if (!pic) {
+ *got_packet = 0;
+ } else {
+ err = vaapi_encode_output(avctx, pic, pkt);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Output failed: %d.\n", err);
+ goto fail;
+ }
+
+ if (ctx->output_delay == 0) {
+ pkt->dts = pkt->pts;
+ } else if (ctx->output_order < ctx->decode_delay) {
+ if (ctx->ts_ring[ctx->output_order] < INT64_MIN + ctx->dts_pts_diff)
+ pkt->dts = INT64_MIN;
+ else
+ pkt->dts = ctx->ts_ring[ctx->output_order] - ctx->dts_pts_diff;
+ } else {
+ pkt->dts = ctx->ts_ring[(ctx->output_order - ctx->decode_delay) %
+ (3 * ctx->output_delay)];
+ }
+
+ *got_packet = 1;
+ }
+
+ err = vaapi_encode_clear_old(avctx);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "List clearing failed: %d.\n", err);
+ goto fail;
+ }
+
+ return 0;
+
+fail:
+ // Unclear what to clean up on failure. There are probably some things we
+ // could do usefully clean up here, but for now just leave them for uninit()
+ // to do instead.
+ return err;
+}
+
+av_cold int ff_vaapi_encode_init(AVCodecContext *avctx,
+ const VAAPIEncodeType *type)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ AVVAAPIFramesContext *recon_hwctx = NULL;
+ AVVAAPIHWConfig *hwconfig = NULL;
+ AVHWFramesConstraints *constraints = NULL;
+ enum AVPixelFormat recon_format;
+ VAStatus vas;
+ int err, i;
+
+ if (!avctx->hw_frames_ctx) {
+ av_log(avctx, AV_LOG_ERROR, "A hardware frames reference is "
+ "required to associate the encoding device.\n");
+ return AVERROR(EINVAL);
+ }
+
+ ctx->codec = type;
+
+ ctx->priv_data = av_mallocz(type->priv_data_size);
+ if (!ctx->priv_data) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+
+ ctx->input_frames_ref = av_buffer_ref(avctx->hw_frames_ctx);
+ if (!ctx->input_frames_ref) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ ctx->input_frames = (AVHWFramesContext*)ctx->input_frames_ref->data;
+
+ ctx->device_ref = av_buffer_ref(ctx->input_frames->device_ref);
+ if (!ctx->device_ref) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ ctx->device = (AVHWDeviceContext*)ctx->device_ref->data;
+ ctx->hwctx = ctx->device->hwctx;
+
+ err = ctx->codec->init(avctx);
+ if (err < 0)
+ goto fail;
+
+ vas = vaCreateConfig(ctx->hwctx->display,
+ ctx->va_profile, ctx->va_entrypoint,
+ ctx->config_attributes, ctx->nb_config_attributes,
+ &ctx->va_config);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create encode pipeline "
+ "configuration: %d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail;
+ }
+
+ hwconfig = av_hwdevice_hwconfig_alloc(ctx->device_ref);
+ if (!hwconfig) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ hwconfig->config_id = ctx->va_config;
+
+ constraints = av_hwdevice_get_hwframe_constraints(ctx->device_ref,
+ hwconfig);
+ if (!constraints) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+
+ // Probably we can use the input surface format as the surface format
+ // of the reconstructed frames. If not, we just pick the first (only?)
+ // format in the valid list and hope that it all works.
+ recon_format = AV_PIX_FMT_NONE;
+ if (constraints->valid_sw_formats) {
+ for (i = 0; constraints->valid_sw_formats[i] != AV_PIX_FMT_NONE; i++) {
+ if (ctx->input_frames->sw_format ==
+ constraints->valid_sw_formats[i]) {
+ recon_format = ctx->input_frames->sw_format;
+ break;
+ }
+ }
+ if (recon_format == AV_PIX_FMT_NONE)
+ recon_format = constraints->valid_sw_formats[i];
+ } else {
+ // No idea what to use; copy input format.
+ recon_format = ctx->input_frames->sw_format;
+ }
+ av_log(avctx, AV_LOG_DEBUG, "Using %s as format of "
+ "reconstructed frames.\n", av_get_pix_fmt_name(recon_format));
+
+ if (ctx->aligned_width < constraints->min_width ||
+ ctx->aligned_height < constraints->min_height ||
+ ctx->aligned_width > constraints->max_width ||
+ ctx->aligned_height > constraints->max_height) {
+ av_log(avctx, AV_LOG_ERROR, "Hardware does not support encoding at "
+ "size %dx%d (constraints: width %d-%d height %d-%d).\n",
+ ctx->aligned_width, ctx->aligned_height,
+ constraints->min_width, constraints->max_width,
+ constraints->min_height, constraints->max_height);
+ err = AVERROR(EINVAL);
+ goto fail;
+ }
+
+ av_freep(&hwconfig);
+ av_hwframe_constraints_free(&constraints);
+
+ ctx->recon_frames_ref = av_hwframe_ctx_alloc(ctx->device_ref);
+ if (!ctx->recon_frames_ref) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ ctx->recon_frames = (AVHWFramesContext*)ctx->recon_frames_ref->data;
+
+ ctx->recon_frames->format = AV_PIX_FMT_VAAPI;
+ ctx->recon_frames->sw_format = recon_format;
+ ctx->recon_frames->width = ctx->aligned_width;
+ ctx->recon_frames->height = ctx->aligned_height;
+ ctx->recon_frames->initial_pool_size = ctx->nb_recon_frames;
+
+ err = av_hwframe_ctx_init(ctx->recon_frames_ref);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to initialise reconstructed "
+ "frame context: %d.\n", err);
+ goto fail;
+ }
+ recon_hwctx = ctx->recon_frames->hwctx;
+
+ vas = vaCreateContext(ctx->hwctx->display, ctx->va_config,
+ ctx->aligned_width, ctx->aligned_height,
+ VA_PROGRESSIVE,
+ recon_hwctx->surface_ids,
+ recon_hwctx->nb_surfaces,
+ &ctx->va_context);
+ if (vas != VA_STATUS_SUCCESS) {
+ av_log(avctx, AV_LOG_ERROR, "Failed to create encode pipeline "
+ "context: %d (%s).\n", vas, vaErrorStr(vas));
+ err = AVERROR(EIO);
+ goto fail;
+ }
+
+ ctx->input_order = 0;
+ ctx->output_delay = avctx->max_b_frames;
+ ctx->decode_delay = 1;
+ ctx->output_order = - ctx->output_delay - 1;
+
+ if (ctx->codec->sequence_params_size > 0) {
+ ctx->codec_sequence_params =
+ av_mallocz(ctx->codec->sequence_params_size);
+ if (!ctx->codec_sequence_params) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ }
+ if (ctx->codec->picture_params_size > 0) {
+ ctx->codec_picture_params =
+ av_mallocz(ctx->codec->picture_params_size);
+ if (!ctx->codec_picture_params) {
+ err = AVERROR(ENOMEM);
+ goto fail;
+ }
+ }
+
+ if (ctx->codec->init_sequence_params) {
+ err = ctx->codec->init_sequence_params(avctx);
+ if (err < 0) {
+ av_log(avctx, AV_LOG_ERROR, "Codec sequence initialisation "
+ "failed: %d.\n", err);
+ goto fail;
+ }
+ }
+
+ // All I are IDR for now.
+ ctx->i_per_idr = 0;
+ ctx->p_per_i = ((avctx->gop_size + avctx->max_b_frames) /
+ (avctx->max_b_frames + 1));
+ ctx->b_per_p = avctx->max_b_frames;
+
+ // This should be configurable somehow. (Needs testing on a machine
+ // where it actually overlaps properly, though.)
+ ctx->issue_mode = ISSUE_MODE_MAXIMISE_THROUGHPUT;
+
+ return 0;
+
+fail:
+ av_freep(&hwconfig);
+ av_hwframe_constraints_free(&constraints);
+ ff_vaapi_encode_close(avctx);
+ return err;
+}
+
+av_cold int ff_vaapi_encode_close(AVCodecContext *avctx)
+{
+ VAAPIEncodeContext *ctx = avctx->priv_data;
+ VAAPIEncodePicture *pic, *next;
+
+ for (pic = ctx->pic_start; pic; pic = next) {
+ next = pic->next;
+ vaapi_encode_free(avctx, pic);
+ }
+
+ if (ctx->va_context != VA_INVALID_ID)
+ vaDestroyContext(ctx->hwctx->display, ctx->va_context);
+
+ if (ctx->va_config != VA_INVALID_ID)
+ vaDestroyConfig(ctx->hwctx->display, ctx->va_config);
+
+ if (ctx->codec->close)
+ ctx->codec->close(avctx);
+
+ av_freep(&ctx->codec_sequence_params);
+ av_freep(&ctx->codec_picture_params);
+
+ av_buffer_unref(&ctx->recon_frames_ref);
+ av_buffer_unref(&ctx->input_frames_ref);
+ av_buffer_unref(&ctx->device_ref);
+
+ av_freep(&ctx->priv_data);
+
+ return 0;
+}
diff --git a/libavcodec/vaapi_encode.h b/libavcodec/vaapi_encode.h
new file mode 100644
index 0000000000..1aea45884b
--- /dev/null
+++ b/libavcodec/vaapi_encode.h
@@ -0,0 +1,217 @@
+/*
+ * This file is part of Libav.
+ *
+ * Libav is free software; you can redistribute it and/or
+ * modify it under the terms of the GNU Lesser General Public
+ * License as published by the Free Software Foundation; either
+ * version 2.1 of the License, or (at your option) any later version.
+ *
+ * Libav is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
+ * Lesser General Public License for more details.
+ *
+ * You should have received a copy of the GNU Lesser General Public
+ * License along with Libav; if not, write to the Free Software
+ * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
+ */
+
+#ifndef AVCODEC_VAAPI_ENCODE_H
+#define AVCODEC_VAAPI_ENCODE_H
+
+#include <stdint.h>
+
+#include <va/va.h>
+
+#include "libavutil/hwcontext.h"
+#include "libavutil/hwcontext_vaapi.h"
+
+#include "avcodec.h"
+
+struct VAAPIEncodeType;
+struct VAAPIEncodePicture;
+
+enum {
+ MAX_PICTURE_REFERENCES = 2,
+ MAX_PICTURE_SLICES = 1,
+ MAX_PARAM_BUFFERS = 16,
+ MAX_REORDER_DELAY = 16,
+ MAX_PARAM_BUFFER_SIZE = 1024,
+ MAX_OUTPUT_BUFFER_SIZE = 1024 * 1024,
+};
+
+enum {
+ PICTURE_TYPE_IDR = 0,
+ PICTURE_TYPE_I = 1,
+ PICTURE_TYPE_P = 2,
+ PICTURE_TYPE_B = 3,
+};
+
+enum {
+ // All encode operations are done independently.
+ ISSUE_MODE_SERIALISE_EVERYTHING = 0,
+ // Overlap as many operations as possible.
+ ISSUE_MODE_MAXIMISE_THROUGHPUT,
+ // Overlap operations only when satisfying parallel dependencies.
+ ISSUE_MODE_MINIMISE_LATENCY,
+};
+
+typedef struct VAAPIEncodeSlice {
+ void *priv_data;
+ void *codec_slice_params;
+} VAAPIEncodeSlice;
+
+typedef struct VAAPIEncodePicture {
+ struct VAAPIEncodePicture *next;
+
+ int64_t display_order;
+ int64_t encode_order;
+ int64_t pts;
+
+ int type;
+ int input_available;
+ int encode_issued;
+ int encode_complete;
+
+ AVFrame *input_image;
+ VASurfaceID input_surface;
+
+ AVFrame *recon_image;
+ VASurfaceID recon_surface;
+
+ int nb_param_buffers;
+ VABufferID param_buffers[MAX_PARAM_BUFFERS];
+
+ VABufferID output_buffer;
+
+ void *priv_data;
+ void *codec_picture_params;
+
+ int nb_refs;
+ struct VAAPIEncodePicture *refs[MAX_PICTURE_REFERENCES];
+
+ int nb_slices;
+ VAAPIEncodeSlice *slices[MAX_PICTURE_SLICES];
+} VAAPIEncodePicture;
+
+typedef struct VAAPIEncodeContext {
+ const AVClass *class;
+
+ // Codec-specific hooks.
+ const struct VAAPIEncodeType *codec;
+
+ // Codec-specific state.
+ void *priv_data;
+
+ VAProfile va_profile;
+ VAEntrypoint va_entrypoint;
+ VAConfigID va_config;
+ VAContextID va_context;
+
+ int va_rc_mode;
+
+ AVBufferRef *device_ref;
+ AVHWDeviceContext *device;
+ AVVAAPIDeviceContext *hwctx;
+
+ AVBufferRef *input_frames_ref;
+ AVHWFramesContext *input_frames;
+
+ // Input size, set from input frames.
+ int input_width;
+ int input_height;
+ // Aligned size, set by codec init, becomes hwframe size.
+ int aligned_width;
+ int aligned_height;
+
+ int nb_recon_frames;
+ AVBufferRef *recon_frames_ref;
+ AVHWFramesContext *recon_frames;
+
+ VAConfigAttrib *config_attributes;
+ int nb_config_attributes;
+
+ // Per-sequence parameter structure (VAEncSequenceParameterBuffer*).
+ void *codec_sequence_params;
+
+ // Per-sequence parameters found in the per-picture parameter
+ // structure (VAEncPictureParameterBuffer*).
+ void *codec_picture_params;
+
+ // Current encoding window, in display (input) order.
+ VAAPIEncodePicture *pic_start, *pic_end;
+
+ // Next input order index (display order).
+ int64_t input_order;
+ // Number of frames that output is behind input.
+ int64_t output_delay;
+ // Number of frames decode output will need to be delayed.
+ int64_t decode_delay;
+ // Next output order index (encode order).
+ int64_t output_order;
+
+ int issue_mode;
+
+ // Timestamp handling.
+ int64_t first_pts;
+ int64_t dts_pts_diff;
+ int64_t ts_ring[MAX_REORDER_DELAY * 3];
+
+ // Frame type decision.
+ int i_per_idr;
+ int p_per_i;
+ int b_per_p;
+ int idr_counter;
+ int i_counter;
+ int p_counter;
+ int end_of_stream;
+
+} VAAPIEncodeContext;
+
+
+typedef struct VAAPIEncodeType {
+ size_t priv_data_size;
+
+ int (*init)(AVCodecContext *avctx);
+ int (*close)(AVCodecContext *avctx);
+
+ size_t sequence_params_size;
+ size_t picture_params_size;
+ size_t slice_params_size;
+
+ int (*init_sequence_params)(AVCodecContext *avctx);
+ int (*init_picture_params)(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic);
+ int (*init_slice_params)(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ VAAPIEncodeSlice *slice);
+
+ int sequence_header_type;
+ int picture_header_type;
+ int slice_header_type;
+
+ int (*write_sequence_header)(AVCodecContext *avctx,
+ char *data, size_t *data_len);
+ int (*write_picture_header)(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ char *data, size_t *data_len);
+ int (*write_slice_header)(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ VAAPIEncodeSlice *slice,
+ char *data, size_t *data_len);
+
+ int (*write_extra_buffer)(AVCodecContext *avctx,
+ VAAPIEncodePicture *pic,
+ int index, int *type,
+ char *data, size_t *data_len);
+} VAAPIEncodeType;
+
+
+int ff_vaapi_encode2(AVCodecContext *avctx, AVPacket *pkt,
+ const AVFrame *input_image, int *got_packet);
+
+int ff_vaapi_encode_init(AVCodecContext *avctx,
+ const VAAPIEncodeType *type);
+int ff_vaapi_encode_close(AVCodecContext *avctx);
+
+#endif /* AVCODEC_VAAPI_ENCODE_H */