| /* |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include <inttypes.h> |
| #include <string.h> |
| |
| #include "libavutil/avassert.h" |
| #include "libavutil/common.h" |
| #include "libavutil/log.h" |
| #include "libavutil/pixdesc.h" |
| |
| #include "vaapi_encode.h" |
| #include "encode.h" |
| #include "avcodec.h" |
| |
| const AVCodecHWConfigInternal *const ff_vaapi_encode_hw_configs[] = { |
| HW_CONFIG_ENCODER_FRAMES(VAAPI, VAAPI), |
| NULL, |
| }; |
| |
| static const char * const picture_type_name[] = { "IDR", "I", "P", "B" }; |
| |
| static int vaapi_encode_make_packed_header(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| int type, char *data, size_t bit_len) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VABufferID param_buffer, data_buffer; |
| VABufferID *tmp; |
| VAEncPackedHeaderParameterBuffer params = { |
| .type = type, |
| .bit_length = bit_len, |
| .has_emulation_bytes = 1, |
| }; |
| |
| tmp = av_realloc_array(pic->param_buffers, sizeof(*tmp), pic->nb_param_buffers + 2); |
| if (!tmp) |
| return AVERROR(ENOMEM); |
| pic->param_buffers = tmp; |
| |
| vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, |
| VAEncPackedHeaderParameterBufferType, |
| sizeof(params), 1, ¶ms, ¶m_buffer); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer " |
| "for packed header (type %d): %d (%s).\n", |
| type, vas, vaErrorStr(vas)); |
| return AVERROR(EIO); |
| } |
| pic->param_buffers[pic->nb_param_buffers++] = param_buffer; |
| |
| vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, |
| VAEncPackedHeaderDataBufferType, |
| (bit_len + 7) / 8, 1, data, &data_buffer); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create data buffer " |
| "for packed header (type %d): %d (%s).\n", |
| type, vas, vaErrorStr(vas)); |
| return AVERROR(EIO); |
| } |
| pic->param_buffers[pic->nb_param_buffers++] = data_buffer; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Packed header buffer (%d) is %#x/%#x " |
| "(%zu bits).\n", type, param_buffer, data_buffer, bit_len); |
| return 0; |
| } |
| |
| static int vaapi_encode_make_param_buffer(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| int type, char *data, size_t len) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VABufferID *tmp; |
| VABufferID buffer; |
| |
| tmp = av_realloc_array(pic->param_buffers, sizeof(*tmp), pic->nb_param_buffers + 1); |
| if (!tmp) |
| return AVERROR(ENOMEM); |
| pic->param_buffers = tmp; |
| |
| vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, |
| type, len, 1, data, &buffer); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create parameter buffer " |
| "(type %d): %d (%s).\n", type, vas, vaErrorStr(vas)); |
| return AVERROR(EIO); |
| } |
| pic->param_buffers[pic->nb_param_buffers++] = buffer; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Param buffer (%d) is %#x.\n", |
| type, buffer); |
| return 0; |
| } |
| |
| static int vaapi_encode_make_misc_param_buffer(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| int type, |
| const void *data, size_t len) |
| { |
| // Construct the buffer on the stack - 1KB is much larger than any |
| // current misc parameter buffer type (the largest is EncQuality at |
| // 224 bytes). |
| uint8_t buffer[1024]; |
| VAEncMiscParameterBuffer header = { |
| .type = type, |
| }; |
| size_t buffer_size = sizeof(header) + len; |
| av_assert0(buffer_size <= sizeof(buffer)); |
| |
| memcpy(buffer, &header, sizeof(header)); |
| memcpy(buffer + sizeof(header), data, len); |
| |
| return vaapi_encode_make_param_buffer(avctx, pic, |
| VAEncMiscParameterBufferType, |
| buffer, buffer_size); |
| } |
| |
| static int vaapi_encode_wait(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| |
| av_assert0(pic->encode_issued); |
| |
| if (pic->encode_complete) { |
| // Already waited for this picture. |
| return 0; |
| } |
| |
| av_log(avctx, AV_LOG_DEBUG, "Sync to pic %"PRId64"/%"PRId64" " |
| "(input surface %#x).\n", pic->display_order, |
| pic->encode_order, pic->input_surface); |
| |
| vas = vaSyncSurface(ctx->hwctx->display, pic->input_surface); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to sync to picture completion: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR(EIO); |
| } |
| |
| // Input is definitely finished with now. |
| av_frame_free(&pic->input_image); |
| |
| pic->encode_complete = 1; |
| return 0; |
| } |
| |
| static int vaapi_encode_make_row_slice(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeSlice *slice; |
| int i, rounding; |
| |
| for (i = 0; i < pic->nb_slices; i++) |
| pic->slices[i].row_size = ctx->slice_size; |
| |
| rounding = ctx->slice_block_rows - ctx->nb_slices * ctx->slice_size; |
| if (rounding > 0) { |
| // Place rounding error at top and bottom of frame. |
| av_assert0(rounding < pic->nb_slices); |
| // Some Intel drivers contain a bug where the encoder will fail |
| // if the last slice is smaller than the one before it. Since |
| // that's straightforward to avoid here, just do so. |
| if (rounding <= 2) { |
| for (i = 0; i < rounding; i++) |
| ++pic->slices[i].row_size; |
| } else { |
| for (i = 0; i < (rounding + 1) / 2; i++) |
| ++pic->slices[pic->nb_slices - i - 1].row_size; |
| for (i = 0; i < rounding / 2; i++) |
| ++pic->slices[i].row_size; |
| } |
| } else if (rounding < 0) { |
| // Remove rounding error from last slice only. |
| av_assert0(rounding < ctx->slice_size); |
| pic->slices[pic->nb_slices - 1].row_size += rounding; |
| } |
| |
| for (i = 0; i < pic->nb_slices; i++) { |
| slice = &pic->slices[i]; |
| slice->index = i; |
| if (i == 0) { |
| slice->row_start = 0; |
| slice->block_start = 0; |
| } else { |
| const VAAPIEncodeSlice *prev = &pic->slices[i - 1]; |
| slice->row_start = prev->row_start + prev->row_size; |
| slice->block_start = prev->block_start + prev->block_size; |
| } |
| slice->block_size = slice->row_size * ctx->slice_block_cols; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Slice %d: %d-%d (%d rows), " |
| "%d-%d (%d blocks).\n", i, slice->row_start, |
| slice->row_start + slice->row_size - 1, slice->row_size, |
| slice->block_start, slice->block_start + slice->block_size - 1, |
| slice->block_size); |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_make_tile_slice(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeSlice *slice; |
| int i, j, index; |
| |
| for (i = 0; i < ctx->tile_cols; i++) { |
| for (j = 0; j < ctx->tile_rows; j++) { |
| index = j * ctx->tile_cols + i; |
| slice = &pic->slices[index]; |
| slice->index = index; |
| |
| pic->slices[index].block_start = ctx->col_bd[i] + |
| ctx->row_bd[j] * ctx->slice_block_cols; |
| pic->slices[index].block_size = ctx->row_height[j] * ctx->col_width[i]; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Slice %2d: (%2d, %2d) start at: %4d " |
| "width:%2d height:%2d (%d blocks).\n", index, ctx->col_bd[i], |
| ctx->row_bd[j], slice->block_start, ctx->col_width[i], |
| ctx->row_height[j], slice->block_size); |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_issue(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeSlice *slice; |
| VAStatus vas; |
| int err, i; |
| char data[MAX_PARAM_BUFFER_SIZE]; |
| size_t bit_len; |
| av_unused AVFrameSideData *sd; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Issuing encode for pic %"PRId64"/%"PRId64" " |
| "as type %s.\n", pic->display_order, pic->encode_order, |
| picture_type_name[pic->type]); |
| if (pic->nb_refs == 0) { |
| av_log(avctx, AV_LOG_DEBUG, "No reference pictures.\n"); |
| } else { |
| av_log(avctx, AV_LOG_DEBUG, "Refers to:"); |
| for (i = 0; i < pic->nb_refs; i++) { |
| av_log(avctx, AV_LOG_DEBUG, " %"PRId64"/%"PRId64, |
| pic->refs[i]->display_order, pic->refs[i]->encode_order); |
| } |
| av_log(avctx, AV_LOG_DEBUG, ".\n"); |
| } |
| |
| av_assert0(!pic->encode_issued); |
| for (i = 0; i < pic->nb_refs; i++) { |
| av_assert0(pic->refs[i]); |
| av_assert0(pic->refs[i]->encode_issued); |
| } |
| |
| av_log(avctx, AV_LOG_DEBUG, "Input surface is %#x.\n", pic->input_surface); |
| |
| pic->recon_image = av_frame_alloc(); |
| if (!pic->recon_image) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| |
| err = av_hwframe_get_buffer(ctx->recon_frames_ref, pic->recon_image, 0); |
| if (err < 0) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| pic->recon_surface = (VASurfaceID)(uintptr_t)pic->recon_image->data[3]; |
| av_log(avctx, AV_LOG_DEBUG, "Recon surface is %#x.\n", pic->recon_surface); |
| |
| pic->output_buffer_ref = av_buffer_pool_get(ctx->output_buffer_pool); |
| if (!pic->output_buffer_ref) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| pic->output_buffer = (VABufferID)(uintptr_t)pic->output_buffer_ref->data; |
| av_log(avctx, AV_LOG_DEBUG, "Output buffer is %#x.\n", |
| pic->output_buffer); |
| |
| if (ctx->codec->picture_params_size > 0) { |
| pic->codec_picture_params = av_malloc(ctx->codec->picture_params_size); |
| if (!pic->codec_picture_params) |
| goto fail; |
| memcpy(pic->codec_picture_params, ctx->codec_picture_params, |
| ctx->codec->picture_params_size); |
| } else { |
| av_assert0(!ctx->codec_picture_params); |
| } |
| |
| pic->nb_param_buffers = 0; |
| |
| if (pic->type == PICTURE_TYPE_IDR && ctx->codec->init_sequence_params) { |
| err = vaapi_encode_make_param_buffer(avctx, pic, |
| VAEncSequenceParameterBufferType, |
| ctx->codec_sequence_params, |
| ctx->codec->sequence_params_size); |
| if (err < 0) |
| goto fail; |
| } |
| |
| if (pic->type == PICTURE_TYPE_IDR) { |
| for (i = 0; i < ctx->nb_global_params; i++) { |
| err = vaapi_encode_make_misc_param_buffer(avctx, pic, |
| ctx->global_params_type[i], |
| ctx->global_params[i], |
| ctx->global_params_size[i]); |
| if (err < 0) |
| goto fail; |
| } |
| } |
| |
| if (ctx->codec->init_picture_params) { |
| err = ctx->codec->init_picture_params(avctx, pic); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to initialise picture " |
| "parameters: %d.\n", err); |
| goto fail; |
| } |
| err = vaapi_encode_make_param_buffer(avctx, pic, |
| VAEncPictureParameterBufferType, |
| pic->codec_picture_params, |
| ctx->codec->picture_params_size); |
| if (err < 0) |
| goto fail; |
| } |
| |
| if (pic->type == PICTURE_TYPE_IDR) { |
| if (ctx->va_packed_headers & VA_ENC_PACKED_HEADER_SEQUENCE && |
| ctx->codec->write_sequence_header) { |
| bit_len = 8 * sizeof(data); |
| err = ctx->codec->write_sequence_header(avctx, data, &bit_len); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write per-sequence " |
| "header: %d.\n", err); |
| goto fail; |
| } |
| err = vaapi_encode_make_packed_header(avctx, pic, |
| ctx->codec->sequence_header_type, |
| data, bit_len); |
| if (err < 0) |
| goto fail; |
| } |
| } |
| |
| if (ctx->va_packed_headers & VA_ENC_PACKED_HEADER_PICTURE && |
| ctx->codec->write_picture_header) { |
| bit_len = 8 * sizeof(data); |
| err = ctx->codec->write_picture_header(avctx, pic, data, &bit_len); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write per-picture " |
| "header: %d.\n", err); |
| goto fail; |
| } |
| err = vaapi_encode_make_packed_header(avctx, pic, |
| ctx->codec->picture_header_type, |
| data, bit_len); |
| if (err < 0) |
| goto fail; |
| } |
| |
| if (ctx->codec->write_extra_buffer) { |
| for (i = 0;; i++) { |
| size_t len = sizeof(data); |
| int type; |
| err = ctx->codec->write_extra_buffer(avctx, pic, i, &type, |
| data, &len); |
| if (err == AVERROR_EOF) |
| break; |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write extra " |
| "buffer %d: %d.\n", i, err); |
| goto fail; |
| } |
| |
| err = vaapi_encode_make_param_buffer(avctx, pic, type, |
| data, len); |
| if (err < 0) |
| goto fail; |
| } |
| } |
| |
| if (ctx->va_packed_headers & VA_ENC_PACKED_HEADER_MISC && |
| ctx->codec->write_extra_header) { |
| for (i = 0;; i++) { |
| int type; |
| bit_len = 8 * sizeof(data); |
| err = ctx->codec->write_extra_header(avctx, pic, i, &type, |
| data, &bit_len); |
| if (err == AVERROR_EOF) |
| break; |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write extra " |
| "header %d: %d.\n", i, err); |
| goto fail; |
| } |
| |
| err = vaapi_encode_make_packed_header(avctx, pic, type, |
| data, bit_len); |
| if (err < 0) |
| goto fail; |
| } |
| } |
| |
| if (pic->nb_slices == 0) |
| pic->nb_slices = ctx->nb_slices; |
| if (pic->nb_slices > 0) { |
| pic->slices = av_mallocz_array(pic->nb_slices, sizeof(*pic->slices)); |
| if (!pic->slices) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| |
| if (ctx->tile_rows && ctx->tile_cols) |
| vaapi_encode_make_tile_slice(avctx, pic); |
| else |
| vaapi_encode_make_row_slice(avctx, pic); |
| } |
| |
| for (i = 0; i < pic->nb_slices; i++) { |
| slice = &pic->slices[i]; |
| |
| if (ctx->codec->slice_params_size > 0) { |
| slice->codec_slice_params = av_mallocz(ctx->codec->slice_params_size); |
| if (!slice->codec_slice_params) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| } |
| |
| if (ctx->codec->init_slice_params) { |
| err = ctx->codec->init_slice_params(avctx, pic, slice); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to initialise slice " |
| "parameters: %d.\n", err); |
| goto fail; |
| } |
| } |
| |
| if (ctx->va_packed_headers & VA_ENC_PACKED_HEADER_SLICE && |
| ctx->codec->write_slice_header) { |
| bit_len = 8 * sizeof(data); |
| err = ctx->codec->write_slice_header(avctx, pic, slice, |
| data, &bit_len); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write per-slice " |
| "header: %d.\n", err); |
| goto fail; |
| } |
| err = vaapi_encode_make_packed_header(avctx, pic, |
| ctx->codec->slice_header_type, |
| data, bit_len); |
| if (err < 0) |
| goto fail; |
| } |
| |
| if (ctx->codec->init_slice_params) { |
| err = vaapi_encode_make_param_buffer(avctx, pic, |
| VAEncSliceParameterBufferType, |
| slice->codec_slice_params, |
| ctx->codec->slice_params_size); |
| if (err < 0) |
| goto fail; |
| } |
| } |
| |
| #if VA_CHECK_VERSION(1, 0, 0) |
| sd = av_frame_get_side_data(pic->input_image, |
| AV_FRAME_DATA_REGIONS_OF_INTEREST); |
| if (sd && ctx->roi_allowed) { |
| const AVRegionOfInterest *roi; |
| uint32_t roi_size; |
| VAEncMiscParameterBufferROI param_roi; |
| int nb_roi, i, v; |
| |
| roi = (const AVRegionOfInterest*)sd->data; |
| roi_size = roi->self_size; |
| av_assert0(roi_size && sd->size % roi_size == 0); |
| nb_roi = sd->size / roi_size; |
| if (nb_roi > ctx->roi_max_regions) { |
| if (!ctx->roi_warned) { |
| av_log(avctx, AV_LOG_WARNING, "More ROIs set than " |
| "supported by driver (%d > %d).\n", |
| nb_roi, ctx->roi_max_regions); |
| ctx->roi_warned = 1; |
| } |
| nb_roi = ctx->roi_max_regions; |
| } |
| |
| pic->roi = av_mallocz_array(nb_roi, sizeof(*pic->roi)); |
| if (!pic->roi) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| // For overlapping regions, the first in the array takes priority. |
| for (i = 0; i < nb_roi; i++) { |
| roi = (const AVRegionOfInterest*)(sd->data + roi_size * i); |
| |
| av_assert0(roi->qoffset.den != 0); |
| v = roi->qoffset.num * ctx->roi_quant_range / roi->qoffset.den; |
| av_log(avctx, AV_LOG_DEBUG, "ROI: (%d,%d)-(%d,%d) -> %+d.\n", |
| roi->top, roi->left, roi->bottom, roi->right, v); |
| |
| pic->roi[i] = (VAEncROI) { |
| .roi_rectangle = { |
| .x = roi->left, |
| .y = roi->top, |
| .width = roi->right - roi->left, |
| .height = roi->bottom - roi->top, |
| }, |
| .roi_value = av_clip_int8(v), |
| }; |
| } |
| |
| param_roi = (VAEncMiscParameterBufferROI) { |
| .num_roi = nb_roi, |
| .max_delta_qp = INT8_MAX, |
| .min_delta_qp = INT8_MIN, |
| .roi = pic->roi, |
| .roi_flags.bits.roi_value_is_qp_delta = 1, |
| }; |
| |
| err = vaapi_encode_make_misc_param_buffer(avctx, pic, |
| VAEncMiscParameterTypeROI, |
| ¶m_roi, |
| sizeof(param_roi)); |
| if (err < 0) |
| goto fail; |
| } |
| #endif |
| |
| vas = vaBeginPicture(ctx->hwctx->display, ctx->va_context, |
| pic->input_surface); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to begin picture encode issue: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail_with_picture; |
| } |
| |
| vas = vaRenderPicture(ctx->hwctx->display, ctx->va_context, |
| pic->param_buffers, pic->nb_param_buffers); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to upload encode parameters: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail_with_picture; |
| } |
| |
| vas = vaEndPicture(ctx->hwctx->display, ctx->va_context); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to end picture encode issue: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| // vaRenderPicture() has been called here, so we should not destroy |
| // the parameter buffers unless separate destruction is required. |
| if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks & |
| AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) |
| goto fail; |
| else |
| goto fail_at_end; |
| } |
| |
| if (CONFIG_VAAPI_1 || ctx->hwctx->driver_quirks & |
| AV_VAAPI_DRIVER_QUIRK_RENDER_PARAM_BUFFERS) { |
| for (i = 0; i < pic->nb_param_buffers; i++) { |
| vas = vaDestroyBuffer(ctx->hwctx->display, |
| pic->param_buffers[i]); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to destroy " |
| "param buffer %#x: %d (%s).\n", |
| pic->param_buffers[i], vas, vaErrorStr(vas)); |
| // And ignore. |
| } |
| } |
| } |
| |
| pic->encode_issued = 1; |
| |
| return 0; |
| |
| fail_with_picture: |
| vaEndPicture(ctx->hwctx->display, ctx->va_context); |
| fail: |
| for(i = 0; i < pic->nb_param_buffers; i++) |
| vaDestroyBuffer(ctx->hwctx->display, pic->param_buffers[i]); |
| if (pic->slices) { |
| for (i = 0; i < pic->nb_slices; i++) |
| av_freep(&pic->slices[i].codec_slice_params); |
| } |
| fail_at_end: |
| av_freep(&pic->codec_picture_params); |
| av_freep(&pic->param_buffers); |
| av_freep(&pic->slices); |
| av_freep(&pic->roi); |
| av_frame_free(&pic->recon_image); |
| av_buffer_unref(&pic->output_buffer_ref); |
| pic->output_buffer = VA_INVALID_ID; |
| return err; |
| } |
| |
| static int vaapi_encode_output(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, AVPacket *pkt) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VACodedBufferSegment *buf_list, *buf; |
| VAStatus vas; |
| int total_size = 0; |
| uint8_t *ptr; |
| int err; |
| |
| err = vaapi_encode_wait(avctx, pic); |
| if (err < 0) |
| return err; |
| |
| buf_list = NULL; |
| vas = vaMapBuffer(ctx->hwctx->display, pic->output_buffer, |
| (void**)&buf_list); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to map output buffers: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail; |
| } |
| |
| for (buf = buf_list; buf; buf = buf->next) |
| total_size += buf->size; |
| |
| err = av_new_packet(pkt, total_size); |
| ptr = pkt->data; |
| |
| if (err < 0) |
| goto fail_mapped; |
| |
| for (buf = buf_list; buf; buf = buf->next) { |
| av_log(avctx, AV_LOG_DEBUG, "Output buffer: %u bytes " |
| "(status %08x).\n", buf->size, buf->status); |
| |
| memcpy(ptr, buf->buf, buf->size); |
| ptr += buf->size; |
| } |
| |
| if (pic->type == PICTURE_TYPE_IDR) |
| pkt->flags |= AV_PKT_FLAG_KEY; |
| |
| pkt->pts = pic->pts; |
| |
| vas = vaUnmapBuffer(ctx->hwctx->display, pic->output_buffer); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to unmap output buffers: " |
| "%d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail; |
| } |
| |
| av_buffer_unref(&pic->output_buffer_ref); |
| pic->output_buffer = VA_INVALID_ID; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Output read for pic %"PRId64"/%"PRId64".\n", |
| pic->display_order, pic->encode_order); |
| return 0; |
| |
| fail_mapped: |
| vaUnmapBuffer(ctx->hwctx->display, pic->output_buffer); |
| fail: |
| av_buffer_unref(&pic->output_buffer_ref); |
| pic->output_buffer = VA_INVALID_ID; |
| return err; |
| } |
| |
| static int vaapi_encode_discard(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| vaapi_encode_wait(avctx, pic); |
| |
| if (pic->output_buffer_ref) { |
| av_log(avctx, AV_LOG_DEBUG, "Discard output for pic " |
| "%"PRId64"/%"PRId64".\n", |
| pic->display_order, pic->encode_order); |
| |
| av_buffer_unref(&pic->output_buffer_ref); |
| pic->output_buffer = VA_INVALID_ID; |
| } |
| |
| return 0; |
| } |
| |
| static VAAPIEncodePicture *vaapi_encode_alloc(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic; |
| |
| pic = av_mallocz(sizeof(*pic)); |
| if (!pic) |
| return NULL; |
| |
| if (ctx->codec->picture_priv_data_size > 0) { |
| pic->priv_data = av_mallocz(ctx->codec->picture_priv_data_size); |
| if (!pic->priv_data) { |
| av_freep(&pic); |
| return NULL; |
| } |
| } |
| |
| pic->input_surface = VA_INVALID_ID; |
| pic->recon_surface = VA_INVALID_ID; |
| pic->output_buffer = VA_INVALID_ID; |
| |
| return pic; |
| } |
| |
| static int vaapi_encode_free(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic) |
| { |
| int i; |
| |
| if (pic->encode_issued) |
| vaapi_encode_discard(avctx, pic); |
| |
| if (pic->slices) { |
| for (i = 0; i < pic->nb_slices; i++) |
| av_freep(&pic->slices[i].codec_slice_params); |
| } |
| av_freep(&pic->codec_picture_params); |
| |
| av_frame_free(&pic->input_image); |
| av_frame_free(&pic->recon_image); |
| |
| av_freep(&pic->param_buffers); |
| av_freep(&pic->slices); |
| // Output buffer should already be destroyed. |
| av_assert0(pic->output_buffer == VA_INVALID_ID); |
| |
| av_freep(&pic->priv_data); |
| av_freep(&pic->codec_picture_params); |
| av_freep(&pic->roi); |
| |
| av_free(pic); |
| |
| return 0; |
| } |
| |
| static void vaapi_encode_add_ref(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| VAAPIEncodePicture *target, |
| int is_ref, int in_dpb, int prev) |
| { |
| int refs = 0; |
| |
| if (is_ref) { |
| av_assert0(pic != target); |
| av_assert0(pic->nb_refs < MAX_PICTURE_REFERENCES); |
| pic->refs[pic->nb_refs++] = target; |
| ++refs; |
| } |
| |
| if (in_dpb) { |
| av_assert0(pic->nb_dpb_pics < MAX_DPB_SIZE); |
| pic->dpb[pic->nb_dpb_pics++] = target; |
| ++refs; |
| } |
| |
| if (prev) { |
| av_assert0(!pic->prev); |
| pic->prev = target; |
| ++refs; |
| } |
| |
| target->ref_count[0] += refs; |
| target->ref_count[1] += refs; |
| } |
| |
| static void vaapi_encode_remove_refs(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| int level) |
| { |
| int i; |
| |
| if (pic->ref_removed[level]) |
| return; |
| |
| for (i = 0; i < pic->nb_refs; i++) { |
| av_assert0(pic->refs[i]); |
| --pic->refs[i]->ref_count[level]; |
| av_assert0(pic->refs[i]->ref_count[level] >= 0); |
| } |
| |
| for (i = 0; i < pic->nb_dpb_pics; i++) { |
| av_assert0(pic->dpb[i]); |
| --pic->dpb[i]->ref_count[level]; |
| av_assert0(pic->dpb[i]->ref_count[level] >= 0); |
| } |
| |
| av_assert0(pic->prev || pic->type == PICTURE_TYPE_IDR); |
| if (pic->prev) { |
| --pic->prev->ref_count[level]; |
| av_assert0(pic->prev->ref_count[level] >= 0); |
| } |
| |
| pic->ref_removed[level] = 1; |
| } |
| |
| static void vaapi_encode_set_b_pictures(AVCodecContext *avctx, |
| VAAPIEncodePicture *start, |
| VAAPIEncodePicture *end, |
| VAAPIEncodePicture *prev, |
| int current_depth, |
| VAAPIEncodePicture **last) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic, *next, *ref; |
| int i, len; |
| |
| av_assert0(start && end && start != end && start->next != end); |
| |
| // If we are at the maximum depth then encode all pictures as |
| // non-referenced B-pictures. Also do this if there is exactly one |
| // picture left, since there will be nothing to reference it. |
| if (current_depth == ctx->max_b_depth || start->next->next == end) { |
| for (pic = start->next; pic; pic = pic->next) { |
| if (pic == end) |
| break; |
| pic->type = PICTURE_TYPE_B; |
| pic->b_depth = current_depth; |
| |
| vaapi_encode_add_ref(avctx, pic, start, 1, 1, 0); |
| vaapi_encode_add_ref(avctx, pic, end, 1, 1, 0); |
| vaapi_encode_add_ref(avctx, pic, prev, 0, 0, 1); |
| |
| for (ref = end->refs[1]; ref; ref = ref->refs[1]) |
| vaapi_encode_add_ref(avctx, pic, ref, 0, 1, 0); |
| } |
| *last = prev; |
| |
| } else { |
| // Split the current list at the midpoint with a referenced |
| // B-picture, then descend into each side separately. |
| len = 0; |
| for (pic = start->next; pic != end; pic = pic->next) |
| ++len; |
| for (pic = start->next, i = 1; 2 * i < len; pic = pic->next, i++); |
| |
| pic->type = PICTURE_TYPE_B; |
| pic->b_depth = current_depth; |
| |
| pic->is_reference = 1; |
| |
| vaapi_encode_add_ref(avctx, pic, pic, 0, 1, 0); |
| vaapi_encode_add_ref(avctx, pic, start, 1, 1, 0); |
| vaapi_encode_add_ref(avctx, pic, end, 1, 1, 0); |
| vaapi_encode_add_ref(avctx, pic, prev, 0, 0, 1); |
| |
| for (ref = end->refs[1]; ref; ref = ref->refs[1]) |
| vaapi_encode_add_ref(avctx, pic, ref, 0, 1, 0); |
| |
| if (i > 1) |
| vaapi_encode_set_b_pictures(avctx, start, pic, pic, |
| current_depth + 1, &next); |
| else |
| next = pic; |
| |
| vaapi_encode_set_b_pictures(avctx, pic, end, next, |
| current_depth + 1, last); |
| } |
| } |
| |
| static int vaapi_encode_pick_next(AVCodecContext *avctx, |
| VAAPIEncodePicture **pic_out) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic = NULL, *next, *start; |
| int i, b_counter, closed_gop_end; |
| |
| // If there are any B-frames already queued, the next one to encode |
| // is the earliest not-yet-issued frame for which all references are |
| // available. |
| for (pic = ctx->pic_start; pic; pic = pic->next) { |
| if (pic->encode_issued) |
| continue; |
| if (pic->type != PICTURE_TYPE_B) |
| continue; |
| for (i = 0; i < pic->nb_refs; i++) { |
| if (!pic->refs[i]->encode_issued) |
| break; |
| } |
| if (i == pic->nb_refs) |
| break; |
| } |
| |
| if (pic) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick B-picture at depth %d to " |
| "encode next.\n", pic->b_depth); |
| *pic_out = pic; |
| return 0; |
| } |
| |
| // Find the B-per-Pth available picture to become the next picture |
| // on the top layer. |
| start = NULL; |
| b_counter = 0; |
| closed_gop_end = ctx->closed_gop || |
| ctx->idr_counter == ctx->gop_per_idr; |
| for (pic = ctx->pic_start; pic; pic = next) { |
| next = pic->next; |
| if (pic->encode_issued) { |
| start = pic; |
| continue; |
| } |
| // If the next available picture is force-IDR, encode it to start |
| // a new GOP immediately. |
| if (pic->force_idr) |
| break; |
| if (b_counter == ctx->b_per_p) |
| break; |
| // If this picture ends a closed GOP or starts a new GOP then it |
| // needs to be in the top layer. |
| if (ctx->gop_counter + b_counter + closed_gop_end >= ctx->gop_size) |
| break; |
| // If the picture after this one is force-IDR, we need to encode |
| // this one in the top layer. |
| if (next && next->force_idr) |
| break; |
| ++b_counter; |
| } |
| |
| // At the end of the stream the last picture must be in the top layer. |
| if (!pic && ctx->end_of_stream) { |
| --b_counter; |
| pic = ctx->pic_end; |
| if (pic->encode_issued) |
| return AVERROR_EOF; |
| } |
| |
| if (!pic) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick nothing to encode next - " |
| "need more input for reference pictures.\n"); |
| return AVERROR(EAGAIN); |
| } |
| if (ctx->input_order <= ctx->decode_delay && !ctx->end_of_stream) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick nothing to encode next - " |
| "need more input for timestamps.\n"); |
| return AVERROR(EAGAIN); |
| } |
| |
| if (pic->force_idr) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick forced IDR-picture to " |
| "encode next.\n"); |
| pic->type = PICTURE_TYPE_IDR; |
| ctx->idr_counter = 1; |
| ctx->gop_counter = 1; |
| |
| } else if (ctx->gop_counter + b_counter >= ctx->gop_size) { |
| if (ctx->idr_counter == ctx->gop_per_idr) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick new-GOP IDR-picture to " |
| "encode next.\n"); |
| pic->type = PICTURE_TYPE_IDR; |
| ctx->idr_counter = 1; |
| } else { |
| av_log(avctx, AV_LOG_DEBUG, "Pick new-GOP I-picture to " |
| "encode next.\n"); |
| pic->type = PICTURE_TYPE_I; |
| ++ctx->idr_counter; |
| } |
| ctx->gop_counter = 1; |
| |
| } else { |
| if (ctx->gop_counter + b_counter + closed_gop_end == ctx->gop_size) { |
| av_log(avctx, AV_LOG_DEBUG, "Pick group-end P-picture to " |
| "encode next.\n"); |
| } else { |
| av_log(avctx, AV_LOG_DEBUG, "Pick normal P-picture to " |
| "encode next.\n"); |
| } |
| pic->type = PICTURE_TYPE_P; |
| av_assert0(start); |
| ctx->gop_counter += 1 + b_counter; |
| } |
| pic->is_reference = 1; |
| *pic_out = pic; |
| |
| vaapi_encode_add_ref(avctx, pic, pic, 0, 1, 0); |
| if (pic->type != PICTURE_TYPE_IDR) { |
| vaapi_encode_add_ref(avctx, pic, start, |
| pic->type == PICTURE_TYPE_P, |
| b_counter > 0, 0); |
| vaapi_encode_add_ref(avctx, pic, ctx->next_prev, 0, 0, 1); |
| } |
| if (ctx->next_prev) |
| --ctx->next_prev->ref_count[0]; |
| |
| if (b_counter > 0) { |
| vaapi_encode_set_b_pictures(avctx, start, pic, pic, 1, |
| &ctx->next_prev); |
| } else { |
| ctx->next_prev = pic; |
| } |
| ++ctx->next_prev->ref_count[0]; |
| return 0; |
| } |
| |
| static int vaapi_encode_clear_old(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic, *prev, *next; |
| |
| av_assert0(ctx->pic_start); |
| |
| // Remove direct references once each picture is complete. |
| for (pic = ctx->pic_start; pic; pic = pic->next) { |
| if (pic->encode_complete && pic->next) |
| vaapi_encode_remove_refs(avctx, pic, 0); |
| } |
| |
| // Remove indirect references once a picture has no direct references. |
| for (pic = ctx->pic_start; pic; pic = pic->next) { |
| if (pic->encode_complete && pic->ref_count[0] == 0) |
| vaapi_encode_remove_refs(avctx, pic, 1); |
| } |
| |
| // Clear out all complete pictures with no remaining references. |
| prev = NULL; |
| for (pic = ctx->pic_start; pic; pic = next) { |
| next = pic->next; |
| if (pic->encode_complete && pic->ref_count[1] == 0) { |
| av_assert0(pic->ref_removed[0] && pic->ref_removed[1]); |
| if (prev) |
| prev->next = next; |
| else |
| ctx->pic_start = next; |
| vaapi_encode_free(avctx, pic); |
| } else { |
| prev = pic; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_check_frame(AVCodecContext *avctx, |
| const AVFrame *frame) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| |
| if ((frame->crop_top || frame->crop_bottom || |
| frame->crop_left || frame->crop_right) && !ctx->crop_warned) { |
| av_log(avctx, AV_LOG_WARNING, "Cropping information on input " |
| "frames ignored due to lack of API support.\n"); |
| ctx->crop_warned = 1; |
| } |
| |
| if (!ctx->roi_allowed) { |
| AVFrameSideData *sd = |
| av_frame_get_side_data(frame, AV_FRAME_DATA_REGIONS_OF_INTEREST); |
| |
| if (sd && !ctx->roi_warned) { |
| av_log(avctx, AV_LOG_WARNING, "ROI side data on input " |
| "frames ignored due to lack of driver support.\n"); |
| ctx->roi_warned = 1; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_send_frame(AVCodecContext *avctx, AVFrame *frame) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic; |
| int err; |
| |
| if (frame) { |
| av_log(avctx, AV_LOG_DEBUG, "Input frame: %ux%u (%"PRId64").\n", |
| frame->width, frame->height, frame->pts); |
| |
| err = vaapi_encode_check_frame(avctx, frame); |
| if (err < 0) |
| return err; |
| |
| pic = vaapi_encode_alloc(avctx); |
| if (!pic) |
| return AVERROR(ENOMEM); |
| |
| pic->input_image = av_frame_alloc(); |
| if (!pic->input_image) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| |
| if (ctx->input_order == 0 || frame->pict_type == AV_PICTURE_TYPE_I) |
| pic->force_idr = 1; |
| |
| pic->input_surface = (VASurfaceID)(uintptr_t)frame->data[3]; |
| pic->pts = frame->pts; |
| |
| av_frame_move_ref(pic->input_image, frame); |
| |
| if (ctx->input_order == 0) |
| ctx->first_pts = pic->pts; |
| if (ctx->input_order == ctx->decode_delay) |
| ctx->dts_pts_diff = pic->pts - ctx->first_pts; |
| if (ctx->output_delay > 0) |
| ctx->ts_ring[ctx->input_order % (3 * ctx->output_delay)] = pic->pts; |
| |
| pic->display_order = ctx->input_order; |
| ++ctx->input_order; |
| |
| if (ctx->pic_start) { |
| ctx->pic_end->next = pic; |
| ctx->pic_end = pic; |
| } else { |
| ctx->pic_start = pic; |
| ctx->pic_end = pic; |
| } |
| |
| } else { |
| ctx->end_of_stream = 1; |
| |
| // Fix timestamps if we hit end-of-stream before the initial decode |
| // delay has elapsed. |
| if (ctx->input_order < ctx->decode_delay) |
| ctx->dts_pts_diff = ctx->pic_end->pts - ctx->first_pts; |
| } |
| |
| return 0; |
| |
| fail: |
| vaapi_encode_free(avctx, pic); |
| return err; |
| } |
| |
| int ff_vaapi_encode_receive_packet(AVCodecContext *avctx, AVPacket *pkt) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic; |
| AVFrame *frame = ctx->frame; |
| int err; |
| |
| err = ff_encode_get_frame(avctx, frame); |
| if (err < 0 && err != AVERROR_EOF) |
| return err; |
| |
| if (err == AVERROR_EOF) |
| frame = NULL; |
| |
| err = vaapi_encode_send_frame(avctx, frame); |
| if (err < 0) |
| return err; |
| |
| if (!ctx->pic_start) { |
| if (ctx->end_of_stream) |
| return AVERROR_EOF; |
| else |
| return AVERROR(EAGAIN); |
| } |
| |
| pic = NULL; |
| err = vaapi_encode_pick_next(avctx, &pic); |
| if (err < 0) |
| return err; |
| av_assert0(pic); |
| |
| pic->encode_order = ctx->encode_order++; |
| |
| err = vaapi_encode_issue(avctx, pic); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Encode failed: %d.\n", err); |
| return err; |
| } |
| |
| err = vaapi_encode_output(avctx, pic, pkt); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Output failed: %d.\n", err); |
| return err; |
| } |
| |
| if (ctx->output_delay == 0) { |
| pkt->dts = pkt->pts; |
| } else if (pic->encode_order < ctx->decode_delay) { |
| if (ctx->ts_ring[pic->encode_order] < INT64_MIN + ctx->dts_pts_diff) |
| pkt->dts = INT64_MIN; |
| else |
| pkt->dts = ctx->ts_ring[pic->encode_order] - ctx->dts_pts_diff; |
| } else { |
| pkt->dts = ctx->ts_ring[(pic->encode_order - ctx->decode_delay) % |
| (3 * ctx->output_delay)]; |
| } |
| av_log(avctx, AV_LOG_DEBUG, "Output packet: pts %"PRId64" dts %"PRId64".\n", |
| pkt->pts, pkt->dts); |
| |
| ctx->output_order = pic->encode_order; |
| vaapi_encode_clear_old(avctx); |
| |
| return 0; |
| } |
| |
| |
| static av_cold void vaapi_encode_add_global_param(AVCodecContext *avctx, int type, |
| void *buffer, size_t size) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| |
| av_assert0(ctx->nb_global_params < MAX_GLOBAL_PARAMS); |
| |
| ctx->global_params_type[ctx->nb_global_params] = type; |
| ctx->global_params [ctx->nb_global_params] = buffer; |
| ctx->global_params_size[ctx->nb_global_params] = size; |
| |
| ++ctx->nb_global_params; |
| } |
| |
| typedef struct VAAPIEncodeRTFormat { |
| const char *name; |
| unsigned int value; |
| int depth; |
| int nb_components; |
| int log2_chroma_w; |
| int log2_chroma_h; |
| } VAAPIEncodeRTFormat; |
| |
| static const VAAPIEncodeRTFormat vaapi_encode_rt_formats[] = { |
| { "YUV400", VA_RT_FORMAT_YUV400, 8, 1, }, |
| { "YUV420", VA_RT_FORMAT_YUV420, 8, 3, 1, 1 }, |
| { "YUV422", VA_RT_FORMAT_YUV422, 8, 3, 1, 0 }, |
| #if VA_CHECK_VERSION(1, 2, 0) |
| { "YUV422_10", VA_RT_FORMAT_YUV422_10, 10, 3, 1, 0 }, |
| #endif |
| { "YUV444", VA_RT_FORMAT_YUV444, 8, 3, 0, 0 }, |
| { "YUV411", VA_RT_FORMAT_YUV411, 8, 3, 2, 0 }, |
| #if VA_CHECK_VERSION(0, 38, 1) |
| { "YUV420_10", VA_RT_FORMAT_YUV420_10BPP, 10, 3, 1, 1 }, |
| #endif |
| }; |
| |
| static const VAEntrypoint vaapi_encode_entrypoints_normal[] = { |
| VAEntrypointEncSlice, |
| VAEntrypointEncPicture, |
| #if VA_CHECK_VERSION(0, 39, 2) |
| VAEntrypointEncSliceLP, |
| #endif |
| 0 |
| }; |
| #if VA_CHECK_VERSION(0, 39, 2) |
| static const VAEntrypoint vaapi_encode_entrypoints_low_power[] = { |
| VAEntrypointEncSliceLP, |
| 0 |
| }; |
| #endif |
| |
| static av_cold int vaapi_encode_profile_entrypoint(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAProfile *va_profiles = NULL; |
| VAEntrypoint *va_entrypoints = NULL; |
| VAStatus vas; |
| const VAEntrypoint *usable_entrypoints; |
| const VAAPIEncodeProfile *profile; |
| const AVPixFmtDescriptor *desc; |
| VAConfigAttrib rt_format_attr; |
| const VAAPIEncodeRTFormat *rt_format; |
| const char *profile_string, *entrypoint_string; |
| int i, j, n, depth, err; |
| |
| |
| if (ctx->low_power) { |
| #if VA_CHECK_VERSION(0, 39, 2) |
| usable_entrypoints = vaapi_encode_entrypoints_low_power; |
| #else |
| av_log(avctx, AV_LOG_ERROR, "Low-power encoding is not " |
| "supported with this VAAPI version.\n"); |
| return AVERROR(EINVAL); |
| #endif |
| } else { |
| usable_entrypoints = vaapi_encode_entrypoints_normal; |
| } |
| |
| desc = av_pix_fmt_desc_get(ctx->input_frames->sw_format); |
| if (!desc) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid input pixfmt (%d).\n", |
| ctx->input_frames->sw_format); |
| return AVERROR(EINVAL); |
| } |
| depth = desc->comp[0].depth; |
| for (i = 1; i < desc->nb_components; i++) { |
| if (desc->comp[i].depth != depth) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid input pixfmt (%s).\n", |
| desc->name); |
| return AVERROR(EINVAL); |
| } |
| } |
| av_log(avctx, AV_LOG_VERBOSE, "Input surface format is %s.\n", |
| desc->name); |
| |
| n = vaMaxNumProfiles(ctx->hwctx->display); |
| va_profiles = av_malloc_array(n, sizeof(VAProfile)); |
| if (!va_profiles) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| vas = vaQueryConfigProfiles(ctx->hwctx->display, va_profiles, &n); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query profiles: %d (%s).\n", |
| vas, vaErrorStr(vas)); |
| err = AVERROR_EXTERNAL; |
| goto fail; |
| } |
| |
| av_assert0(ctx->codec->profiles); |
| for (i = 0; (ctx->codec->profiles[i].av_profile != |
| FF_PROFILE_UNKNOWN); i++) { |
| profile = &ctx->codec->profiles[i]; |
| if (depth != profile->depth || |
| desc->nb_components != profile->nb_components) |
| continue; |
| if (desc->nb_components > 1 && |
| (desc->log2_chroma_w != profile->log2_chroma_w || |
| desc->log2_chroma_h != profile->log2_chroma_h)) |
| continue; |
| if (avctx->profile != profile->av_profile && |
| avctx->profile != FF_PROFILE_UNKNOWN) |
| continue; |
| |
| #if VA_CHECK_VERSION(1, 0, 0) |
| profile_string = vaProfileStr(profile->va_profile); |
| #else |
| profile_string = "(no profile names)"; |
| #endif |
| |
| for (j = 0; j < n; j++) { |
| if (va_profiles[j] == profile->va_profile) |
| break; |
| } |
| if (j >= n) { |
| av_log(avctx, AV_LOG_VERBOSE, "Compatible profile %s (%d) " |
| "is not supported by driver.\n", profile_string, |
| profile->va_profile); |
| continue; |
| } |
| |
| ctx->profile = profile; |
| break; |
| } |
| if (!ctx->profile) { |
| av_log(avctx, AV_LOG_ERROR, "No usable encoding profile found.\n"); |
| err = AVERROR(ENOSYS); |
| goto fail; |
| } |
| |
| avctx->profile = profile->av_profile; |
| ctx->va_profile = profile->va_profile; |
| av_log(avctx, AV_LOG_VERBOSE, "Using VAAPI profile %s (%d).\n", |
| profile_string, ctx->va_profile); |
| |
| n = vaMaxNumEntrypoints(ctx->hwctx->display); |
| va_entrypoints = av_malloc_array(n, sizeof(VAEntrypoint)); |
| if (!va_entrypoints) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| vas = vaQueryConfigEntrypoints(ctx->hwctx->display, ctx->va_profile, |
| va_entrypoints, &n); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query entrypoints for " |
| "profile %s (%d): %d (%s).\n", profile_string, |
| ctx->va_profile, vas, vaErrorStr(vas)); |
| err = AVERROR_EXTERNAL; |
| goto fail; |
| } |
| |
| for (i = 0; i < n; i++) { |
| for (j = 0; usable_entrypoints[j]; j++) { |
| if (va_entrypoints[i] == usable_entrypoints[j]) |
| break; |
| } |
| if (usable_entrypoints[j]) |
| break; |
| } |
| if (i >= n) { |
| av_log(avctx, AV_LOG_ERROR, "No usable encoding entrypoint found " |
| "for profile %s (%d).\n", profile_string, ctx->va_profile); |
| err = AVERROR(ENOSYS); |
| goto fail; |
| } |
| |
| ctx->va_entrypoint = va_entrypoints[i]; |
| #if VA_CHECK_VERSION(1, 0, 0) |
| entrypoint_string = vaEntrypointStr(ctx->va_entrypoint); |
| #else |
| entrypoint_string = "(no entrypoint names)"; |
| #endif |
| av_log(avctx, AV_LOG_VERBOSE, "Using VAAPI entrypoint %s (%d).\n", |
| entrypoint_string, ctx->va_entrypoint); |
| |
| for (i = 0; i < FF_ARRAY_ELEMS(vaapi_encode_rt_formats); i++) { |
| rt_format = &vaapi_encode_rt_formats[i]; |
| if (rt_format->depth == depth && |
| rt_format->nb_components == profile->nb_components && |
| rt_format->log2_chroma_w == profile->log2_chroma_w && |
| rt_format->log2_chroma_h == profile->log2_chroma_h) |
| break; |
| } |
| if (i >= FF_ARRAY_ELEMS(vaapi_encode_rt_formats)) { |
| av_log(avctx, AV_LOG_ERROR, "No usable render target format " |
| "found for profile %s (%d) entrypoint %s (%d).\n", |
| profile_string, ctx->va_profile, |
| entrypoint_string, ctx->va_entrypoint); |
| err = AVERROR(ENOSYS); |
| goto fail; |
| } |
| |
| rt_format_attr = (VAConfigAttrib) { VAConfigAttribRTFormat }; |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, ctx->va_entrypoint, |
| &rt_format_attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query RT format " |
| "config attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR_EXTERNAL; |
| goto fail; |
| } |
| |
| if (rt_format_attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_VERBOSE, "RT format config attribute not " |
| "supported by driver: assuming surface RT format %s " |
| "is valid.\n", rt_format->name); |
| } else if (!(rt_format_attr.value & rt_format->value)) { |
| av_log(avctx, AV_LOG_ERROR, "Surface RT format %s not supported " |
| "by driver for encoding profile %s (%d) entrypoint %s (%d).\n", |
| rt_format->name, profile_string, ctx->va_profile, |
| entrypoint_string, ctx->va_entrypoint); |
| err = AVERROR(ENOSYS); |
| goto fail; |
| } else { |
| av_log(avctx, AV_LOG_VERBOSE, "Using VAAPI render target " |
| "format %s (%#x).\n", rt_format->name, rt_format->value); |
| ctx->config_attributes[ctx->nb_config_attributes++] = |
| (VAConfigAttrib) { |
| .type = VAConfigAttribRTFormat, |
| .value = rt_format->value, |
| }; |
| } |
| |
| err = 0; |
| fail: |
| av_freep(&va_profiles); |
| av_freep(&va_entrypoints); |
| return err; |
| } |
| |
| static const VAAPIEncodeRCMode vaapi_encode_rc_modes[] = { |
| // Bitrate Quality |
| // | Maxrate | HRD/VBV |
| { 0 }, // | | | | |
| { RC_MODE_CQP, "CQP", 1, VA_RC_CQP, 0, 0, 1, 0 }, |
| { RC_MODE_CBR, "CBR", 1, VA_RC_CBR, 1, 0, 0, 1 }, |
| { RC_MODE_VBR, "VBR", 1, VA_RC_VBR, 1, 1, 0, 1 }, |
| #if VA_CHECK_VERSION(1, 1, 0) |
| { RC_MODE_ICQ, "ICQ", 1, VA_RC_ICQ, 0, 0, 1, 0 }, |
| #else |
| { RC_MODE_ICQ, "ICQ", 0 }, |
| #endif |
| #if VA_CHECK_VERSION(1, 3, 0) |
| { RC_MODE_QVBR, "QVBR", 1, VA_RC_QVBR, 1, 1, 1, 1 }, |
| { RC_MODE_AVBR, "AVBR", 0, VA_RC_AVBR, 1, 0, 0, 0 }, |
| #else |
| { RC_MODE_QVBR, "QVBR", 0 }, |
| { RC_MODE_AVBR, "AVBR", 0 }, |
| #endif |
| }; |
| |
| static av_cold int vaapi_encode_init_rate_control(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| uint32_t supported_va_rc_modes; |
| const VAAPIEncodeRCMode *rc_mode; |
| int64_t rc_bits_per_second; |
| int rc_target_percentage; |
| int rc_window_size; |
| int rc_quality; |
| int64_t hrd_buffer_size; |
| int64_t hrd_initial_buffer_fullness; |
| int fr_num, fr_den; |
| VAConfigAttrib rc_attr = { VAConfigAttribRateControl }; |
| VAStatus vas; |
| char supported_rc_modes_string[64]; |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, ctx->va_entrypoint, |
| &rc_attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query rate control " |
| "config attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| if (rc_attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_VERBOSE, "Driver does not report any " |
| "supported rate control modes: assuming CQP only.\n"); |
| supported_va_rc_modes = VA_RC_CQP; |
| strcpy(supported_rc_modes_string, "unknown"); |
| } else { |
| char *str = supported_rc_modes_string; |
| size_t len = sizeof(supported_rc_modes_string); |
| int i, first = 1, res; |
| |
| supported_va_rc_modes = rc_attr.value; |
| for (i = 0; i < FF_ARRAY_ELEMS(vaapi_encode_rc_modes); i++) { |
| rc_mode = &vaapi_encode_rc_modes[i]; |
| if (supported_va_rc_modes & rc_mode->va_mode) { |
| res = snprintf(str, len, "%s%s", |
| first ? "" : ", ", rc_mode->name); |
| first = 0; |
| if (res < 0) { |
| *str = 0; |
| break; |
| } |
| len -= res; |
| str += res; |
| if (len == 0) |
| break; |
| } |
| } |
| |
| av_log(avctx, AV_LOG_DEBUG, "Driver supports RC modes %s.\n", |
| supported_rc_modes_string); |
| } |
| |
| // Rate control mode selection: |
| // * If the user has set a mode explicitly with the rc_mode option, |
| // use it and fail if it is not available. |
| // * If an explicit QP option has been set, use CQP. |
| // * If the codec is CQ-only, use CQP. |
| // * If the QSCALE avcodec option is set, use CQP. |
| // * If bitrate and quality are both set, try QVBR. |
| // * If quality is set, try ICQ, then CQP. |
| // * If bitrate and maxrate are set and have the same value, try CBR. |
| // * If a bitrate is set, try AVBR, then VBR, then CBR. |
| // * If no bitrate is set, try ICQ, then CQP. |
| |
| #define TRY_RC_MODE(mode, fail) do { \ |
| rc_mode = &vaapi_encode_rc_modes[mode]; \ |
| if (!(rc_mode->va_mode & supported_va_rc_modes)) { \ |
| if (fail) { \ |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support %s " \ |
| "RC mode (supported modes: %s).\n", rc_mode->name, \ |
| supported_rc_modes_string); \ |
| return AVERROR(EINVAL); \ |
| } \ |
| av_log(avctx, AV_LOG_DEBUG, "Driver does not support %s " \ |
| "RC mode.\n", rc_mode->name); \ |
| rc_mode = NULL; \ |
| } else { \ |
| goto rc_mode_found; \ |
| } \ |
| } while (0) |
| |
| if (ctx->explicit_rc_mode) |
| TRY_RC_MODE(ctx->explicit_rc_mode, 1); |
| |
| if (ctx->explicit_qp) |
| TRY_RC_MODE(RC_MODE_CQP, 1); |
| |
| if (ctx->codec->flags & FLAG_CONSTANT_QUALITY_ONLY) |
| TRY_RC_MODE(RC_MODE_CQP, 1); |
| |
| if (avctx->flags & AV_CODEC_FLAG_QSCALE) |
| TRY_RC_MODE(RC_MODE_CQP, 1); |
| |
| if (avctx->bit_rate > 0 && avctx->global_quality > 0) |
| TRY_RC_MODE(RC_MODE_QVBR, 0); |
| |
| if (avctx->global_quality > 0) { |
| TRY_RC_MODE(RC_MODE_ICQ, 0); |
| TRY_RC_MODE(RC_MODE_CQP, 0); |
| } |
| |
| if (avctx->bit_rate > 0 && avctx->rc_max_rate == avctx->bit_rate) |
| TRY_RC_MODE(RC_MODE_CBR, 0); |
| |
| if (avctx->bit_rate > 0) { |
| TRY_RC_MODE(RC_MODE_AVBR, 0); |
| TRY_RC_MODE(RC_MODE_VBR, 0); |
| TRY_RC_MODE(RC_MODE_CBR, 0); |
| } else { |
| TRY_RC_MODE(RC_MODE_ICQ, 0); |
| TRY_RC_MODE(RC_MODE_CQP, 0); |
| } |
| |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support any " |
| "RC mode compatible with selected options " |
| "(supported modes: %s).\n", supported_rc_modes_string); |
| return AVERROR(EINVAL); |
| |
| rc_mode_found: |
| if (rc_mode->bitrate) { |
| if (avctx->bit_rate <= 0) { |
| av_log(avctx, AV_LOG_ERROR, "Bitrate must be set for %s " |
| "RC mode.\n", rc_mode->name); |
| return AVERROR(EINVAL); |
| } |
| |
| if (rc_mode->mode == RC_MODE_AVBR) { |
| // For maximum confusion AVBR is hacked into the existing API |
| // by overloading some of the fields with completely different |
| // meanings. |
| |
| // Target percentage does not apply in AVBR mode. |
| rc_bits_per_second = avctx->bit_rate; |
| |
| // Accuracy tolerance range for meeting the specified target |
| // bitrate. It's very unclear how this is actually intended |
| // to work - since we do want to get the specified bitrate, |
| // set the accuracy to 100% for now. |
| rc_target_percentage = 100; |
| |
| // Convergence period in frames. The GOP size reflects the |
| // user's intended block size for cutting, so reusing that |
| // as the convergence period seems a reasonable default. |
| rc_window_size = avctx->gop_size > 0 ? avctx->gop_size : 60; |
| |
| } else if (rc_mode->maxrate) { |
| if (avctx->rc_max_rate > 0) { |
| if (avctx->rc_max_rate < avctx->bit_rate) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid bitrate settings: " |
| "bitrate (%"PRId64") must not be greater than " |
| "maxrate (%"PRId64").\n", avctx->bit_rate, |
| avctx->rc_max_rate); |
| return AVERROR(EINVAL); |
| } |
| rc_bits_per_second = avctx->rc_max_rate; |
| rc_target_percentage = (avctx->bit_rate * 100) / |
| avctx->rc_max_rate; |
| } else { |
| // We only have a target bitrate, but this mode requires |
| // that a maximum rate be supplied as well. Since the |
| // user does not want this to be a constraint, arbitrarily |
| // pick a maximum rate of double the target rate. |
| rc_bits_per_second = 2 * avctx->bit_rate; |
| rc_target_percentage = 50; |
| } |
| } else { |
| if (avctx->rc_max_rate > avctx->bit_rate) { |
| av_log(avctx, AV_LOG_WARNING, "Max bitrate is ignored " |
| "in %s RC mode.\n", rc_mode->name); |
| } |
| rc_bits_per_second = avctx->bit_rate; |
| rc_target_percentage = 100; |
| } |
| } else { |
| rc_bits_per_second = 0; |
| rc_target_percentage = 100; |
| } |
| |
| if (rc_mode->quality) { |
| if (ctx->explicit_qp) { |
| rc_quality = ctx->explicit_qp; |
| } else if (avctx->global_quality > 0) { |
| rc_quality = avctx->global_quality; |
| } else { |
| rc_quality = ctx->codec->default_quality; |
| av_log(avctx, AV_LOG_WARNING, "No quality level set; " |
| "using default (%d).\n", rc_quality); |
| } |
| } else { |
| rc_quality = 0; |
| } |
| |
| if (rc_mode->hrd) { |
| if (avctx->rc_buffer_size) |
| hrd_buffer_size = avctx->rc_buffer_size; |
| else if (avctx->rc_max_rate > 0) |
| hrd_buffer_size = avctx->rc_max_rate; |
| else |
| hrd_buffer_size = avctx->bit_rate; |
| if (avctx->rc_initial_buffer_occupancy) { |
| if (avctx->rc_initial_buffer_occupancy > hrd_buffer_size) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid RC buffer settings: " |
| "must have initial buffer size (%d) <= " |
| "buffer size (%"PRId64").\n", |
| avctx->rc_initial_buffer_occupancy, hrd_buffer_size); |
| return AVERROR(EINVAL); |
| } |
| hrd_initial_buffer_fullness = avctx->rc_initial_buffer_occupancy; |
| } else { |
| hrd_initial_buffer_fullness = hrd_buffer_size * 3 / 4; |
| } |
| |
| rc_window_size = (hrd_buffer_size * 1000) / rc_bits_per_second; |
| } else { |
| if (avctx->rc_buffer_size || avctx->rc_initial_buffer_occupancy) { |
| av_log(avctx, AV_LOG_WARNING, "Buffering settings are ignored " |
| "in %s RC mode.\n", rc_mode->name); |
| } |
| |
| hrd_buffer_size = 0; |
| hrd_initial_buffer_fullness = 0; |
| |
| if (rc_mode->mode != RC_MODE_AVBR) { |
| // Already set (with completely different meaning) for AVBR. |
| rc_window_size = 1000; |
| } |
| } |
| |
| if (rc_bits_per_second > UINT32_MAX || |
| hrd_buffer_size > UINT32_MAX || |
| hrd_initial_buffer_fullness > UINT32_MAX) { |
| av_log(avctx, AV_LOG_ERROR, "RC parameters of 2^32 or " |
| "greater are not supported by VAAPI.\n"); |
| return AVERROR(EINVAL); |
| } |
| |
| ctx->rc_mode = rc_mode; |
| ctx->rc_quality = rc_quality; |
| ctx->va_rc_mode = rc_mode->va_mode; |
| ctx->va_bit_rate = rc_bits_per_second; |
| |
| av_log(avctx, AV_LOG_VERBOSE, "RC mode: %s.\n", rc_mode->name); |
| if (rc_attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| // This driver does not want the RC mode attribute to be set. |
| } else { |
| ctx->config_attributes[ctx->nb_config_attributes++] = |
| (VAConfigAttrib) { |
| .type = VAConfigAttribRateControl, |
| .value = ctx->va_rc_mode, |
| }; |
| } |
| |
| if (rc_mode->quality) |
| av_log(avctx, AV_LOG_VERBOSE, "RC quality: %d.\n", rc_quality); |
| |
| if (rc_mode->va_mode != VA_RC_CQP) { |
| if (rc_mode->mode == RC_MODE_AVBR) { |
| av_log(avctx, AV_LOG_VERBOSE, "RC target: %"PRId64" bps " |
| "converging in %d frames with %d%% accuracy.\n", |
| rc_bits_per_second, rc_window_size, |
| rc_target_percentage); |
| } else if (rc_mode->bitrate) { |
| av_log(avctx, AV_LOG_VERBOSE, "RC target: %d%% of " |
| "%"PRId64" bps over %d ms.\n", rc_target_percentage, |
| rc_bits_per_second, rc_window_size); |
| } |
| |
| ctx->rc_params = (VAEncMiscParameterRateControl) { |
| .bits_per_second = rc_bits_per_second, |
| .target_percentage = rc_target_percentage, |
| .window_size = rc_window_size, |
| .initial_qp = 0, |
| .min_qp = (avctx->qmin > 0 ? avctx->qmin : 0), |
| .basic_unit_size = 0, |
| #if VA_CHECK_VERSION(1, 1, 0) |
| .ICQ_quality_factor = av_clip(rc_quality, 1, 51), |
| .max_qp = (avctx->qmax > 0 ? avctx->qmax : 0), |
| #endif |
| #if VA_CHECK_VERSION(1, 3, 0) |
| .quality_factor = rc_quality, |
| #endif |
| }; |
| vaapi_encode_add_global_param(avctx, |
| VAEncMiscParameterTypeRateControl, |
| &ctx->rc_params, |
| sizeof(ctx->rc_params)); |
| } |
| |
| if (rc_mode->hrd) { |
| av_log(avctx, AV_LOG_VERBOSE, "RC buffer: %"PRId64" bits, " |
| "initial fullness %"PRId64" bits.\n", |
| hrd_buffer_size, hrd_initial_buffer_fullness); |
| |
| ctx->hrd_params = (VAEncMiscParameterHRD) { |
| .initial_buffer_fullness = hrd_initial_buffer_fullness, |
| .buffer_size = hrd_buffer_size, |
| }; |
| vaapi_encode_add_global_param(avctx, |
| VAEncMiscParameterTypeHRD, |
| &ctx->hrd_params, |
| sizeof(ctx->hrd_params)); |
| } |
| |
| if (avctx->framerate.num > 0 && avctx->framerate.den > 0) |
| av_reduce(&fr_num, &fr_den, |
| avctx->framerate.num, avctx->framerate.den, 65535); |
| else |
| av_reduce(&fr_num, &fr_den, |
| avctx->time_base.den, avctx->time_base.num, 65535); |
| |
| av_log(avctx, AV_LOG_VERBOSE, "RC framerate: %d/%d (%.2f fps).\n", |
| fr_num, fr_den, (double)fr_num / fr_den); |
| |
| ctx->fr_params = (VAEncMiscParameterFrameRate) { |
| .framerate = (unsigned int)fr_den << 16 | fr_num, |
| }; |
| #if VA_CHECK_VERSION(0, 40, 0) |
| vaapi_encode_add_global_param(avctx, |
| VAEncMiscParameterTypeFrameRate, |
| &ctx->fr_params, |
| sizeof(ctx->fr_params)); |
| #endif |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_gop_structure(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VAConfigAttrib attr = { VAConfigAttribEncMaxRefFrames }; |
| uint32_t ref_l0, ref_l1; |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, |
| ctx->va_entrypoint, |
| &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query reference frames " |
| "attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| |
| if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| ref_l0 = ref_l1 = 0; |
| } else { |
| ref_l0 = attr.value & 0xffff; |
| ref_l1 = attr.value >> 16 & 0xffff; |
| } |
| |
| if (ctx->codec->flags & FLAG_INTRA_ONLY || |
| avctx->gop_size <= 1) { |
| av_log(avctx, AV_LOG_VERBOSE, "Using intra frames only.\n"); |
| ctx->gop_size = 1; |
| } else if (ref_l0 < 1) { |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support any " |
| "reference frames.\n"); |
| return AVERROR(EINVAL); |
| } else if (!(ctx->codec->flags & FLAG_B_PICTURES) || |
| ref_l1 < 1 || avctx->max_b_frames < 1) { |
| av_log(avctx, AV_LOG_VERBOSE, "Using intra and P-frames " |
| "(supported references: %d / %d).\n", ref_l0, ref_l1); |
| ctx->gop_size = avctx->gop_size; |
| ctx->p_per_i = INT_MAX; |
| ctx->b_per_p = 0; |
| } else { |
| av_log(avctx, AV_LOG_VERBOSE, "Using intra, P- and B-frames " |
| "(supported references: %d / %d).\n", ref_l0, ref_l1); |
| ctx->gop_size = avctx->gop_size; |
| ctx->p_per_i = INT_MAX; |
| ctx->b_per_p = avctx->max_b_frames; |
| if (ctx->codec->flags & FLAG_B_PICTURE_REFERENCES) { |
| ctx->max_b_depth = FFMIN(ctx->desired_b_depth, |
| av_log2(ctx->b_per_p) + 1); |
| } else { |
| ctx->max_b_depth = 1; |
| } |
| } |
| |
| if (ctx->codec->flags & FLAG_NON_IDR_KEY_PICTURES) { |
| ctx->closed_gop = !!(avctx->flags & AV_CODEC_FLAG_CLOSED_GOP); |
| ctx->gop_per_idr = ctx->idr_interval + 1; |
| } else { |
| ctx->closed_gop = 1; |
| ctx->gop_per_idr = 1; |
| } |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_row_slice_structure(AVCodecContext *avctx, |
| uint32_t slice_structure) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| int req_slices; |
| |
| // For fixed-size slices currently we only support whole rows, making |
| // rectangular slices. This could be extended to arbitrary runs of |
| // blocks, but since slices tend to be a conformance requirement and |
| // most cases (such as broadcast or bluray) want rectangular slices |
| // only it would need to be gated behind another option. |
| if (avctx->slices > ctx->slice_block_rows) { |
| av_log(avctx, AV_LOG_WARNING, "Not enough rows to use " |
| "configured number of slices (%d < %d); using " |
| "maximum.\n", ctx->slice_block_rows, avctx->slices); |
| req_slices = ctx->slice_block_rows; |
| } else { |
| req_slices = avctx->slices; |
| } |
| if (slice_structure & VA_ENC_SLICE_STRUCTURE_ARBITRARY_ROWS || |
| slice_structure & VA_ENC_SLICE_STRUCTURE_ARBITRARY_MACROBLOCKS) { |
| ctx->nb_slices = req_slices; |
| ctx->slice_size = ctx->slice_block_rows / ctx->nb_slices; |
| } else if (slice_structure & VA_ENC_SLICE_STRUCTURE_POWER_OF_TWO_ROWS) { |
| int k; |
| for (k = 1;; k *= 2) { |
| if (2 * k * (req_slices - 1) + 1 >= ctx->slice_block_rows) |
| break; |
| } |
| ctx->nb_slices = (ctx->slice_block_rows + k - 1) / k; |
| ctx->slice_size = k; |
| #if VA_CHECK_VERSION(1, 0, 0) |
| } else if (slice_structure & VA_ENC_SLICE_STRUCTURE_EQUAL_ROWS) { |
| ctx->nb_slices = ctx->slice_block_rows; |
| ctx->slice_size = 1; |
| #endif |
| } else { |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support any usable " |
| "slice structure modes (%#x).\n", slice_structure); |
| return AVERROR(EINVAL); |
| } |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_tile_slice_structure(AVCodecContext *avctx, |
| uint32_t slice_structure) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| int i, req_tiles; |
| |
| if (!(slice_structure & VA_ENC_SLICE_STRUCTURE_ARBITRARY_MACROBLOCKS || |
| (slice_structure & VA_ENC_SLICE_STRUCTURE_ARBITRARY_ROWS && |
| ctx->tile_cols == 1))) { |
| av_log(avctx, AV_LOG_ERROR, "Supported slice structure (%#x) doesn't work for " |
| "current tile requirement.\n", slice_structure); |
| return AVERROR(EINVAL); |
| } |
| |
| if (ctx->tile_rows > ctx->slice_block_rows || |
| ctx->tile_cols > ctx->slice_block_cols) { |
| av_log(avctx, AV_LOG_WARNING, "Not enough block rows/cols (%d x %d) " |
| "for configured number of tile (%d x %d); ", |
| ctx->slice_block_rows, ctx->slice_block_cols, |
| ctx->tile_rows, ctx->tile_cols); |
| ctx->tile_rows = ctx->tile_rows > ctx->slice_block_rows ? |
| ctx->slice_block_rows : ctx->tile_rows; |
| ctx->tile_cols = ctx->tile_cols > ctx->slice_block_cols ? |
| ctx->slice_block_cols : ctx->tile_cols; |
| av_log(avctx, AV_LOG_WARNING, "using allowed maximum (%d x %d).\n", |
| ctx->tile_rows, ctx->tile_cols); |
| } |
| |
| req_tiles = ctx->tile_rows * ctx->tile_cols; |
| |
| // Tile slice is not allowed to cross the boundary of a tile due to |
| // the constraints of media-driver. Currently we support one slice |
| // per tile. This could be extended to multiple slices per tile. |
| if (avctx->slices != req_tiles) |
| av_log(avctx, AV_LOG_WARNING, "The number of requested slices " |
| "mismatches with configured number of tile (%d != %d); " |
| "using requested tile number for slice.\n", |
| avctx->slices, req_tiles); |
| |
| ctx->nb_slices = req_tiles; |
| |
| // Default in uniform spacing |
| // 6-3, 6-5 |
| for (i = 0; i < ctx->tile_cols; i++) { |
| ctx->col_width[i] = ( i + 1 ) * ctx->slice_block_cols / ctx->tile_cols - |
| i * ctx->slice_block_cols / ctx->tile_cols; |
| ctx->col_bd[i + 1] = ctx->col_bd[i] + ctx->col_width[i]; |
| } |
| // 6-4, 6-6 |
| for (i = 0; i < ctx->tile_rows; i++) { |
| ctx->row_height[i] = ( i + 1 ) * ctx->slice_block_rows / ctx->tile_rows - |
| i * ctx->slice_block_rows / ctx->tile_rows; |
| ctx->row_bd[i + 1] = ctx->row_bd[i] + ctx->row_height[i]; |
| } |
| |
| av_log(avctx, AV_LOG_VERBOSE, "Encoding pictures with %d x %d tile.\n", |
| ctx->tile_rows, ctx->tile_cols); |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_slice_structure(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAConfigAttrib attr[3] = { { VAConfigAttribEncMaxSlices }, |
| { VAConfigAttribEncSliceStructure }, |
| #if VA_CHECK_VERSION(1, 1, 0) |
| { VAConfigAttribEncTileSupport }, |
| #endif |
| }; |
| VAStatus vas; |
| uint32_t max_slices, slice_structure; |
| int ret; |
| |
| if (!(ctx->codec->flags & FLAG_SLICE_CONTROL)) { |
| if (avctx->slices > 0) { |
| av_log(avctx, AV_LOG_WARNING, "Multiple slices were requested " |
| "but this codec does not support controlling slices.\n"); |
| } |
| return 0; |
| } |
| |
| ctx->slice_block_rows = (avctx->height + ctx->slice_block_height - 1) / |
| ctx->slice_block_height; |
| ctx->slice_block_cols = (avctx->width + ctx->slice_block_width - 1) / |
| ctx->slice_block_width; |
| |
| if (avctx->slices <= 1 && !ctx->tile_rows && !ctx->tile_cols) { |
| ctx->nb_slices = 1; |
| ctx->slice_size = ctx->slice_block_rows; |
| return 0; |
| } |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, |
| ctx->va_entrypoint, |
| attr, FF_ARRAY_ELEMS(attr)); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query slice " |
| "attributes: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| max_slices = attr[0].value; |
| slice_structure = attr[1].value; |
| if (max_slices == VA_ATTRIB_NOT_SUPPORTED || |
| slice_structure == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support encoding " |
| "pictures as multiple slices.\n."); |
| return AVERROR(EINVAL); |
| } |
| |
| if (ctx->tile_rows && ctx->tile_cols) { |
| #if VA_CHECK_VERSION(1, 1, 0) |
| uint32_t tile_support = attr[2].value; |
| if (tile_support == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support encoding " |
| "pictures as multiple tiles.\n."); |
| return AVERROR(EINVAL); |
| } |
| #else |
| av_log(avctx, AV_LOG_ERROR, "Tile encoding option is " |
| "not supported with this VAAPI version.\n"); |
| return AVERROR(EINVAL); |
| #endif |
| } |
| |
| if (ctx->tile_rows && ctx->tile_cols) |
| ret = vaapi_encode_init_tile_slice_structure(avctx, slice_structure); |
| else |
| ret = vaapi_encode_init_row_slice_structure(avctx, slice_structure); |
| if (ret < 0) |
| return ret; |
| |
| if (ctx->nb_slices > avctx->slices) { |
| av_log(avctx, AV_LOG_WARNING, "Slice count rounded up to " |
| "%d (from %d) due to driver constraints on slice " |
| "structure.\n", ctx->nb_slices, avctx->slices); |
| } |
| if (ctx->nb_slices > max_slices) { |
| av_log(avctx, AV_LOG_ERROR, "Driver does not support " |
| "encoding with %d slices (max %"PRIu32").\n", |
| ctx->nb_slices, max_slices); |
| return AVERROR(EINVAL); |
| } |
| |
| av_log(avctx, AV_LOG_VERBOSE, "Encoding pictures with %d slices.\n", |
| ctx->nb_slices); |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_packed_headers(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VAConfigAttrib attr = { VAConfigAttribEncPackedHeaders }; |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, |
| ctx->va_entrypoint, |
| &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query packed headers " |
| "attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| |
| if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| if (ctx->desired_packed_headers) { |
| av_log(avctx, AV_LOG_WARNING, "Driver does not support any " |
| "packed headers (wanted %#x).\n", |
| ctx->desired_packed_headers); |
| } else { |
| av_log(avctx, AV_LOG_VERBOSE, "Driver does not support any " |
| "packed headers (none wanted).\n"); |
| } |
| ctx->va_packed_headers = 0; |
| } else { |
| if (ctx->desired_packed_headers & ~attr.value) { |
| av_log(avctx, AV_LOG_WARNING, "Driver does not support some " |
| "wanted packed headers (wanted %#x, found %#x).\n", |
| ctx->desired_packed_headers, attr.value); |
| } else { |
| av_log(avctx, AV_LOG_VERBOSE, "All wanted packed headers " |
| "available (wanted %#x, found %#x).\n", |
| ctx->desired_packed_headers, attr.value); |
| } |
| ctx->va_packed_headers = ctx->desired_packed_headers & attr.value; |
| } |
| |
| if (ctx->va_packed_headers) { |
| ctx->config_attributes[ctx->nb_config_attributes++] = |
| (VAConfigAttrib) { |
| .type = VAConfigAttribEncPackedHeaders, |
| .value = ctx->va_packed_headers, |
| }; |
| } |
| |
| if ( (ctx->desired_packed_headers & VA_ENC_PACKED_HEADER_SEQUENCE) && |
| !(ctx->va_packed_headers & VA_ENC_PACKED_HEADER_SEQUENCE) && |
| (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER)) { |
| av_log(avctx, AV_LOG_WARNING, "Driver does not support packed " |
| "sequence headers, but a global header is requested.\n"); |
| av_log(avctx, AV_LOG_WARNING, "No global header will be written: " |
| "this may result in a stream which is not usable for some " |
| "purposes (e.g. not muxable to some containers).\n"); |
| } |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_quality(AVCodecContext *avctx) |
| { |
| #if VA_CHECK_VERSION(0, 36, 0) |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VAConfigAttrib attr = { VAConfigAttribEncQualityRange }; |
| int quality = avctx->compression_level; |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, |
| ctx->va_entrypoint, |
| &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query quality " |
| "config attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| |
| if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| if (quality != 0) { |
| av_log(avctx, AV_LOG_WARNING, "Quality attribute is not " |
| "supported: will use default quality level.\n"); |
| } |
| } else { |
| if (quality > attr.value) { |
| av_log(avctx, AV_LOG_WARNING, "Invalid quality level: " |
| "valid range is 0-%d, using %d.\n", |
| attr.value, attr.value); |
| quality = attr.value; |
| } |
| |
| ctx->quality_params = (VAEncMiscParameterBufferQualityLevel) { |
| .quality_level = quality, |
| }; |
| vaapi_encode_add_global_param(avctx, |
| VAEncMiscParameterTypeQualityLevel, |
| &ctx->quality_params, |
| sizeof(ctx->quality_params)); |
| } |
| #else |
| av_log(avctx, AV_LOG_WARNING, "The encode quality option is " |
| "not supported with this VAAPI version.\n"); |
| #endif |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_init_roi(AVCodecContext *avctx) |
| { |
| #if VA_CHECK_VERSION(1, 0, 0) |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAStatus vas; |
| VAConfigAttrib attr = { VAConfigAttribEncROI }; |
| |
| vas = vaGetConfigAttributes(ctx->hwctx->display, |
| ctx->va_profile, |
| ctx->va_entrypoint, |
| &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query ROI " |
| "config attribute: %d (%s).\n", vas, vaErrorStr(vas)); |
| return AVERROR_EXTERNAL; |
| } |
| |
| if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| ctx->roi_allowed = 0; |
| } else { |
| VAConfigAttribValEncROI roi = { |
| .value = attr.value, |
| }; |
| |
| ctx->roi_max_regions = roi.bits.num_roi_regions; |
| ctx->roi_allowed = ctx->roi_max_regions > 0 && |
| (ctx->va_rc_mode == VA_RC_CQP || |
| roi.bits.roi_rc_qp_delta_support); |
| } |
| #endif |
| return 0; |
| } |
| |
| static void vaapi_encode_free_output_buffer(void *opaque, |
| uint8_t *data) |
| { |
| AVCodecContext *avctx = opaque; |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VABufferID buffer_id; |
| |
| buffer_id = (VABufferID)(uintptr_t)data; |
| |
| vaDestroyBuffer(ctx->hwctx->display, buffer_id); |
| |
| av_log(avctx, AV_LOG_DEBUG, "Freed output buffer %#x\n", buffer_id); |
| } |
| |
| static AVBufferRef *vaapi_encode_alloc_output_buffer(void *opaque, |
| int size) |
| { |
| AVCodecContext *avctx = opaque; |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VABufferID buffer_id; |
| VAStatus vas; |
| AVBufferRef *ref; |
| |
| // The output buffer size is fixed, so it needs to be large enough |
| // to hold the largest possible compressed frame. We assume here |
| // that the uncompressed frame plus some header data is an upper |
| // bound on that. |
| vas = vaCreateBuffer(ctx->hwctx->display, ctx->va_context, |
| VAEncCodedBufferType, |
| 3 * ctx->surface_width * ctx->surface_height + |
| (1 << 16), 1, 0, &buffer_id); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create bitstream " |
| "output buffer: %d (%s).\n", vas, vaErrorStr(vas)); |
| return NULL; |
| } |
| |
| av_log(avctx, AV_LOG_DEBUG, "Allocated output buffer %#x\n", buffer_id); |
| |
| ref = av_buffer_create((uint8_t*)(uintptr_t)buffer_id, |
| sizeof(buffer_id), |
| &vaapi_encode_free_output_buffer, |
| avctx, AV_BUFFER_FLAG_READONLY); |
| if (!ref) { |
| vaDestroyBuffer(ctx->hwctx->display, buffer_id); |
| return NULL; |
| } |
| |
| return ref; |
| } |
| |
| static av_cold int vaapi_encode_create_recon_frames(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| AVVAAPIHWConfig *hwconfig = NULL; |
| AVHWFramesConstraints *constraints = NULL; |
| enum AVPixelFormat recon_format; |
| int err, i; |
| |
| hwconfig = av_hwdevice_hwconfig_alloc(ctx->device_ref); |
| if (!hwconfig) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| hwconfig->config_id = ctx->va_config; |
| |
| constraints = av_hwdevice_get_hwframe_constraints(ctx->device_ref, |
| hwconfig); |
| if (!constraints) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| |
| // Probably we can use the input surface format as the surface format |
| // of the reconstructed frames. If not, we just pick the first (only?) |
| // format in the valid list and hope that it all works. |
| recon_format = AV_PIX_FMT_NONE; |
| if (constraints->valid_sw_formats) { |
| for (i = 0; constraints->valid_sw_formats[i] != AV_PIX_FMT_NONE; i++) { |
| if (ctx->input_frames->sw_format == |
| constraints->valid_sw_formats[i]) { |
| recon_format = ctx->input_frames->sw_format; |
| break; |
| } |
| } |
| if (recon_format == AV_PIX_FMT_NONE) { |
| // No match. Just use the first in the supported list and |
| // hope for the best. |
| recon_format = constraints->valid_sw_formats[0]; |
| } |
| } else { |
| // No idea what to use; copy input format. |
| recon_format = ctx->input_frames->sw_format; |
| } |
| av_log(avctx, AV_LOG_DEBUG, "Using %s as format of " |
| "reconstructed frames.\n", av_get_pix_fmt_name(recon_format)); |
| |
| if (ctx->surface_width < constraints->min_width || |
| ctx->surface_height < constraints->min_height || |
| ctx->surface_width > constraints->max_width || |
| ctx->surface_height > constraints->max_height) { |
| av_log(avctx, AV_LOG_ERROR, "Hardware does not support encoding at " |
| "size %dx%d (constraints: width %d-%d height %d-%d).\n", |
| ctx->surface_width, ctx->surface_height, |
| constraints->min_width, constraints->max_width, |
| constraints->min_height, constraints->max_height); |
| err = AVERROR(EINVAL); |
| goto fail; |
| } |
| |
| av_freep(&hwconfig); |
| av_hwframe_constraints_free(&constraints); |
| |
| ctx->recon_frames_ref = av_hwframe_ctx_alloc(ctx->device_ref); |
| if (!ctx->recon_frames_ref) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| ctx->recon_frames = (AVHWFramesContext*)ctx->recon_frames_ref->data; |
| |
| ctx->recon_frames->format = AV_PIX_FMT_VAAPI; |
| ctx->recon_frames->sw_format = recon_format; |
| ctx->recon_frames->width = ctx->surface_width; |
| ctx->recon_frames->height = ctx->surface_height; |
| |
| err = av_hwframe_ctx_init(ctx->recon_frames_ref); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to initialise reconstructed " |
| "frame context: %d.\n", err); |
| goto fail; |
| } |
| |
| err = 0; |
| fail: |
| av_freep(&hwconfig); |
| av_hwframe_constraints_free(&constraints); |
| return err; |
| } |
| |
| av_cold int ff_vaapi_encode_init(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| AVVAAPIFramesContext *recon_hwctx = NULL; |
| VAStatus vas; |
| int err; |
| |
| ctx->frame = av_frame_alloc(); |
| if (!ctx->frame) { |
| return AVERROR(ENOMEM); |
| } |
| |
| if (!avctx->hw_frames_ctx) { |
| av_log(avctx, AV_LOG_ERROR, "A hardware frames reference is " |
| "required to associate the encoding device.\n"); |
| return AVERROR(EINVAL); |
| } |
| |
| ctx->va_config = VA_INVALID_ID; |
| ctx->va_context = VA_INVALID_ID; |
| |
| ctx->input_frames_ref = av_buffer_ref(avctx->hw_frames_ctx); |
| if (!ctx->input_frames_ref) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| ctx->input_frames = (AVHWFramesContext*)ctx->input_frames_ref->data; |
| |
| ctx->device_ref = av_buffer_ref(ctx->input_frames->device_ref); |
| if (!ctx->device_ref) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| ctx->device = (AVHWDeviceContext*)ctx->device_ref->data; |
| ctx->hwctx = ctx->device->hwctx; |
| |
| err = vaapi_encode_profile_entrypoint(avctx); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_init_rate_control(avctx); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_init_gop_structure(avctx); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_init_slice_structure(avctx); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_init_packed_headers(avctx); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_init_roi(avctx); |
| if (err < 0) |
| goto fail; |
| |
| if (avctx->compression_level >= 0) { |
| err = vaapi_encode_init_quality(avctx); |
| if (err < 0) |
| goto fail; |
| } |
| |
| vas = vaCreateConfig(ctx->hwctx->display, |
| ctx->va_profile, ctx->va_entrypoint, |
| ctx->config_attributes, ctx->nb_config_attributes, |
| &ctx->va_config); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create encode pipeline " |
| "configuration: %d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail; |
| } |
| |
| err = vaapi_encode_create_recon_frames(avctx); |
| if (err < 0) |
| goto fail; |
| |
| recon_hwctx = ctx->recon_frames->hwctx; |
| vas = vaCreateContext(ctx->hwctx->display, ctx->va_config, |
| ctx->surface_width, ctx->surface_height, |
| VA_PROGRESSIVE, |
| recon_hwctx->surface_ids, |
| recon_hwctx->nb_surfaces, |
| &ctx->va_context); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to create encode pipeline " |
| "context: %d (%s).\n", vas, vaErrorStr(vas)); |
| err = AVERROR(EIO); |
| goto fail; |
| } |
| |
| ctx->output_buffer_pool = |
| av_buffer_pool_init2(sizeof(VABufferID), avctx, |
| &vaapi_encode_alloc_output_buffer, NULL); |
| if (!ctx->output_buffer_pool) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| |
| if (ctx->codec->configure) { |
| err = ctx->codec->configure(avctx); |
| if (err < 0) |
| goto fail; |
| } |
| |
| ctx->output_delay = ctx->b_per_p; |
| ctx->decode_delay = ctx->max_b_depth; |
| |
| if (ctx->codec->sequence_params_size > 0) { |
| ctx->codec_sequence_params = |
| av_mallocz(ctx->codec->sequence_params_size); |
| if (!ctx->codec_sequence_params) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| } |
| if (ctx->codec->picture_params_size > 0) { |
| ctx->codec_picture_params = |
| av_mallocz(ctx->codec->picture_params_size); |
| if (!ctx->codec_picture_params) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| } |
| |
| if (ctx->codec->init_sequence_params) { |
| err = ctx->codec->init_sequence_params(avctx); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Codec sequence initialisation " |
| "failed: %d.\n", err); |
| goto fail; |
| } |
| } |
| |
| if (ctx->va_packed_headers & VA_ENC_PACKED_HEADER_SEQUENCE && |
| ctx->codec->write_sequence_header && |
| avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) { |
| char data[MAX_PARAM_BUFFER_SIZE]; |
| size_t bit_len = 8 * sizeof(data); |
| |
| err = ctx->codec->write_sequence_header(avctx, data, &bit_len); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write sequence header " |
| "for extradata: %d.\n", err); |
| goto fail; |
| } else { |
| avctx->extradata_size = (bit_len + 7) / 8; |
| avctx->extradata = av_mallocz(avctx->extradata_size + |
| AV_INPUT_BUFFER_PADDING_SIZE); |
| if (!avctx->extradata) { |
| err = AVERROR(ENOMEM); |
| goto fail; |
| } |
| memcpy(avctx->extradata, data, avctx->extradata_size); |
| } |
| } |
| |
| return 0; |
| |
| fail: |
| return err; |
| } |
| |
| av_cold int ff_vaapi_encode_close(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodePicture *pic, *next; |
| |
| for (pic = ctx->pic_start; pic; pic = next) { |
| next = pic->next; |
| vaapi_encode_free(avctx, pic); |
| } |
| |
| av_buffer_pool_uninit(&ctx->output_buffer_pool); |
| |
| if (ctx->va_context != VA_INVALID_ID) { |
| vaDestroyContext(ctx->hwctx->display, ctx->va_context); |
| ctx->va_context = VA_INVALID_ID; |
| } |
| |
| if (ctx->va_config != VA_INVALID_ID) { |
| vaDestroyConfig(ctx->hwctx->display, ctx->va_config); |
| ctx->va_config = VA_INVALID_ID; |
| } |
| |
| av_frame_free(&ctx->frame); |
| |
| av_freep(&ctx->codec_sequence_params); |
| av_freep(&ctx->codec_picture_params); |
| |
| av_buffer_unref(&ctx->recon_frames_ref); |
| av_buffer_unref(&ctx->input_frames_ref); |
| av_buffer_unref(&ctx->device_ref); |
| |
| return 0; |
| } |