| /* |
| * This file is part of FFmpeg. |
| * |
| * FFmpeg is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU Lesser General Public |
| * License as published by the Free Software Foundation; either |
| * version 2.1 of the License, or (at your option) any later version. |
| * |
| * FFmpeg is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * Lesser General Public License for more details. |
| * |
| * You should have received a copy of the GNU Lesser General Public |
| * License along with FFmpeg; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA |
| */ |
| |
| #include <string.h> |
| |
| #include <va/va.h> |
| #include <va/va_enc_hevc.h> |
| |
| #include "libavutil/avassert.h" |
| #include "libavutil/common.h" |
| #include "libavutil/mem.h" |
| #include "libavutil/pixdesc.h" |
| #include "libavutil/opt.h" |
| #include "libavutil/mastering_display_metadata.h" |
| |
| #include "atsc_a53.h" |
| #include "avcodec.h" |
| #include "cbs.h" |
| #include "cbs_h265.h" |
| #include "hw_base_encode_h265.h" |
| #include "codec_internal.h" |
| #include "h2645data.h" |
| #include "h265_profile_level.h" |
| #include "vaapi_encode.h" |
| |
| #include "hevc/hevc.h" |
| |
| enum { |
| SEI_MASTERING_DISPLAY = 0x08, |
| SEI_CONTENT_LIGHT_LEVEL = 0x10, |
| SEI_A53_CC = 0x20, |
| }; |
| |
| typedef struct VAAPIEncodeH265Picture { |
| int pic_order_cnt; |
| |
| int64_t last_idr_frame; |
| |
| int slice_nal_unit; |
| int slice_type; |
| int pic_type; |
| } VAAPIEncodeH265Picture; |
| |
| typedef struct VAAPIEncodeH265Context { |
| VAAPIEncodeContext common; |
| |
| // Encoder features. |
| uint32_t va_features; |
| // Block size info. |
| uint32_t va_bs; |
| uint32_t ctu_size; |
| uint32_t min_cb_size; |
| |
| // User options. |
| int qp; |
| int aud; |
| int profile; |
| int level; |
| int sei; |
| |
| // Derived settings. |
| int fixed_qp_p; |
| int fixed_qp_b; |
| |
| // Writer structures. |
| FFHWBaseEncodeH265 units; |
| FFHWBaseEncodeH265Opts unit_opts; |
| H265RawAUD raw_aud; |
| H265RawSlice raw_slice; |
| |
| SEIRawMasteringDisplayColourVolume sei_mastering_display; |
| SEIRawContentLightLevelInfo sei_content_light_level; |
| SEIRawUserDataRegistered sei_a53cc; |
| void *sei_a53cc_data; |
| |
| CodedBitstreamContext *cbc; |
| CodedBitstreamFragment current_access_unit; |
| int aud_needed; |
| int sei_needed; |
| } VAAPIEncodeH265Context; |
| |
| |
| static int vaapi_encode_h265_write_access_unit(AVCodecContext *avctx, |
| char *data, size_t *data_len, |
| CodedBitstreamFragment *au) |
| { |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| int err; |
| |
| err = ff_cbs_write_fragment_data(priv->cbc, au); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to write packed header.\n"); |
| return err; |
| } |
| |
| if (*data_len < 8 * au->data_size - au->data_bit_padding) { |
| av_log(avctx, AV_LOG_ERROR, "Access unit too large: " |
| "%zu < %zu.\n", *data_len, |
| 8 * au->data_size - au->data_bit_padding); |
| return AVERROR(ENOSPC); |
| } |
| |
| memcpy(data, au->data, au->data_size); |
| *data_len = 8 * au->data_size - au->data_bit_padding; |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_h265_add_nal(AVCodecContext *avctx, |
| CodedBitstreamFragment *au, |
| void *nal_unit) |
| { |
| H265RawNALUnitHeader *header = nal_unit; |
| int err; |
| |
| err = ff_cbs_insert_unit_content(au, -1, |
| header->nal_unit_type, nal_unit, NULL); |
| if (err < 0) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to add NAL unit: " |
| "type = %d.\n", header->nal_unit_type); |
| return err; |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_h265_write_sequence_header(AVCodecContext *avctx, |
| char *data, size_t *data_len) |
| { |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| CodedBitstreamFragment *au = &priv->current_access_unit; |
| int err; |
| |
| if (priv->aud_needed) { |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_aud); |
| if (err < 0) |
| goto fail; |
| priv->aud_needed = 0; |
| } |
| |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->units.raw_vps); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->units.raw_sps); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->units.raw_pps); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au); |
| fail: |
| ff_cbs_fragment_reset(au); |
| return err; |
| } |
| |
| static int vaapi_encode_h265_write_slice_header(AVCodecContext *avctx, |
| VAAPIEncodePicture *pic, |
| VAAPIEncodeSlice *slice, |
| char *data, size_t *data_len) |
| { |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| CodedBitstreamFragment *au = &priv->current_access_unit; |
| int err; |
| |
| if (priv->aud_needed) { |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_aud); |
| if (err < 0) |
| goto fail; |
| priv->aud_needed = 0; |
| } |
| |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->raw_slice); |
| if (err < 0) |
| goto fail; |
| |
| err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au); |
| fail: |
| ff_cbs_fragment_reset(au); |
| return err; |
| } |
| |
| static int vaapi_encode_h265_write_extra_header(AVCodecContext *avctx, |
| FFHWBaseEncodePicture *base, |
| int index, int *type, |
| char *data, size_t *data_len) |
| { |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| CodedBitstreamFragment *au = &priv->current_access_unit; |
| int err; |
| |
| if (priv->sei_needed) { |
| if (priv->aud_needed) { |
| err = vaapi_encode_h265_add_nal(avctx, au, &priv->aud); |
| if (err < 0) |
| goto fail; |
| priv->aud_needed = 0; |
| } |
| |
| if (priv->sei_needed & SEI_MASTERING_DISPLAY) { |
| err = ff_cbs_sei_add_message(priv->cbc, au, 1, |
| SEI_TYPE_MASTERING_DISPLAY_COLOUR_VOLUME, |
| &priv->sei_mastering_display, NULL); |
| if (err < 0) |
| goto fail; |
| } |
| |
| if (priv->sei_needed & SEI_CONTENT_LIGHT_LEVEL) { |
| err = ff_cbs_sei_add_message(priv->cbc, au, 1, |
| SEI_TYPE_CONTENT_LIGHT_LEVEL_INFO, |
| &priv->sei_content_light_level, NULL); |
| if (err < 0) |
| goto fail; |
| } |
| if (priv->sei_needed & SEI_A53_CC) { |
| err = ff_cbs_sei_add_message(priv->cbc, au, 1, |
| SEI_TYPE_USER_DATA_REGISTERED_ITU_T_T35, |
| &priv->sei_a53cc, NULL); |
| if (err < 0) |
| goto fail; |
| } |
| |
| priv->sei_needed = 0; |
| |
| err = vaapi_encode_h265_write_access_unit(avctx, data, data_len, au); |
| if (err < 0) |
| goto fail; |
| |
| ff_cbs_fragment_reset(au); |
| |
| *type = VAEncPackedHeaderRawData; |
| return 0; |
| } else { |
| return AVERROR_EOF; |
| } |
| |
| fail: |
| ff_cbs_fragment_reset(au); |
| return err; |
| } |
| |
| static int vaapi_encode_h265_init_sequence_params(AVCodecContext *avctx) |
| { |
| FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| H265RawVPS *vps = &priv->units.raw_vps; |
| H265RawSPS *sps = &priv->units.raw_sps; |
| H265RawPPS *pps = &priv->units.raw_pps; |
| VAEncSequenceParameterBufferHEVC *vseq = ctx->codec_sequence_params; |
| VAEncPictureParameterBufferHEVC *vpic = ctx->codec_picture_params; |
| int i, err; |
| |
| // priv->unit_opts.tier already set |
| // priv->unit_opts.fixed_qp_idr already set |
| priv->unit_opts.cu_qp_delta_enabled_flag = (ctx->va_rc_mode != VA_RC_CQP); |
| priv->unit_opts.tile_rows = ctx->tile_rows; |
| priv->unit_opts.tile_cols = ctx->tile_cols; |
| priv->unit_opts.nb_slices = ctx->nb_slices; |
| priv->unit_opts.slice_block_rows = ctx->slice_block_rows; |
| priv->unit_opts.slice_block_cols = ctx->slice_block_cols; |
| memcpy(priv->unit_opts.col_width, ctx->col_width, |
| ctx->tile_rows*sizeof(*priv->unit_opts.col_width)); |
| memcpy(priv->unit_opts.row_height, ctx->row_height, |
| ctx->tile_cols*sizeof(*priv->unit_opts.row_height)); |
| |
| err = ff_hw_base_encode_init_params_h265(base_ctx, avctx, |
| &priv->units, &priv->unit_opts); |
| if (err < 0) |
| return err; |
| |
| #if VA_CHECK_VERSION(1, 13, 0) |
| // update sps setting according to queried result |
| if (priv->va_features) { |
| VAConfigAttribValEncHEVCFeatures features = { .value = priv->va_features }; |
| |
| // Enable feature if get queried result is VA_FEATURE_SUPPORTED | VA_FEATURE_REQUIRED |
| sps->amp_enabled_flag = |
| !!features.bits.amp; |
| sps->sample_adaptive_offset_enabled_flag = |
| !!features.bits.sao; |
| sps->sps_temporal_mvp_enabled_flag = |
| !!features.bits.temporal_mvp; |
| sps->pcm_enabled_flag = |
| !!features.bits.pcm; |
| } |
| |
| if (priv->va_bs) { |
| VAConfigAttribValEncHEVCBlockSizes bs = { .value = priv->va_bs }; |
| sps->log2_min_luma_coding_block_size_minus3 = |
| ff_ctz(priv->min_cb_size) - 3; |
| sps->log2_diff_max_min_luma_coding_block_size = |
| ff_ctz(priv->ctu_size) - ff_ctz(priv->min_cb_size); |
| |
| sps->log2_min_luma_transform_block_size_minus2 = |
| bs.bits.log2_min_luma_transform_block_size_minus2; |
| sps->log2_diff_max_min_luma_transform_block_size = |
| bs.bits.log2_max_luma_transform_block_size_minus2 - |
| bs.bits.log2_min_luma_transform_block_size_minus2; |
| |
| sps->max_transform_hierarchy_depth_inter = |
| bs.bits.max_max_transform_hierarchy_depth_inter; |
| sps->max_transform_hierarchy_depth_intra = |
| bs.bits.max_max_transform_hierarchy_depth_intra; |
| } |
| |
| // update pps setting according to queried result |
| if (priv->va_features) { |
| VAConfigAttribValEncHEVCFeatures features = { .value = priv->va_features }; |
| if (ctx->va_rc_mode != VA_RC_CQP) |
| pps->cu_qp_delta_enabled_flag = |
| !!features.bits.cu_qp_delta; |
| |
| pps->transform_skip_enabled_flag = |
| !!features.bits.transform_skip; |
| // set diff_cu_qp_delta_depth as its max value if cu_qp_delta enabled. Otherwise |
| // 0 will make cu_qp_delta invalid. |
| if (pps->cu_qp_delta_enabled_flag) |
| pps->diff_cu_qp_delta_depth = sps->log2_diff_max_min_luma_coding_block_size; |
| } |
| #endif |
| |
| // Fill VAAPI parameter buffers. |
| |
| *vseq = (VAEncSequenceParameterBufferHEVC) { |
| .general_profile_idc = vps->profile_tier_level.general_profile_idc, |
| .general_level_idc = vps->profile_tier_level.general_level_idc, |
| .general_tier_flag = vps->profile_tier_level.general_tier_flag, |
| |
| .intra_period = base_ctx->gop_size, |
| .intra_idr_period = base_ctx->gop_size, |
| .ip_period = base_ctx->b_per_p + 1, |
| .bits_per_second = ctx->va_bit_rate, |
| |
| .pic_width_in_luma_samples = sps->pic_width_in_luma_samples, |
| .pic_height_in_luma_samples = sps->pic_height_in_luma_samples, |
| |
| .seq_fields.bits = { |
| .chroma_format_idc = sps->chroma_format_idc, |
| .separate_colour_plane_flag = sps->separate_colour_plane_flag, |
| .bit_depth_luma_minus8 = sps->bit_depth_luma_minus8, |
| .bit_depth_chroma_minus8 = sps->bit_depth_chroma_minus8, |
| .scaling_list_enabled_flag = sps->scaling_list_enabled_flag, |
| .strong_intra_smoothing_enabled_flag = |
| sps->strong_intra_smoothing_enabled_flag, |
| .amp_enabled_flag = sps->amp_enabled_flag, |
| .sample_adaptive_offset_enabled_flag = |
| sps->sample_adaptive_offset_enabled_flag, |
| .pcm_enabled_flag = sps->pcm_enabled_flag, |
| .pcm_loop_filter_disabled_flag = sps->pcm_loop_filter_disabled_flag, |
| .sps_temporal_mvp_enabled_flag = sps->sps_temporal_mvp_enabled_flag, |
| }, |
| |
| .log2_min_luma_coding_block_size_minus3 = |
| sps->log2_min_luma_coding_block_size_minus3, |
| .log2_diff_max_min_luma_coding_block_size = |
| sps->log2_diff_max_min_luma_coding_block_size, |
| .log2_min_transform_block_size_minus2 = |
| sps->log2_min_luma_transform_block_size_minus2, |
| .log2_diff_max_min_transform_block_size = |
| sps->log2_diff_max_min_luma_transform_block_size, |
| .max_transform_hierarchy_depth_inter = |
| sps->max_transform_hierarchy_depth_inter, |
| .max_transform_hierarchy_depth_intra = |
| sps->max_transform_hierarchy_depth_intra, |
| |
| .pcm_sample_bit_depth_luma_minus1 = |
| sps->pcm_sample_bit_depth_luma_minus1, |
| .pcm_sample_bit_depth_chroma_minus1 = |
| sps->pcm_sample_bit_depth_chroma_minus1, |
| .log2_min_pcm_luma_coding_block_size_minus3 = |
| sps->log2_min_pcm_luma_coding_block_size_minus3, |
| .log2_max_pcm_luma_coding_block_size_minus3 = |
| sps->log2_min_pcm_luma_coding_block_size_minus3 + |
| sps->log2_diff_max_min_pcm_luma_coding_block_size, |
| |
| .vui_parameters_present_flag = 0, |
| }; |
| |
| *vpic = (VAEncPictureParameterBufferHEVC) { |
| .decoded_curr_pic = { |
| .picture_id = VA_INVALID_ID, |
| .flags = VA_PICTURE_HEVC_INVALID, |
| }, |
| |
| .coded_buf = VA_INVALID_ID, |
| |
| .collocated_ref_pic_index = sps->sps_temporal_mvp_enabled_flag ? |
| 0 : 0xff, |
| .last_picture = 0, |
| |
| .pic_init_qp = pps->init_qp_minus26 + 26, |
| .diff_cu_qp_delta_depth = pps->diff_cu_qp_delta_depth, |
| .pps_cb_qp_offset = pps->pps_cb_qp_offset, |
| .pps_cr_qp_offset = pps->pps_cr_qp_offset, |
| |
| .num_tile_columns_minus1 = pps->num_tile_columns_minus1, |
| .num_tile_rows_minus1 = pps->num_tile_rows_minus1, |
| |
| .log2_parallel_merge_level_minus2 = pps->log2_parallel_merge_level_minus2, |
| .ctu_max_bitsize_allowed = 0, |
| |
| .num_ref_idx_l0_default_active_minus1 = |
| pps->num_ref_idx_l0_default_active_minus1, |
| .num_ref_idx_l1_default_active_minus1 = |
| pps->num_ref_idx_l1_default_active_minus1, |
| |
| .slice_pic_parameter_set_id = pps->pps_pic_parameter_set_id, |
| |
| .pic_fields.bits = { |
| .sign_data_hiding_enabled_flag = pps->sign_data_hiding_enabled_flag, |
| .constrained_intra_pred_flag = pps->constrained_intra_pred_flag, |
| .transform_skip_enabled_flag = pps->transform_skip_enabled_flag, |
| .cu_qp_delta_enabled_flag = pps->cu_qp_delta_enabled_flag, |
| .weighted_pred_flag = pps->weighted_pred_flag, |
| .weighted_bipred_flag = pps->weighted_bipred_flag, |
| .transquant_bypass_enabled_flag = pps->transquant_bypass_enabled_flag, |
| .tiles_enabled_flag = pps->tiles_enabled_flag, |
| .entropy_coding_sync_enabled_flag = pps->entropy_coding_sync_enabled_flag, |
| .loop_filter_across_tiles_enabled_flag = |
| pps->loop_filter_across_tiles_enabled_flag, |
| .pps_loop_filter_across_slices_enabled_flag = |
| pps->pps_loop_filter_across_slices_enabled_flag, |
| .scaling_list_data_present_flag = (sps->sps_scaling_list_data_present_flag | |
| pps->pps_scaling_list_data_present_flag), |
| .screen_content_flag = 0, |
| .enable_gpu_weighted_prediction = 0, |
| .no_output_of_prior_pics_flag = 0, |
| }, |
| }; |
| |
| if (pps->tiles_enabled_flag) { |
| for (i = 0; i <= vpic->num_tile_rows_minus1; i++) |
| vpic->row_height_minus1[i] = pps->row_height_minus1[i]; |
| for (i = 0; i <= vpic->num_tile_columns_minus1; i++) |
| vpic->column_width_minus1[i] = pps->column_width_minus1[i]; |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_h265_init_picture_params(AVCodecContext *avctx, |
| FFHWBaseEncodePicture *pic) |
| { |
| FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| VAAPIEncodePicture *vaapi_pic = pic->priv; |
| VAAPIEncodeH265Picture *hpic = pic->codec_priv; |
| FFHWBaseEncodePicture *prev = pic->prev; |
| VAAPIEncodeH265Picture *hprev = prev ? prev->codec_priv : NULL; |
| VAEncPictureParameterBufferHEVC *vpic = vaapi_pic->codec_picture_params; |
| int i, j = 0; |
| |
| if (pic->type == FF_HW_PICTURE_TYPE_IDR) { |
| av_assert0(pic->display_order == pic->encode_order); |
| |
| hpic->last_idr_frame = pic->display_order; |
| |
| hpic->slice_nal_unit = HEVC_NAL_IDR_W_RADL; |
| hpic->slice_type = HEVC_SLICE_I; |
| hpic->pic_type = 0; |
| } else { |
| av_assert0(prev); |
| hpic->last_idr_frame = hprev->last_idr_frame; |
| |
| if (pic->type == FF_HW_PICTURE_TYPE_I) { |
| hpic->slice_nal_unit = HEVC_NAL_CRA_NUT; |
| hpic->slice_type = HEVC_SLICE_I; |
| hpic->pic_type = 0; |
| } else if (pic->type == FF_HW_PICTURE_TYPE_P) { |
| av_assert0(pic->refs[0]); |
| hpic->slice_nal_unit = HEVC_NAL_TRAIL_R; |
| hpic->slice_type = HEVC_SLICE_P; |
| hpic->pic_type = 1; |
| } else { |
| FFHWBaseEncodePicture *irap_ref; |
| av_assert0(pic->refs[0][0] && pic->refs[1][0]); |
| for (irap_ref = pic; irap_ref; irap_ref = irap_ref->refs[1][0]) { |
| if (irap_ref->type == FF_HW_PICTURE_TYPE_I) |
| break; |
| } |
| if (pic->b_depth == base_ctx->max_b_depth) { |
| hpic->slice_nal_unit = irap_ref ? HEVC_NAL_RASL_N |
| : HEVC_NAL_TRAIL_N; |
| } else { |
| hpic->slice_nal_unit = irap_ref ? HEVC_NAL_RASL_R |
| : HEVC_NAL_TRAIL_R; |
| } |
| hpic->slice_type = HEVC_SLICE_B; |
| hpic->pic_type = 2; |
| } |
| } |
| hpic->pic_order_cnt = pic->display_order - hpic->last_idr_frame; |
| |
| if (priv->aud) { |
| priv->aud_needed = 1; |
| priv->raw_aud = (H265RawAUD) { |
| .nal_unit_header = { |
| .nal_unit_type = HEVC_NAL_AUD, |
| .nuh_layer_id = 0, |
| .nuh_temporal_id_plus1 = 1, |
| }, |
| .pic_type = hpic->pic_type, |
| }; |
| } else { |
| priv->aud_needed = 0; |
| } |
| |
| priv->sei_needed = 0; |
| |
| // Only look for the metadata on I/IDR frame on the output. We |
| // may force an IDR frame on the output where the metadata gets |
| // changed on the input frame. |
| if ((priv->sei & SEI_MASTERING_DISPLAY) && |
| (pic->type == FF_HW_PICTURE_TYPE_I || pic->type == FF_HW_PICTURE_TYPE_IDR)) { |
| AVFrameSideData *sd = |
| av_frame_get_side_data(pic->input_image, |
| AV_FRAME_DATA_MASTERING_DISPLAY_METADATA); |
| |
| if (sd) { |
| AVMasteringDisplayMetadata *mdm = |
| (AVMasteringDisplayMetadata *)sd->data; |
| |
| // SEI is needed when both the primaries and luminance are set |
| if (mdm->has_primaries && mdm->has_luminance) { |
| SEIRawMasteringDisplayColourVolume *mdcv = |
| &priv->sei_mastering_display; |
| const int mapping[3] = {1, 2, 0}; |
| const int chroma_den = 50000; |
| const int luma_den = 10000; |
| |
| for (i = 0; i < 3; i++) { |
| const int j = mapping[i]; |
| mdcv->display_primaries_x[i] = |
| FFMIN(lrint(chroma_den * |
| av_q2d(mdm->display_primaries[j][0])), |
| chroma_den); |
| mdcv->display_primaries_y[i] = |
| FFMIN(lrint(chroma_den * |
| av_q2d(mdm->display_primaries[j][1])), |
| chroma_den); |
| } |
| |
| mdcv->white_point_x = |
| FFMIN(lrint(chroma_den * av_q2d(mdm->white_point[0])), |
| chroma_den); |
| mdcv->white_point_y = |
| FFMIN(lrint(chroma_den * av_q2d(mdm->white_point[1])), |
| chroma_den); |
| |
| mdcv->max_display_mastering_luminance = |
| lrint(luma_den * av_q2d(mdm->max_luminance)); |
| mdcv->min_display_mastering_luminance = |
| FFMIN(lrint(luma_den * av_q2d(mdm->min_luminance)), |
| mdcv->max_display_mastering_luminance); |
| |
| priv->sei_needed |= SEI_MASTERING_DISPLAY; |
| } |
| } |
| } |
| |
| if ((priv->sei & SEI_CONTENT_LIGHT_LEVEL) && |
| (pic->type == FF_HW_PICTURE_TYPE_I || pic->type == FF_HW_PICTURE_TYPE_IDR)) { |
| AVFrameSideData *sd = |
| av_frame_get_side_data(pic->input_image, |
| AV_FRAME_DATA_CONTENT_LIGHT_LEVEL); |
| |
| if (sd) { |
| AVContentLightMetadata *clm = |
| (AVContentLightMetadata *)sd->data; |
| SEIRawContentLightLevelInfo *clli = |
| &priv->sei_content_light_level; |
| |
| clli->max_content_light_level = FFMIN(clm->MaxCLL, 65535); |
| clli->max_pic_average_light_level = FFMIN(clm->MaxFALL, 65535); |
| |
| priv->sei_needed |= SEI_CONTENT_LIGHT_LEVEL; |
| } |
| } |
| |
| if (priv->sei & SEI_A53_CC) { |
| int err; |
| size_t sei_a53cc_len; |
| av_freep(&priv->sei_a53cc_data); |
| err = ff_alloc_a53_sei(pic->input_image, 0, &priv->sei_a53cc_data, &sei_a53cc_len); |
| if (err < 0) |
| return err; |
| if (priv->sei_a53cc_data != NULL) { |
| priv->sei_a53cc.itu_t_t35_country_code = 181; |
| priv->sei_a53cc.data = (uint8_t *)priv->sei_a53cc_data + 1; |
| priv->sei_a53cc.data_length = sei_a53cc_len - 1; |
| |
| priv->sei_needed |= SEI_A53_CC; |
| } |
| } |
| |
| vpic->decoded_curr_pic = (VAPictureHEVC) { |
| .picture_id = vaapi_pic->recon_surface, |
| .pic_order_cnt = hpic->pic_order_cnt, |
| .flags = 0, |
| }; |
| |
| for (int k = 0; k < MAX_REFERENCE_LIST_NUM; k++) { |
| for (i = 0; i < pic->nb_refs[k]; i++) { |
| FFHWBaseEncodePicture *ref = pic->refs[k][i]; |
| VAAPIEncodeH265Picture *href; |
| |
| av_assert0(ref && ref->encode_order < pic->encode_order); |
| href = ref->codec_priv; |
| |
| vpic->reference_frames[j++] = (VAPictureHEVC) { |
| .picture_id = ((VAAPIEncodePicture *)ref->priv)->recon_surface, |
| .pic_order_cnt = href->pic_order_cnt, |
| .flags = (ref->display_order < pic->display_order ? |
| VA_PICTURE_HEVC_RPS_ST_CURR_BEFORE : 0) | |
| (ref->display_order > pic->display_order ? |
| VA_PICTURE_HEVC_RPS_ST_CURR_AFTER : 0), |
| }; |
| } |
| } |
| |
| for (; j < FF_ARRAY_ELEMS(vpic->reference_frames); j++) { |
| vpic->reference_frames[j] = (VAPictureHEVC) { |
| .picture_id = VA_INVALID_ID, |
| .flags = VA_PICTURE_HEVC_INVALID, |
| }; |
| } |
| |
| vpic->coded_buf = vaapi_pic->output_buffer; |
| |
| vpic->nal_unit_type = hpic->slice_nal_unit; |
| |
| vpic->pic_fields.bits.reference_pic_flag = pic->is_reference; |
| switch (pic->type) { |
| case FF_HW_PICTURE_TYPE_IDR: |
| vpic->pic_fields.bits.idr_pic_flag = 1; |
| vpic->pic_fields.bits.coding_type = 1; |
| break; |
| case FF_HW_PICTURE_TYPE_I: |
| vpic->pic_fields.bits.idr_pic_flag = 0; |
| vpic->pic_fields.bits.coding_type = 1; |
| break; |
| case FF_HW_PICTURE_TYPE_P: |
| vpic->pic_fields.bits.idr_pic_flag = 0; |
| vpic->pic_fields.bits.coding_type = 2; |
| break; |
| case FF_HW_PICTURE_TYPE_B: |
| vpic->pic_fields.bits.idr_pic_flag = 0; |
| vpic->pic_fields.bits.coding_type = 3; |
| break; |
| default: |
| av_assert0(0 && "invalid picture type"); |
| } |
| |
| return 0; |
| } |
| |
| static int vaapi_encode_h265_init_slice_params(AVCodecContext *avctx, |
| FFHWBaseEncodePicture *pic, |
| VAAPIEncodeSlice *slice) |
| { |
| FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| VAAPIEncodePicture *vaapi_pic = pic->priv; |
| VAAPIEncodeH265Picture *hpic = pic->codec_priv; |
| const H265RawSPS *sps = &priv->units.raw_sps; |
| const H265RawPPS *pps = &priv->units.raw_pps; |
| H265RawSliceHeader *sh = &priv->raw_slice.header; |
| VAEncPictureParameterBufferHEVC *vpic = vaapi_pic->codec_picture_params; |
| VAEncSliceParameterBufferHEVC *vslice = slice->codec_slice_params; |
| int i; |
| |
| sh->nal_unit_header = (H265RawNALUnitHeader) { |
| .nal_unit_type = hpic->slice_nal_unit, |
| .nuh_layer_id = 0, |
| .nuh_temporal_id_plus1 = 1, |
| }; |
| |
| sh->slice_pic_parameter_set_id = pps->pps_pic_parameter_set_id; |
| |
| sh->first_slice_segment_in_pic_flag = slice->index == 0; |
| sh->slice_segment_address = slice->block_start; |
| |
| sh->slice_type = hpic->slice_type; |
| |
| if (sh->slice_type == HEVC_SLICE_P && base_ctx->p_to_gpb) |
| sh->slice_type = HEVC_SLICE_B; |
| |
| sh->slice_pic_order_cnt_lsb = hpic->pic_order_cnt & |
| (1 << (sps->log2_max_pic_order_cnt_lsb_minus4 + 4)) - 1; |
| |
| if (pic->type != FF_HW_PICTURE_TYPE_IDR) { |
| H265RawSTRefPicSet *rps; |
| const VAAPIEncodeH265Picture *strp; |
| int rps_poc[MAX_DPB_SIZE]; |
| int rps_used[MAX_DPB_SIZE]; |
| int i, j, poc, rps_pics; |
| |
| sh->short_term_ref_pic_set_sps_flag = 0; |
| |
| rps = &sh->short_term_ref_pic_set; |
| memset(rps, 0, sizeof(*rps)); |
| |
| rps_pics = 0; |
| for (i = 0; i < MAX_REFERENCE_LIST_NUM; i++) { |
| for (j = 0; j < pic->nb_refs[i]; j++) { |
| strp = pic->refs[i][j]->codec_priv; |
| rps_poc[rps_pics] = strp->pic_order_cnt; |
| rps_used[rps_pics] = 1; |
| ++rps_pics; |
| } |
| } |
| |
| for (i = 0; i < pic->nb_dpb_pics; i++) { |
| if (pic->dpb[i] == pic) |
| continue; |
| |
| for (j = 0; j < pic->nb_refs[0]; j++) { |
| if (pic->dpb[i] == pic->refs[0][j]) |
| break; |
| } |
| if (j < pic->nb_refs[0]) |
| continue; |
| |
| for (j = 0; j < pic->nb_refs[1]; j++) { |
| if (pic->dpb[i] == pic->refs[1][j]) |
| break; |
| } |
| if (j < pic->nb_refs[1]) |
| continue; |
| |
| strp = pic->dpb[i]->codec_priv; |
| rps_poc[rps_pics] = strp->pic_order_cnt; |
| rps_used[rps_pics] = 0; |
| ++rps_pics; |
| } |
| |
| for (i = 1; i < rps_pics; i++) { |
| for (j = i; j > 0; j--) { |
| if (rps_poc[j] > rps_poc[j - 1]) |
| break; |
| av_assert0(rps_poc[j] != rps_poc[j - 1]); |
| FFSWAP(int, rps_poc[j], rps_poc[j - 1]); |
| FFSWAP(int, rps_used[j], rps_used[j - 1]); |
| } |
| } |
| |
| av_log(avctx, AV_LOG_DEBUG, "RPS for POC %d:", |
| hpic->pic_order_cnt); |
| for (i = 0; i < rps_pics; i++) { |
| av_log(avctx, AV_LOG_DEBUG, " (%d,%d)", |
| rps_poc[i], rps_used[i]); |
| } |
| av_log(avctx, AV_LOG_DEBUG, "\n"); |
| |
| for (i = 0; i < rps_pics; i++) { |
| av_assert0(rps_poc[i] != hpic->pic_order_cnt); |
| if (rps_poc[i] > hpic->pic_order_cnt) |
| break; |
| } |
| |
| rps->num_negative_pics = i; |
| poc = hpic->pic_order_cnt; |
| for (j = i - 1; j >= 0; j--) { |
| rps->delta_poc_s0_minus1[i - 1 - j] = poc - rps_poc[j] - 1; |
| rps->used_by_curr_pic_s0_flag[i - 1 - j] = rps_used[j]; |
| poc = rps_poc[j]; |
| } |
| |
| rps->num_positive_pics = rps_pics - i; |
| poc = hpic->pic_order_cnt; |
| for (j = i; j < rps_pics; j++) { |
| rps->delta_poc_s1_minus1[j - i] = rps_poc[j] - poc - 1; |
| rps->used_by_curr_pic_s1_flag[j - i] = rps_used[j]; |
| poc = rps_poc[j]; |
| } |
| |
| sh->num_long_term_sps = 0; |
| sh->num_long_term_pics = 0; |
| |
| // when this flag is not present, it is inerred to 1. |
| sh->collocated_from_l0_flag = 1; |
| sh->slice_temporal_mvp_enabled_flag = |
| sps->sps_temporal_mvp_enabled_flag; |
| if (sh->slice_temporal_mvp_enabled_flag) { |
| if (sh->slice_type == HEVC_SLICE_B) |
| sh->collocated_from_l0_flag = 1; |
| sh->collocated_ref_idx = 0; |
| } |
| |
| sh->num_ref_idx_active_override_flag = 0; |
| sh->num_ref_idx_l0_active_minus1 = pps->num_ref_idx_l0_default_active_minus1; |
| sh->num_ref_idx_l1_active_minus1 = pps->num_ref_idx_l1_default_active_minus1; |
| } |
| |
| sh->slice_sao_luma_flag = sh->slice_sao_chroma_flag = |
| sps->sample_adaptive_offset_enabled_flag; |
| |
| if (pic->type == FF_HW_PICTURE_TYPE_B) |
| sh->slice_qp_delta = priv->fixed_qp_b - (pps->init_qp_minus26 + 26); |
| else if (pic->type == FF_HW_PICTURE_TYPE_P) |
| sh->slice_qp_delta = priv->fixed_qp_p - (pps->init_qp_minus26 + 26); |
| else |
| sh->slice_qp_delta = priv->unit_opts.fixed_qp_idr - (pps->init_qp_minus26 + 26); |
| |
| |
| *vslice = (VAEncSliceParameterBufferHEVC) { |
| .slice_segment_address = sh->slice_segment_address, |
| .num_ctu_in_slice = slice->block_size, |
| |
| .slice_type = sh->slice_type, |
| .slice_pic_parameter_set_id = sh->slice_pic_parameter_set_id, |
| |
| .num_ref_idx_l0_active_minus1 = sh->num_ref_idx_l0_active_minus1, |
| .num_ref_idx_l1_active_minus1 = sh->num_ref_idx_l1_active_minus1, |
| |
| .luma_log2_weight_denom = sh->luma_log2_weight_denom, |
| .delta_chroma_log2_weight_denom = sh->delta_chroma_log2_weight_denom, |
| |
| .max_num_merge_cand = 5 - sh->five_minus_max_num_merge_cand, |
| |
| .slice_qp_delta = sh->slice_qp_delta, |
| .slice_cb_qp_offset = sh->slice_cb_qp_offset, |
| .slice_cr_qp_offset = sh->slice_cr_qp_offset, |
| |
| .slice_beta_offset_div2 = sh->slice_beta_offset_div2, |
| .slice_tc_offset_div2 = sh->slice_tc_offset_div2, |
| |
| .slice_fields.bits = { |
| .last_slice_of_pic_flag = slice->index == vaapi_pic->nb_slices - 1, |
| .dependent_slice_segment_flag = sh->dependent_slice_segment_flag, |
| .colour_plane_id = sh->colour_plane_id, |
| .slice_temporal_mvp_enabled_flag = |
| sh->slice_temporal_mvp_enabled_flag, |
| .slice_sao_luma_flag = sh->slice_sao_luma_flag, |
| .slice_sao_chroma_flag = sh->slice_sao_chroma_flag, |
| .num_ref_idx_active_override_flag = |
| sh->num_ref_idx_active_override_flag, |
| .mvd_l1_zero_flag = sh->mvd_l1_zero_flag, |
| .cabac_init_flag = sh->cabac_init_flag, |
| .slice_deblocking_filter_disabled_flag = |
| sh->slice_deblocking_filter_disabled_flag, |
| .slice_loop_filter_across_slices_enabled_flag = |
| sh->slice_loop_filter_across_slices_enabled_flag, |
| .collocated_from_l0_flag = sh->collocated_from_l0_flag, |
| }, |
| }; |
| |
| for (i = 0; i < FF_ARRAY_ELEMS(vslice->ref_pic_list0); i++) { |
| vslice->ref_pic_list0[i].picture_id = VA_INVALID_ID; |
| vslice->ref_pic_list0[i].flags = VA_PICTURE_HEVC_INVALID; |
| vslice->ref_pic_list1[i].picture_id = VA_INVALID_ID; |
| vslice->ref_pic_list1[i].flags = VA_PICTURE_HEVC_INVALID; |
| } |
| |
| if (pic->nb_refs[0]) { |
| // Backward reference for P- or B-frame. |
| av_assert0(pic->type == FF_HW_PICTURE_TYPE_P || |
| pic->type == FF_HW_PICTURE_TYPE_B); |
| vslice->ref_pic_list0[0] = vpic->reference_frames[0]; |
| if (base_ctx->p_to_gpb && pic->type == FF_HW_PICTURE_TYPE_P) |
| // Reference for GPB B-frame, L0 == L1 |
| vslice->ref_pic_list1[0] = vpic->reference_frames[0]; |
| } |
| if (pic->nb_refs[1]) { |
| // Forward reference for B-frame. |
| av_assert0(pic->type == FF_HW_PICTURE_TYPE_B); |
| vslice->ref_pic_list1[0] = vpic->reference_frames[1]; |
| } |
| |
| if (pic->type == FF_HW_PICTURE_TYPE_P && base_ctx->p_to_gpb) { |
| vslice->slice_type = HEVC_SLICE_B; |
| for (i = 0; i < FF_ARRAY_ELEMS(vslice->ref_pic_list0); i++) { |
| vslice->ref_pic_list1[i].picture_id = vslice->ref_pic_list0[i].picture_id; |
| vslice->ref_pic_list1[i].flags = vslice->ref_pic_list0[i].flags; |
| } |
| } |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_h265_get_encoder_caps(AVCodecContext *avctx) |
| { |
| FFHWBaseEncodeContext *base_ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| |
| #if VA_CHECK_VERSION(1, 13, 0) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAConfigAttribValEncHEVCBlockSizes block_size; |
| VAConfigAttrib attr; |
| VAStatus vas; |
| |
| attr.type = VAConfigAttribEncHEVCFeatures; |
| vas = vaGetConfigAttributes(ctx->hwctx->display, ctx->va_profile, |
| ctx->va_entrypoint, &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query encoder " |
| "features, using guessed defaults.\n"); |
| return AVERROR_EXTERNAL; |
| } else if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_WARNING, "Driver does not advertise " |
| "encoder features, using guessed defaults.\n"); |
| } else { |
| priv->va_features = attr.value; |
| } |
| |
| attr.type = VAConfigAttribEncHEVCBlockSizes; |
| vas = vaGetConfigAttributes(ctx->hwctx->display, ctx->va_profile, |
| ctx->va_entrypoint, &attr, 1); |
| if (vas != VA_STATUS_SUCCESS) { |
| av_log(avctx, AV_LOG_ERROR, "Failed to query encoder " |
| "block size, using guessed defaults.\n"); |
| return AVERROR_EXTERNAL; |
| } else if (attr.value == VA_ATTRIB_NOT_SUPPORTED) { |
| av_log(avctx, AV_LOG_WARNING, "Driver does not advertise " |
| "encoder block size, using guessed defaults.\n"); |
| } else { |
| priv->va_bs = block_size.value = attr.value; |
| |
| priv->ctu_size = |
| 1 << block_size.bits.log2_max_coding_tree_block_size_minus3 + 3; |
| priv->min_cb_size = |
| 1 << block_size.bits.log2_min_luma_coding_block_size_minus3 + 3; |
| } |
| } |
| #endif |
| |
| if (!priv->ctu_size) { |
| priv->ctu_size = 32; |
| priv->min_cb_size = 16; |
| } |
| av_log(avctx, AV_LOG_VERBOSE, "Using CTU size %dx%d, " |
| "min CB size %dx%d.\n", priv->ctu_size, priv->ctu_size, |
| priv->min_cb_size, priv->min_cb_size); |
| |
| base_ctx->surface_width = FFALIGN(avctx->width, |
| FFMAX(priv->min_cb_size, priv->common.surface_alignment_width)); |
| base_ctx->surface_height = FFALIGN(avctx->height, |
| FFMAX(priv->min_cb_size, priv->common.surface_alignment_height)); |
| |
| base_ctx->slice_block_width = base_ctx->slice_block_height = priv->ctu_size; |
| |
| return 0; |
| } |
| |
| static av_cold int vaapi_encode_h265_configure(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| int err; |
| |
| err = ff_cbs_init(&priv->cbc, AV_CODEC_ID_HEVC, avctx); |
| if (err < 0) |
| return err; |
| |
| if (ctx->va_rc_mode == VA_RC_CQP) { |
| // Note that VAAPI only supports positive QP values - the range is |
| // therefore always bounded below by 1, even in 10-bit mode where |
| // it should go down to -12. |
| |
| priv->fixed_qp_p = av_clip(ctx->rc_quality, 1, 51); |
| if (avctx->i_quant_factor > 0.0) |
| priv->unit_opts.fixed_qp_idr = |
| av_clip((avctx->i_quant_factor * priv->fixed_qp_p + |
| avctx->i_quant_offset) + 0.5, 1, 51); |
| else |
| priv->unit_opts.fixed_qp_idr = priv->fixed_qp_p; |
| if (avctx->b_quant_factor > 0.0) |
| priv->fixed_qp_b = |
| av_clip((avctx->b_quant_factor * priv->fixed_qp_p + |
| avctx->b_quant_offset) + 0.5, 1, 51); |
| else |
| priv->fixed_qp_b = priv->fixed_qp_p; |
| |
| av_log(avctx, AV_LOG_DEBUG, "Using fixed QP = " |
| "%d / %d / %d for IDR- / P- / B-frames.\n", |
| priv->unit_opts.fixed_qp_idr, priv->fixed_qp_p, priv->fixed_qp_b); |
| |
| } else { |
| // These still need to be set for init_qp/slice_qp_delta. |
| priv->unit_opts.fixed_qp_idr = 30; |
| priv->fixed_qp_p = 30; |
| priv->fixed_qp_b = 30; |
| } |
| |
| ctx->roi_quant_range = 51 + 6 * (ctx->profile->depth - 8); |
| |
| return 0; |
| } |
| |
| static const VAAPIEncodeProfile vaapi_encode_h265_profiles[] = { |
| { AV_PROFILE_HEVC_MAIN, 8, 3, 1, 1, VAProfileHEVCMain }, |
| { AV_PROFILE_HEVC_REXT, 8, 3, 1, 1, VAProfileHEVCMain }, |
| #if VA_CHECK_VERSION(0, 37, 0) |
| { AV_PROFILE_HEVC_MAIN_10, 10, 3, 1, 1, VAProfileHEVCMain10 }, |
| { AV_PROFILE_HEVC_REXT, 10, 3, 1, 1, VAProfileHEVCMain10 }, |
| #endif |
| #if VA_CHECK_VERSION(1, 2, 0) |
| { AV_PROFILE_HEVC_REXT, 12, 3, 1, 1, VAProfileHEVCMain12 }, |
| { AV_PROFILE_HEVC_REXT, 8, 3, 1, 0, VAProfileHEVCMain422_10 }, |
| { AV_PROFILE_HEVC_REXT, 10, 3, 1, 0, VAProfileHEVCMain422_10 }, |
| { AV_PROFILE_HEVC_REXT, 12, 3, 1, 0, VAProfileHEVCMain422_12 }, |
| { AV_PROFILE_HEVC_REXT, 8, 3, 0, 0, VAProfileHEVCMain444 }, |
| { AV_PROFILE_HEVC_REXT, 10, 3, 0, 0, VAProfileHEVCMain444_10 }, |
| { AV_PROFILE_HEVC_REXT, 12, 3, 0, 0, VAProfileHEVCMain444_12 }, |
| #endif |
| { AV_PROFILE_UNKNOWN } |
| }; |
| |
| static const VAAPIEncodeType vaapi_encode_type_h265 = { |
| .profiles = vaapi_encode_h265_profiles, |
| |
| .flags = FF_HW_FLAG_SLICE_CONTROL | |
| FF_HW_FLAG_B_PICTURES | |
| FF_HW_FLAG_B_PICTURE_REFERENCES | |
| FF_HW_FLAG_NON_IDR_KEY_PICTURES, |
| |
| .default_quality = 25, |
| |
| .get_encoder_caps = &vaapi_encode_h265_get_encoder_caps, |
| .configure = &vaapi_encode_h265_configure, |
| |
| .picture_priv_data_size = sizeof(VAAPIEncodeH265Picture), |
| |
| .sequence_params_size = sizeof(VAEncSequenceParameterBufferHEVC), |
| .init_sequence_params = &vaapi_encode_h265_init_sequence_params, |
| |
| .picture_params_size = sizeof(VAEncPictureParameterBufferHEVC), |
| .init_picture_params = &vaapi_encode_h265_init_picture_params, |
| |
| .slice_params_size = sizeof(VAEncSliceParameterBufferHEVC), |
| .init_slice_params = &vaapi_encode_h265_init_slice_params, |
| |
| .sequence_header_type = VAEncPackedHeaderSequence, |
| .write_sequence_header = &vaapi_encode_h265_write_sequence_header, |
| |
| .slice_header_type = VAEncPackedHeaderHEVC_Slice, |
| .write_slice_header = &vaapi_encode_h265_write_slice_header, |
| |
| .write_extra_header = &vaapi_encode_h265_write_extra_header, |
| }; |
| |
| static av_cold int vaapi_encode_h265_init(AVCodecContext *avctx) |
| { |
| VAAPIEncodeContext *ctx = avctx->priv_data; |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| |
| ctx->codec = &vaapi_encode_type_h265; |
| |
| if (avctx->profile == AV_PROFILE_UNKNOWN) |
| avctx->profile = priv->profile; |
| if (avctx->level == AV_LEVEL_UNKNOWN) |
| avctx->level = priv->level; |
| |
| if (avctx->level != AV_LEVEL_UNKNOWN && avctx->level & ~0xff) { |
| av_log(avctx, AV_LOG_ERROR, "Invalid level %d: must fit " |
| "in 8-bit unsigned integer.\n", avctx->level); |
| return AVERROR(EINVAL); |
| } |
| |
| ctx->desired_packed_headers = |
| VA_ENC_PACKED_HEADER_SEQUENCE | // VPS, SPS and PPS. |
| VA_ENC_PACKED_HEADER_SLICE | // Slice headers. |
| VA_ENC_PACKED_HEADER_MISC; // SEI |
| |
| if (priv->qp > 0) |
| ctx->explicit_qp = priv->qp; |
| |
| return ff_vaapi_encode_init(avctx); |
| } |
| |
| static av_cold int vaapi_encode_h265_close(AVCodecContext *avctx) |
| { |
| VAAPIEncodeH265Context *priv = avctx->priv_data; |
| |
| ff_cbs_fragment_free(&priv->current_access_unit); |
| ff_cbs_close(&priv->cbc); |
| av_freep(&priv->sei_a53cc_data); |
| |
| return ff_vaapi_encode_close(avctx); |
| } |
| |
| #define OFFSET(x) offsetof(VAAPIEncodeH265Context, x) |
| #define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM) |
| static const AVOption vaapi_encode_h265_options[] = { |
| HW_BASE_ENCODE_COMMON_OPTIONS, |
| VAAPI_ENCODE_COMMON_OPTIONS, |
| VAAPI_ENCODE_RC_OPTIONS, |
| |
| { "qp", "Constant QP (for P-frames; scaled by qfactor/qoffset for I/B)", |
| OFFSET(qp), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 52, FLAGS }, |
| |
| { "aud", "Include AUD", |
| OFFSET(aud), AV_OPT_TYPE_BOOL, { .i64 = 0 }, 0, 1, FLAGS }, |
| |
| { "profile", "Set profile (general_profile_idc)", |
| OFFSET(profile), AV_OPT_TYPE_INT, |
| { .i64 = AV_PROFILE_UNKNOWN }, AV_PROFILE_UNKNOWN, 0xff, FLAGS, .unit = "profile" }, |
| |
| #define PROFILE(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \ |
| { .i64 = value }, 0, 0, FLAGS, .unit = "profile" |
| { PROFILE("main", AV_PROFILE_HEVC_MAIN) }, |
| { PROFILE("main10", AV_PROFILE_HEVC_MAIN_10) }, |
| { PROFILE("rext", AV_PROFILE_HEVC_REXT) }, |
| #undef PROFILE |
| |
| { "tier", "Set tier (general_tier_flag)", |
| OFFSET(unit_opts.tier), AV_OPT_TYPE_INT, |
| { .i64 = 0 }, 0, 1, FLAGS, .unit = "tier" }, |
| { "main", NULL, 0, AV_OPT_TYPE_CONST, |
| { .i64 = 0 }, 0, 0, FLAGS, .unit = "tier" }, |
| { "high", NULL, 0, AV_OPT_TYPE_CONST, |
| { .i64 = 1 }, 0, 0, FLAGS, .unit = "tier" }, |
| |
| { "level", "Set level (general_level_idc)", |
| OFFSET(level), AV_OPT_TYPE_INT, |
| { .i64 = AV_LEVEL_UNKNOWN }, AV_LEVEL_UNKNOWN, 0xff, FLAGS, .unit = "level" }, |
| |
| #define LEVEL(name, value) name, NULL, 0, AV_OPT_TYPE_CONST, \ |
| { .i64 = value }, 0, 0, FLAGS, .unit = "level" |
| { LEVEL("1", 30) }, |
| { LEVEL("2", 60) }, |
| { LEVEL("2.1", 63) }, |
| { LEVEL("3", 90) }, |
| { LEVEL("3.1", 93) }, |
| { LEVEL("4", 120) }, |
| { LEVEL("4.1", 123) }, |
| { LEVEL("5", 150) }, |
| { LEVEL("5.1", 153) }, |
| { LEVEL("5.2", 156) }, |
| { LEVEL("6", 180) }, |
| { LEVEL("6.1", 183) }, |
| { LEVEL("6.2", 186) }, |
| #undef LEVEL |
| |
| { "sei", "Set SEI to include", |
| OFFSET(sei), AV_OPT_TYPE_FLAGS, |
| { .i64 = SEI_MASTERING_DISPLAY | SEI_CONTENT_LIGHT_LEVEL | SEI_A53_CC }, |
| 0, INT_MAX, FLAGS, .unit = "sei" }, |
| { "hdr", |
| "Include HDR metadata for mastering display colour volume " |
| "and content light level information", |
| 0, AV_OPT_TYPE_CONST, |
| { .i64 = SEI_MASTERING_DISPLAY | SEI_CONTENT_LIGHT_LEVEL }, |
| INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, |
| { "a53_cc", |
| "Include A/53 caption data", |
| 0, AV_OPT_TYPE_CONST, |
| { .i64 = SEI_A53_CC }, |
| INT_MIN, INT_MAX, FLAGS, .unit = "sei" }, |
| |
| { "tiles", "Tile columns x rows", |
| OFFSET(common.tile_cols), AV_OPT_TYPE_IMAGE_SIZE, |
| { .str = NULL }, 0, 0, FLAGS }, |
| |
| { NULL }, |
| }; |
| |
| static const FFCodecDefault vaapi_encode_h265_defaults[] = { |
| { "b", "0" }, |
| { "bf", "2" }, |
| { "g", "120" }, |
| { "i_qfactor", "1" }, |
| { "i_qoffset", "0" }, |
| { "b_qfactor", "6/5" }, |
| { "b_qoffset", "0" }, |
| { "qmin", "-1" }, |
| { "qmax", "-1" }, |
| { "refs", "0" }, |
| { NULL }, |
| }; |
| |
| static const AVClass vaapi_encode_h265_class = { |
| .class_name = "h265_vaapi", |
| .item_name = av_default_item_name, |
| .option = vaapi_encode_h265_options, |
| .version = LIBAVUTIL_VERSION_INT, |
| }; |
| |
| const FFCodec ff_hevc_vaapi_encoder = { |
| .p.name = "hevc_vaapi", |
| CODEC_LONG_NAME("H.265/HEVC (VAAPI)"), |
| .p.type = AVMEDIA_TYPE_VIDEO, |
| .p.id = AV_CODEC_ID_HEVC, |
| .priv_data_size = sizeof(VAAPIEncodeH265Context), |
| .init = &vaapi_encode_h265_init, |
| FF_CODEC_RECEIVE_PACKET_CB(&ff_vaapi_encode_receive_packet), |
| .close = &vaapi_encode_h265_close, |
| .p.priv_class = &vaapi_encode_h265_class, |
| .p.capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_HARDWARE | |
| AV_CODEC_CAP_DR1 | AV_CODEC_CAP_ENCODER_REORDERED_OPAQUE, |
| .caps_internal = FF_CODEC_CAP_NOT_INIT_THREADSAFE | |
| FF_CODEC_CAP_INIT_CLEANUP, |
| .defaults = vaapi_encode_h265_defaults, |
| CODEC_PIXFMTS(AV_PIX_FMT_VAAPI), |
| .color_ranges = AVCOL_RANGE_MPEG | AVCOL_RANGE_JPEG, |
| .hw_configs = ff_vaapi_encode_hw_configs, |
| .p.wrapper_name = "vaapi", |
| }; |