dxva2_av1.c (22807B)
1 /* 2 * DXVA2 AV1 HW acceleration. 3 * 4 * copyright (c) 2020 Hendrik Leppkes 5 * 6 * This file is part of FFmpeg. 7 * 8 * FFmpeg is free software; you can redistribute it and/or 9 * modify it under the terms of the GNU Lesser General Public 10 * License as published by the Free Software Foundation; either 11 * version 2.1 of the License, or (at your option) any later version. 12 * 13 * FFmpeg is distributed in the hope that it will be useful, 14 * but WITHOUT ANY WARRANTY; without even the implied warranty of 15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU 16 * Lesser General Public License for more details. 17 * 18 * You should have received a copy of the GNU Lesser General Public 19 * License along with FFmpeg; if not, write to the Free Software 20 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA 21 */ 22 23 #include "config_components.h" 24 25 #include "libavutil/avassert.h" 26 #include "libavutil/mem.h" 27 #include "libavutil/pixdesc.h" 28 29 #include "dxva2_internal.h" 30 #include "av1dec.h" 31 #include "hwaccel_internal.h" 32 33 #define MAX_TILES 256 34 35 struct AV1DXVAContext { 36 FFDXVASharedContext shared; 37 38 unsigned int bitstream_allocated; 39 uint8_t *bitstream_cache; 40 }; 41 42 struct av1_dxva2_picture_context { 43 DXVA_PicParams_AV1 pp; 44 unsigned tile_count; 45 DXVA_Tile_AV1 tiles[MAX_TILES]; 46 uint8_t *bitstream; 47 unsigned bitstream_size; 48 }; 49 50 static int get_bit_depth_from_seq(const AV1RawSequenceHeader *seq) 51 { 52 if (seq->seq_profile == 2 && seq->color_config.high_bitdepth) 53 return seq->color_config.twelve_bit ? 12 : 10; 54 else if (seq->seq_profile <= 2 && seq->color_config.high_bitdepth) 55 return 10; 56 else 57 return 8; 58 } 59 60 int ff_dxva2_av1_fill_picture_parameters(const AVCodecContext *avctx, AVDXVAContext *ctx, 61 DXVA_PicParams_AV1 *pp) 62 { 63 int i,j, uses_lr; 64 const AV1DecContext *h = avctx->priv_data; 65 const AV1RawSequenceHeader *seq = h->raw_seq; 66 const AV1RawFrameHeader *frame_header = h->raw_frame_header; 67 const AV1RawFilmGrainParams *film_grain = &h->cur_frame.film_grain; 68 69 unsigned char remap_lr_type[4] = { AV1_RESTORE_NONE, AV1_RESTORE_SWITCHABLE, AV1_RESTORE_WIENER, AV1_RESTORE_SGRPROJ }; 70 int apply_grain = !(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN) && film_grain->apply_grain; 71 72 memset(pp, 0, sizeof(*pp)); 73 74 pp->width = avctx->width; 75 pp->height = avctx->height; 76 77 pp->max_width = seq->max_frame_width_minus_1 + 1; 78 pp->max_height = seq->max_frame_height_minus_1 + 1; 79 80 pp->superres_denom = frame_header->use_superres ? frame_header->coded_denom + AV1_SUPERRES_DENOM_MIN : AV1_SUPERRES_NUM; 81 pp->bitdepth = get_bit_depth_from_seq(seq); 82 pp->seq_profile = seq->seq_profile; 83 84 /* Tiling info */ 85 pp->tiles.cols = frame_header->tile_cols; 86 pp->tiles.rows = frame_header->tile_rows; 87 pp->tiles.context_update_id = frame_header->context_update_tile_id; 88 89 for (i = 0; i < pp->tiles.cols; i++) 90 pp->tiles.widths[i] = frame_header->width_in_sbs_minus_1[i] + 1; 91 92 for (i = 0; i < pp->tiles.rows; i++) 93 pp->tiles.heights[i] = frame_header->height_in_sbs_minus_1[i] + 1; 94 95 /* Coding tools */ 96 pp->coding.use_128x128_superblock = seq->use_128x128_superblock; 97 pp->coding.intra_edge_filter = seq->enable_intra_edge_filter; 98 pp->coding.interintra_compound = seq->enable_interintra_compound; 99 pp->coding.masked_compound = seq->enable_masked_compound; 100 pp->coding.warped_motion = frame_header->allow_warped_motion; 101 pp->coding.dual_filter = seq->enable_dual_filter; 102 pp->coding.jnt_comp = seq->enable_jnt_comp; 103 pp->coding.screen_content_tools = frame_header->allow_screen_content_tools; 104 pp->coding.integer_mv = h->cur_frame.force_integer_mv; 105 pp->coding.cdef = seq->enable_cdef; 106 pp->coding.restoration = seq->enable_restoration; 107 pp->coding.film_grain = seq->film_grain_params_present && !(avctx->export_side_data & AV_CODEC_EXPORT_DATA_FILM_GRAIN); 108 pp->coding.intrabc = frame_header->allow_intrabc; 109 pp->coding.high_precision_mv = frame_header->allow_high_precision_mv; 110 pp->coding.switchable_motion_mode = frame_header->is_motion_mode_switchable; 111 pp->coding.filter_intra = seq->enable_filter_intra; 112 pp->coding.disable_frame_end_update_cdf = frame_header->disable_frame_end_update_cdf; 113 pp->coding.disable_cdf_update = frame_header->disable_cdf_update; 114 pp->coding.reference_mode = frame_header->reference_select; 115 pp->coding.skip_mode = frame_header->skip_mode_present; 116 pp->coding.reduced_tx_set = frame_header->reduced_tx_set; 117 pp->coding.superres = frame_header->use_superres; 118 pp->coding.tx_mode = frame_header->tx_mode; 119 pp->coding.use_ref_frame_mvs = frame_header->use_ref_frame_mvs; 120 pp->coding.enable_ref_frame_mvs = seq->enable_ref_frame_mvs; 121 pp->coding.reference_frame_update = 1; // 0 for show_existing_frame with key frames, but those are not passed to the hwaccel 122 123 /* Format & Picture Info flags */ 124 pp->format.frame_type = frame_header->frame_type; 125 pp->format.show_frame = frame_header->show_frame; 126 pp->format.showable_frame = frame_header->showable_frame; 127 pp->format.subsampling_x = seq->color_config.subsampling_x; 128 pp->format.subsampling_y = seq->color_config.subsampling_y; 129 pp->format.mono_chrome = seq->color_config.mono_chrome; 130 131 /* References */ 132 pp->primary_ref_frame = frame_header->primary_ref_frame; 133 pp->order_hint = frame_header->order_hint; 134 pp->order_hint_bits = seq->enable_order_hint ? seq->order_hint_bits_minus_1 + 1 : 0; 135 136 memset(pp->RefFrameMapTextureIndex, 0xFF, sizeof(pp->RefFrameMapTextureIndex)); 137 for (i = 0; i < AV1_REFS_PER_FRAME; i++) { 138 int8_t ref_idx = frame_header->ref_frame_idx[i]; 139 AVFrame *ref_frame = h->ref[ref_idx].f; 140 141 pp->frame_refs[i].width = ref_frame ? ref_frame->width : 0; 142 pp->frame_refs[i].height = ref_frame ? ref_frame->height : 0; 143 pp->frame_refs[i].Index = ref_frame ? ref_idx : 0xFF; 144 145 /* Global Motion */ 146 pp->frame_refs[i].wminvalid = h->cur_frame.gm_invalid[AV1_REF_FRAME_LAST + i]; 147 pp->frame_refs[i].wmtype = h->cur_frame.gm_type[AV1_REF_FRAME_LAST + i]; 148 for (j = 0; j < 6; ++j) { 149 pp->frame_refs[i].wmmat[j] = h->cur_frame.gm_params[AV1_REF_FRAME_LAST + i][j]; 150 } 151 } 152 for (i = 0; i < AV1_NUM_REF_FRAMES; i++) { 153 AVFrame *ref_frame = h->ref[i].f; 154 if (ref_frame) 155 pp->RefFrameMapTextureIndex[i] = ff_dxva2_get_surface_index(avctx, ctx, ref_frame, 0); 156 } 157 158 pp->CurrPicTextureIndex = ff_dxva2_get_surface_index(avctx, ctx, h->cur_frame.f, 1); 159 160 /* Loop filter parameters */ 161 pp->loop_filter.filter_level[0] = frame_header->loop_filter_level[0]; 162 pp->loop_filter.filter_level[1] = frame_header->loop_filter_level[1]; 163 pp->loop_filter.filter_level_u = frame_header->loop_filter_level[2]; 164 pp->loop_filter.filter_level_v = frame_header->loop_filter_level[3]; 165 pp->loop_filter.sharpness_level = frame_header->loop_filter_sharpness; 166 pp->loop_filter.mode_ref_delta_enabled = frame_header->loop_filter_delta_enabled; 167 pp->loop_filter.mode_ref_delta_update = frame_header->loop_filter_delta_update; 168 pp->loop_filter.delta_lf_multi = frame_header->delta_lf_multi; 169 pp->loop_filter.delta_lf_present = frame_header->delta_lf_present; 170 pp->loop_filter.delta_lf_res = frame_header->delta_lf_res; 171 172 for (i = 0; i < AV1_TOTAL_REFS_PER_FRAME; i++) { 173 pp->loop_filter.ref_deltas[i] = frame_header->loop_filter_ref_deltas[i]; 174 } 175 176 pp->loop_filter.mode_deltas[0] = frame_header->loop_filter_mode_deltas[0]; 177 pp->loop_filter.mode_deltas[1] = frame_header->loop_filter_mode_deltas[1]; 178 pp->loop_filter.frame_restoration_type[0] = remap_lr_type[frame_header->lr_type[0]]; 179 pp->loop_filter.frame_restoration_type[1] = remap_lr_type[frame_header->lr_type[1]]; 180 pp->loop_filter.frame_restoration_type[2] = remap_lr_type[frame_header->lr_type[2]]; 181 uses_lr = frame_header->lr_type[0] || frame_header->lr_type[1] || frame_header->lr_type[2]; 182 pp->loop_filter.log2_restoration_unit_size[0] = uses_lr ? (6 + frame_header->lr_unit_shift) : 8; 183 pp->loop_filter.log2_restoration_unit_size[1] = uses_lr ? (6 + frame_header->lr_unit_shift - frame_header->lr_uv_shift) : 8; 184 pp->loop_filter.log2_restoration_unit_size[2] = uses_lr ? (6 + frame_header->lr_unit_shift - frame_header->lr_uv_shift) : 8; 185 186 /* Quantization */ 187 pp->quantization.delta_q_present = frame_header->delta_q_present; 188 pp->quantization.delta_q_res = frame_header->delta_q_res; 189 pp->quantization.base_qindex = frame_header->base_q_idx; 190 pp->quantization.y_dc_delta_q = frame_header->delta_q_y_dc; 191 pp->quantization.u_dc_delta_q = frame_header->delta_q_u_dc; 192 pp->quantization.v_dc_delta_q = frame_header->delta_q_v_dc; 193 pp->quantization.u_ac_delta_q = frame_header->delta_q_u_ac; 194 pp->quantization.v_ac_delta_q = frame_header->delta_q_v_ac; 195 pp->quantization.qm_y = frame_header->using_qmatrix ? frame_header->qm_y : 0xFF; 196 pp->quantization.qm_u = frame_header->using_qmatrix ? frame_header->qm_u : 0xFF; 197 pp->quantization.qm_v = frame_header->using_qmatrix ? frame_header->qm_v : 0xFF; 198 199 /* Cdef parameters */ 200 pp->cdef.damping = frame_header->cdef_damping_minus_3; 201 pp->cdef.bits = frame_header->cdef_bits; 202 for (i = 0; i < 8; i++) { 203 pp->cdef.y_strengths[i].primary = frame_header->cdef_y_pri_strength[i]; 204 pp->cdef.y_strengths[i].secondary = frame_header->cdef_y_sec_strength[i]; 205 pp->cdef.uv_strengths[i].primary = frame_header->cdef_uv_pri_strength[i]; 206 pp->cdef.uv_strengths[i].secondary = frame_header->cdef_uv_sec_strength[i]; 207 } 208 209 /* Misc flags */ 210 pp->interp_filter = frame_header->interpolation_filter; 211 212 /* Segmentation */ 213 pp->segmentation.enabled = frame_header->segmentation_enabled; 214 pp->segmentation.update_map = frame_header->segmentation_update_map; 215 pp->segmentation.update_data = frame_header->segmentation_update_data; 216 pp->segmentation.temporal_update = frame_header->segmentation_temporal_update; 217 for (i = 0; i < AV1_MAX_SEGMENTS; i++) { 218 for (j = 0; j < AV1_SEG_LVL_MAX; j++) { 219 pp->segmentation.feature_mask[i].mask |= frame_header->feature_enabled[i][j] << j; 220 pp->segmentation.feature_data[i][j] = frame_header->feature_value[i][j]; 221 } 222 } 223 224 /* Film grain */ 225 if (apply_grain) { 226 pp->film_grain.apply_grain = 1; 227 pp->film_grain.scaling_shift_minus8 = film_grain->grain_scaling_minus_8; 228 pp->film_grain.chroma_scaling_from_luma = film_grain->chroma_scaling_from_luma; 229 pp->film_grain.ar_coeff_lag = film_grain->ar_coeff_lag; 230 pp->film_grain.ar_coeff_shift_minus6 = film_grain->ar_coeff_shift_minus_6; 231 pp->film_grain.grain_scale_shift = film_grain->grain_scale_shift; 232 pp->film_grain.overlap_flag = film_grain->overlap_flag; 233 pp->film_grain.clip_to_restricted_range = film_grain->clip_to_restricted_range; 234 pp->film_grain.matrix_coeff_is_identity = (seq->color_config.matrix_coefficients == AVCOL_SPC_RGB); 235 236 pp->film_grain.grain_seed = film_grain->grain_seed; 237 pp->film_grain.num_y_points = film_grain->num_y_points; 238 for (i = 0; i < film_grain->num_y_points; i++) { 239 pp->film_grain.scaling_points_y[i][0] = film_grain->point_y_value[i]; 240 pp->film_grain.scaling_points_y[i][1] = film_grain->point_y_scaling[i]; 241 } 242 pp->film_grain.num_cb_points = film_grain->num_cb_points; 243 for (i = 0; i < film_grain->num_cb_points; i++) { 244 pp->film_grain.scaling_points_cb[i][0] = film_grain->point_cb_value[i]; 245 pp->film_grain.scaling_points_cb[i][1] = film_grain->point_cb_scaling[i]; 246 } 247 pp->film_grain.num_cr_points = film_grain->num_cr_points; 248 for (i = 0; i < film_grain->num_cr_points; i++) { 249 pp->film_grain.scaling_points_cr[i][0] = film_grain->point_cr_value[i]; 250 pp->film_grain.scaling_points_cr[i][1] = film_grain->point_cr_scaling[i]; 251 } 252 for (i = 0; i < 24; i++) { 253 pp->film_grain.ar_coeffs_y[i] = film_grain->ar_coeffs_y_plus_128[i]; 254 } 255 for (i = 0; i < 25; i++) { 256 pp->film_grain.ar_coeffs_cb[i] = film_grain->ar_coeffs_cb_plus_128[i]; 257 pp->film_grain.ar_coeffs_cr[i] = film_grain->ar_coeffs_cr_plus_128[i]; 258 } 259 pp->film_grain.cb_mult = film_grain->cb_mult; 260 pp->film_grain.cb_luma_mult = film_grain->cb_luma_mult; 261 pp->film_grain.cr_mult = film_grain->cr_mult; 262 pp->film_grain.cr_luma_mult = film_grain->cr_luma_mult; 263 pp->film_grain.cb_offset = film_grain->cb_offset; 264 pp->film_grain.cr_offset = film_grain->cr_offset; 265 pp->film_grain.cr_offset = film_grain->cr_offset; 266 } 267 268 // XXX: Setting the StatusReportFeedbackNumber breaks decoding on some drivers (tested on NVIDIA 457.09) 269 // Status Reporting is not used by FFmpeg, hence not providing a number does not cause any issues 270 //pp->StatusReportFeedbackNumber = 1 + DXVA_CONTEXT_REPORT_ID(avctx, ctx)++; 271 return 0; 272 } 273 274 static int dxva2_av1_start_frame(AVCodecContext *avctx, 275 av_unused const uint8_t *buffer, 276 av_unused uint32_t size) 277 { 278 const AV1DecContext *h = avctx->priv_data; 279 AVDXVAContext *ctx = DXVA_CONTEXT(avctx); 280 struct av1_dxva2_picture_context *ctx_pic = h->cur_frame.hwaccel_picture_private; 281 282 if (!DXVA_CONTEXT_VALID(avctx, ctx)) 283 return -1; 284 av_assert0(ctx_pic); 285 286 /* Fill up DXVA_PicParams_AV1 */ 287 if (ff_dxva2_av1_fill_picture_parameters(avctx, ctx, &ctx_pic->pp) < 0) 288 return -1; 289 290 ctx_pic->bitstream_size = 0; 291 ctx_pic->bitstream = NULL; 292 return 0; 293 } 294 295 static int dxva2_av1_decode_slice(AVCodecContext *avctx, 296 const uint8_t *buffer, 297 uint32_t size) 298 { 299 const AV1DecContext *h = avctx->priv_data; 300 const AV1RawFrameHeader *frame_header = h->raw_frame_header; 301 struct av1_dxva2_picture_context *ctx_pic = h->cur_frame.hwaccel_picture_private; 302 struct AV1DXVAContext *ctx = avctx->internal->hwaccel_priv_data; 303 void *tmp; 304 305 ctx_pic->tile_count = frame_header->tile_cols * frame_header->tile_rows; 306 307 /* too many tiles, exceeding all defined levels in the AV1 spec */ 308 if (ctx_pic->tile_count > MAX_TILES) 309 return AVERROR(ENOSYS); 310 311 /* Shortcut if all tiles are in the same buffer */ 312 if (ctx_pic->tile_count == h->tg_end - h->tg_start + 1) { 313 ctx_pic->bitstream = (uint8_t *)buffer; 314 ctx_pic->bitstream_size = size; 315 316 for (uint32_t tile_num = 0; tile_num < ctx_pic->tile_count; tile_num++) { 317 ctx_pic->tiles[tile_num].DataOffset = h->tile_group_info[tile_num].tile_offset; 318 ctx_pic->tiles[tile_num].DataSize = h->tile_group_info[tile_num].tile_size; 319 ctx_pic->tiles[tile_num].row = h->tile_group_info[tile_num].tile_row; 320 ctx_pic->tiles[tile_num].column = h->tile_group_info[tile_num].tile_column; 321 ctx_pic->tiles[tile_num].anchor_frame = 0xFF; 322 } 323 324 return 0; 325 } 326 327 /* allocate an internal buffer */ 328 tmp = av_fast_realloc(ctx->bitstream_cache, &ctx->bitstream_allocated, 329 ctx_pic->bitstream_size + size); 330 if (!tmp) { 331 return AVERROR(ENOMEM); 332 } 333 ctx_pic->bitstream = ctx->bitstream_cache = tmp; 334 335 memcpy(ctx_pic->bitstream + ctx_pic->bitstream_size, buffer, size); 336 337 for (uint32_t tile_num = h->tg_start; tile_num <= h->tg_end; tile_num++) { 338 ctx_pic->tiles[tile_num].DataOffset = ctx_pic->bitstream_size + h->tile_group_info[tile_num].tile_offset; 339 ctx_pic->tiles[tile_num].DataSize = h->tile_group_info[tile_num].tile_size; 340 ctx_pic->tiles[tile_num].row = h->tile_group_info[tile_num].tile_row; 341 ctx_pic->tiles[tile_num].column = h->tile_group_info[tile_num].tile_column; 342 ctx_pic->tiles[tile_num].anchor_frame = 0xFF; 343 } 344 345 ctx_pic->bitstream_size += size; 346 347 return 0; 348 } 349 350 static int commit_bitstream_and_slice_buffer(AVCodecContext *avctx, 351 DECODER_BUFFER_DESC *bs, 352 DECODER_BUFFER_DESC *sc) 353 { 354 const AV1DecContext *h = avctx->priv_data; 355 AVDXVAContext *ctx = DXVA_CONTEXT(avctx); 356 struct av1_dxva2_picture_context *ctx_pic = h->cur_frame.hwaccel_picture_private; 357 void *dxva_data_ptr = NULL; 358 uint8_t *dxva_data; 359 unsigned dxva_size; 360 unsigned padding; 361 unsigned type; 362 363 #if CONFIG_D3D11VA 364 if (ff_dxva2_is_d3d11(avctx)) { 365 type = D3D11_VIDEO_DECODER_BUFFER_BITSTREAM; 366 if (FAILED(ID3D11VideoContext_GetDecoderBuffer(D3D11VA_CONTEXT(ctx)->video_context, 367 D3D11VA_CONTEXT(ctx)->decoder, 368 type, 369 &dxva_size, &dxva_data_ptr))) 370 return -1; 371 } 372 #endif 373 #if CONFIG_DXVA2 374 if (avctx->pix_fmt == AV_PIX_FMT_DXVA2_VLD) { 375 type = DXVA2_BitStreamDateBufferType; 376 if (FAILED(IDirectXVideoDecoder_GetBuffer(DXVA2_CONTEXT(ctx)->decoder, 377 type, 378 &dxva_data_ptr, &dxva_size))) 379 return -1; 380 } 381 #endif 382 383 dxva_data = dxva_data_ptr; 384 385 if (!dxva_data || ctx_pic->bitstream_size > dxva_size) { 386 av_log(avctx, AV_LOG_ERROR, "Bitstream size exceeds hardware buffer"); 387 return -1; 388 } 389 390 memcpy(dxva_data, ctx_pic->bitstream, ctx_pic->bitstream_size); 391 392 padding = FFMIN(128 - ((ctx_pic->bitstream_size) & 127), dxva_size - ctx_pic->bitstream_size); 393 if (padding > 0) { 394 memset(dxva_data + ctx_pic->bitstream_size, 0, padding); 395 ctx_pic->bitstream_size += padding; 396 } 397 398 #if CONFIG_D3D11VA 399 if (ff_dxva2_is_d3d11(avctx)) 400 if (FAILED(ID3D11VideoContext_ReleaseDecoderBuffer(D3D11VA_CONTEXT(ctx)->video_context, D3D11VA_CONTEXT(ctx)->decoder, type))) 401 return -1; 402 #endif 403 #if CONFIG_DXVA2 404 if (avctx->pix_fmt == AV_PIX_FMT_DXVA2_VLD) 405 if (FAILED(IDirectXVideoDecoder_ReleaseBuffer(DXVA2_CONTEXT(ctx)->decoder, type))) 406 return -1; 407 #endif 408 409 #if CONFIG_D3D11VA 410 if (ff_dxva2_is_d3d11(avctx)) { 411 D3D11_VIDEO_DECODER_BUFFER_DESC *dsc11 = bs; 412 memset(dsc11, 0, sizeof(*dsc11)); 413 dsc11->BufferType = type; 414 dsc11->DataSize = ctx_pic->bitstream_size; 415 dsc11->NumMBsInBuffer = 0; 416 417 type = D3D11_VIDEO_DECODER_BUFFER_SLICE_CONTROL; 418 } 419 #endif 420 #if CONFIG_DXVA2 421 if (avctx->pix_fmt == AV_PIX_FMT_DXVA2_VLD) { 422 DXVA2_DecodeBufferDesc *dsc2 = bs; 423 memset(dsc2, 0, sizeof(*dsc2)); 424 dsc2->CompressedBufferType = type; 425 dsc2->DataSize = ctx_pic->bitstream_size; 426 dsc2->NumMBsInBuffer = 0; 427 428 type = DXVA2_SliceControlBufferType; 429 } 430 #endif 431 432 return ff_dxva2_commit_buffer(avctx, ctx, sc, type, 433 ctx_pic->tiles, sizeof(*ctx_pic->tiles) * ctx_pic->tile_count, 0); 434 } 435 436 static int dxva2_av1_end_frame(AVCodecContext *avctx) 437 { 438 const AV1DecContext *h = avctx->priv_data; 439 struct av1_dxva2_picture_context *ctx_pic = h->cur_frame.hwaccel_picture_private; 440 int ret; 441 442 if (ctx_pic->bitstream_size <= 0) 443 return -1; 444 445 ret = ff_dxva2_common_end_frame(avctx, h->cur_frame.f, 446 &ctx_pic->pp, sizeof(ctx_pic->pp), 447 NULL, 0, 448 commit_bitstream_and_slice_buffer); 449 450 return ret; 451 } 452 453 static int dxva2_av1_uninit(AVCodecContext *avctx) 454 { 455 struct AV1DXVAContext *ctx = avctx->internal->hwaccel_priv_data; 456 457 av_freep(&ctx->bitstream_cache); 458 ctx->bitstream_allocated = 0; 459 460 return ff_dxva2_decode_uninit(avctx); 461 } 462 463 #if CONFIG_AV1_DXVA2_HWACCEL 464 const FFHWAccel ff_av1_dxva2_hwaccel = { 465 .p.name = "av1_dxva2", 466 .p.type = AVMEDIA_TYPE_VIDEO, 467 .p.id = AV_CODEC_ID_AV1, 468 .p.pix_fmt = AV_PIX_FMT_DXVA2_VLD, 469 .init = ff_dxva2_decode_init, 470 .uninit = dxva2_av1_uninit, 471 .start_frame = dxva2_av1_start_frame, 472 .decode_slice = dxva2_av1_decode_slice, 473 .end_frame = dxva2_av1_end_frame, 474 .frame_params = ff_dxva2_common_frame_params, 475 .frame_priv_data_size = sizeof(struct av1_dxva2_picture_context), 476 .priv_data_size = sizeof(struct AV1DXVAContext), 477 }; 478 #endif 479 480 #if CONFIG_AV1_D3D11VA_HWACCEL 481 const FFHWAccel ff_av1_d3d11va_hwaccel = { 482 .p.name = "av1_d3d11va", 483 .p.type = AVMEDIA_TYPE_VIDEO, 484 .p.id = AV_CODEC_ID_AV1, 485 .p.pix_fmt = AV_PIX_FMT_D3D11VA_VLD, 486 .init = ff_dxva2_decode_init, 487 .uninit = dxva2_av1_uninit, 488 .start_frame = dxva2_av1_start_frame, 489 .decode_slice = dxva2_av1_decode_slice, 490 .end_frame = dxva2_av1_end_frame, 491 .frame_params = ff_dxva2_common_frame_params, 492 .frame_priv_data_size = sizeof(struct av1_dxva2_picture_context), 493 .priv_data_size = sizeof(struct AV1DXVAContext), 494 }; 495 #endif 496 497 #if CONFIG_AV1_D3D11VA2_HWACCEL 498 const FFHWAccel ff_av1_d3d11va2_hwaccel = { 499 .p.name = "av1_d3d11va2", 500 .p.type = AVMEDIA_TYPE_VIDEO, 501 .p.id = AV_CODEC_ID_AV1, 502 .p.pix_fmt = AV_PIX_FMT_D3D11, 503 .init = ff_dxva2_decode_init, 504 .uninit = dxva2_av1_uninit, 505 .start_frame = dxva2_av1_start_frame, 506 .decode_slice = dxva2_av1_decode_slice, 507 .end_frame = dxva2_av1_end_frame, 508 .frame_params = ff_dxva2_common_frame_params, 509 .frame_priv_data_size = sizeof(struct av1_dxva2_picture_context), 510 .priv_data_size = sizeof(struct AV1DXVAContext), 511 }; 512 #endif