You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1323 lines
47KB

  1. /*
  2. * This file is part of Libav.
  3. *
  4. * Libav is free software; you can redistribute it and/or
  5. * modify it under the terms of the GNU Lesser General Public
  6. * License as published by the Free Software Foundation; either
  7. * version 2.1 of the License, or (at your option) any later version.
  8. *
  9. * Libav is distributed in the hope that it will be useful,
  10. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  11. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  12. * Lesser General Public License for more details.
  13. *
  14. * You should have received a copy of the GNU Lesser General Public
  15. * License along with Libav; if not, write to the Free Software
  16. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  17. */
  18. #include <va/va.h>
  19. #include <va/va_enc_h264.h>
  20. #include "libavutil/avassert.h"
  21. #include "libavutil/internal.h"
  22. #include "libavutil/opt.h"
  23. #include "libavutil/pixfmt.h"
  24. #include "avcodec.h"
  25. #include "h264.h"
  26. #include "h264_sei.h"
  27. #include "internal.h"
  28. #include "vaapi_encode.h"
  29. #include "vaapi_encode_h26x.h"
  30. enum {
  31. SLICE_TYPE_P = 0,
  32. SLICE_TYPE_B = 1,
  33. SLICE_TYPE_I = 2,
  34. SLICE_TYPE_SP = 3,
  35. SLICE_TYPE_SI = 4,
  36. };
  37. // This structure contains all possibly-useful per-sequence syntax elements
  38. // which are not already contained in the various VAAPI structures.
  39. typedef struct VAAPIEncodeH264MiscSequenceParams {
  40. unsigned int profile_idc;
  41. char constraint_set0_flag;
  42. char constraint_set1_flag;
  43. char constraint_set2_flag;
  44. char constraint_set3_flag;
  45. char constraint_set4_flag;
  46. char constraint_set5_flag;
  47. char separate_colour_plane_flag;
  48. char qpprime_y_zero_transform_bypass_flag;
  49. char gaps_in_frame_num_allowed_flag;
  50. char delta_pic_order_always_zero_flag;
  51. char bottom_field_pic_order_in_frame_present_flag;
  52. unsigned int num_slice_groups_minus1;
  53. unsigned int slice_group_map_type;
  54. int pic_init_qs_minus26;
  55. char overscan_info_present_flag;
  56. char overscan_appropriate_flag;
  57. char video_signal_type_present_flag;
  58. unsigned int video_format;
  59. char video_full_range_flag;
  60. char colour_description_present_flag;
  61. unsigned int colour_primaries;
  62. unsigned int transfer_characteristics;
  63. unsigned int matrix_coefficients;
  64. char chroma_loc_info_present_flag;
  65. unsigned int chroma_sample_loc_type_top_field;
  66. unsigned int chroma_sample_loc_type_bottom_field;
  67. // Some timing elements are in VAEncSequenceParameterBufferH264.
  68. char fixed_frame_rate_flag;
  69. char nal_hrd_parameters_present_flag;
  70. char vcl_hrd_parameters_present_flag;
  71. char low_delay_hrd_flag;
  72. char pic_struct_present_flag;
  73. char bitstream_restriction_flag;
  74. unsigned int cpb_cnt_minus1;
  75. unsigned int bit_rate_scale;
  76. unsigned int cpb_size_scale;
  77. unsigned int bit_rate_value_minus1[32];
  78. unsigned int cpb_size_value_minus1[32];
  79. char cbr_flag[32];
  80. unsigned int initial_cpb_removal_delay_length_minus1;
  81. unsigned int cpb_removal_delay_length_minus1;
  82. unsigned int dpb_output_delay_length_minus1;
  83. unsigned int time_offset_length;
  84. unsigned int initial_cpb_removal_delay;
  85. unsigned int initial_cpb_removal_delay_offset;
  86. unsigned int pic_struct;
  87. } VAAPIEncodeH264MiscSequenceParams;
  88. // This structure contains all possibly-useful per-slice syntax elements
  89. // which are not already contained in the various VAAPI structures.
  90. typedef struct VAAPIEncodeH264MiscSliceParams {
  91. unsigned int nal_unit_type;
  92. unsigned int nal_ref_idc;
  93. unsigned int colour_plane_id;
  94. char field_pic_flag;
  95. char bottom_field_flag;
  96. unsigned int redundant_pic_cnt;
  97. char sp_for_switch_flag;
  98. int slice_qs_delta;
  99. char ref_pic_list_modification_flag_l0;
  100. char ref_pic_list_modification_flag_l1;
  101. char no_output_of_prior_pics_flag;
  102. char long_term_reference_flag;
  103. char adaptive_ref_pic_marking_mode_flag;
  104. } VAAPIEncodeH264MiscSliceParams;
  105. typedef struct VAAPIEncodeH264Slice {
  106. VAAPIEncodeH264MiscSliceParams misc_slice_params;
  107. } VAAPIEncodeH264Slice;
  108. typedef struct VAAPIEncodeH264Context {
  109. VAAPIEncodeH264MiscSequenceParams misc_sequence_params;
  110. int mb_width;
  111. int mb_height;
  112. int fixed_qp_idr;
  113. int fixed_qp_p;
  114. int fixed_qp_b;
  115. int next_frame_num;
  116. int64_t idr_pic_count;
  117. int cpb_delay;
  118. int dpb_delay;
  119. // Rate control configuration.
  120. int send_timing_sei;
  121. struct {
  122. VAEncMiscParameterBuffer misc;
  123. VAEncMiscParameterRateControl rc;
  124. } rc_params;
  125. struct {
  126. VAEncMiscParameterBuffer misc;
  127. VAEncMiscParameterHRD hrd;
  128. } hrd_params;
  129. #if VA_CHECK_VERSION(0, 36, 0)
  130. // Speed-quality tradeoff setting.
  131. struct {
  132. VAEncMiscParameterBuffer misc;
  133. VAEncMiscParameterBufferQualityLevel quality;
  134. } quality_params;
  135. #endif
  136. } VAAPIEncodeH264Context;
  137. typedef struct VAAPIEncodeH264Options {
  138. int qp;
  139. int quality;
  140. int low_power;
  141. } VAAPIEncodeH264Options;
  142. #define vseq_var(name) vseq->name, name
  143. #define vseq_field(name) vseq->seq_fields.bits.name, name
  144. #define vvui_field(name) vseq->vui_fields.bits.name, name
  145. #define vpic_var(name) vpic->name, name
  146. #define vpic_field(name) vpic->pic_fields.bits.name, name
  147. #define vslice_var(name) vslice->name, name
  148. #define vslice_field(name) vslice->slice_fields.bits.name, name
  149. #define mseq_var(name) mseq->name, name
  150. #define mslice_var(name) mslice->name, name
  151. static void vaapi_encode_h264_write_nal_header(PutBitContext *pbc,
  152. int nal_unit_type, int nal_ref_idc)
  153. {
  154. u(1, 0, forbidden_zero_bit);
  155. u(2, nal_ref_idc, nal_ref_idc);
  156. u(5, nal_unit_type, nal_unit_type);
  157. }
  158. static void vaapi_encode_h264_write_trailing_rbsp(PutBitContext *pbc)
  159. {
  160. u(1, 1, rbsp_stop_one_bit);
  161. while (put_bits_count(pbc) & 7)
  162. u(1, 0, rbsp_alignment_zero_bit);
  163. }
  164. static void vaapi_encode_h264_write_vui(PutBitContext *pbc,
  165. VAAPIEncodeContext *ctx)
  166. {
  167. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  168. VAAPIEncodeH264Context *priv = ctx->priv_data;
  169. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  170. int i;
  171. u(1, vvui_field(aspect_ratio_info_present_flag));
  172. if (vseq->vui_fields.bits.aspect_ratio_info_present_flag) {
  173. u(8, vseq_var(aspect_ratio_idc));
  174. if (vseq->aspect_ratio_idc == 255) {
  175. u(16, vseq_var(sar_width));
  176. u(16, vseq_var(sar_height));
  177. }
  178. }
  179. u(1, mseq_var(overscan_info_present_flag));
  180. if (mseq->overscan_info_present_flag)
  181. u(1, mseq_var(overscan_appropriate_flag));
  182. u(1, mseq_var(video_signal_type_present_flag));
  183. if (mseq->video_signal_type_present_flag) {
  184. u(3, mseq_var(video_format));
  185. u(1, mseq_var(video_full_range_flag));
  186. u(1, mseq_var(colour_description_present_flag));
  187. if (mseq->colour_description_present_flag) {
  188. u(8, mseq_var(colour_primaries));
  189. u(8, mseq_var(transfer_characteristics));
  190. u(8, mseq_var(matrix_coefficients));
  191. }
  192. }
  193. u(1, mseq_var(chroma_loc_info_present_flag));
  194. if (mseq->chroma_loc_info_present_flag) {
  195. ue(mseq_var(chroma_sample_loc_type_top_field));
  196. ue(mseq_var(chroma_sample_loc_type_bottom_field));
  197. }
  198. u(1, vvui_field(timing_info_present_flag));
  199. if (vseq->vui_fields.bits.timing_info_present_flag) {
  200. u(32, vseq_var(num_units_in_tick));
  201. u(32, vseq_var(time_scale));
  202. u(1, mseq_var(fixed_frame_rate_flag));
  203. }
  204. u(1, mseq_var(nal_hrd_parameters_present_flag));
  205. if (mseq->nal_hrd_parameters_present_flag) {
  206. ue(mseq_var(cpb_cnt_minus1));
  207. u(4, mseq_var(bit_rate_scale));
  208. u(4, mseq_var(cpb_size_scale));
  209. for (i = 0; i <= mseq->cpb_cnt_minus1; i++) {
  210. ue(mseq_var(bit_rate_value_minus1[i]));
  211. ue(mseq_var(cpb_size_value_minus1[i]));
  212. u(1, mseq_var(cbr_flag[i]));
  213. }
  214. u(5, mseq_var(initial_cpb_removal_delay_length_minus1));
  215. u(5, mseq_var(cpb_removal_delay_length_minus1));
  216. u(5, mseq_var(dpb_output_delay_length_minus1));
  217. u(5, mseq_var(time_offset_length));
  218. }
  219. u(1, mseq_var(vcl_hrd_parameters_present_flag));
  220. if (mseq->vcl_hrd_parameters_present_flag) {
  221. av_assert0(0 && "vcl hrd parameters not supported");
  222. }
  223. if (mseq->nal_hrd_parameters_present_flag ||
  224. mseq->vcl_hrd_parameters_present_flag)
  225. u(1, mseq_var(low_delay_hrd_flag));
  226. u(1, mseq_var(pic_struct_present_flag));
  227. u(1, vvui_field(bitstream_restriction_flag));
  228. if (vseq->vui_fields.bits.bitstream_restriction_flag) {
  229. av_assert0(0 && "bitstream restrictions not supported");
  230. }
  231. }
  232. static void vaapi_encode_h264_write_sps(PutBitContext *pbc,
  233. VAAPIEncodeContext *ctx)
  234. {
  235. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  236. VAAPIEncodeH264Context *priv = ctx->priv_data;
  237. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  238. int i;
  239. vaapi_encode_h264_write_nal_header(pbc, NAL_SPS, 3);
  240. u(8, mseq_var(profile_idc));
  241. u(1, mseq_var(constraint_set0_flag));
  242. u(1, mseq_var(constraint_set1_flag));
  243. u(1, mseq_var(constraint_set2_flag));
  244. u(1, mseq_var(constraint_set3_flag));
  245. u(1, mseq_var(constraint_set4_flag));
  246. u(1, mseq_var(constraint_set5_flag));
  247. u(2, 0, reserved_zero_2bits);
  248. u(8, vseq_var(level_idc));
  249. ue(vseq_var(seq_parameter_set_id));
  250. if (mseq->profile_idc == 100 || mseq->profile_idc == 110 ||
  251. mseq->profile_idc == 122 || mseq->profile_idc == 244 ||
  252. mseq->profile_idc == 44 || mseq->profile_idc == 83 ||
  253. mseq->profile_idc == 86 || mseq->profile_idc == 118 ||
  254. mseq->profile_idc == 128 || mseq->profile_idc == 138) {
  255. ue(vseq_field(chroma_format_idc));
  256. if (vseq->seq_fields.bits.chroma_format_idc == 3)
  257. u(1, mseq_var(separate_colour_plane_flag));
  258. ue(vseq_var(bit_depth_luma_minus8));
  259. ue(vseq_var(bit_depth_chroma_minus8));
  260. u(1, mseq_var(qpprime_y_zero_transform_bypass_flag));
  261. u(1, vseq_field(seq_scaling_matrix_present_flag));
  262. if (vseq->seq_fields.bits.seq_scaling_matrix_present_flag) {
  263. av_assert0(0 && "scaling matrices not supported");
  264. }
  265. }
  266. ue(vseq_field(log2_max_frame_num_minus4));
  267. ue(vseq_field(pic_order_cnt_type));
  268. if (vseq->seq_fields.bits.pic_order_cnt_type == 0) {
  269. ue(vseq_field(log2_max_pic_order_cnt_lsb_minus4));
  270. } else if (vseq->seq_fields.bits.pic_order_cnt_type == 1) {
  271. u(1, mseq_var(delta_pic_order_always_zero_flag));
  272. se(vseq_var(offset_for_non_ref_pic));
  273. se(vseq_var(offset_for_top_to_bottom_field));
  274. ue(vseq_var(num_ref_frames_in_pic_order_cnt_cycle));
  275. for (i = 0; i < vseq->num_ref_frames_in_pic_order_cnt_cycle; i++)
  276. se(vseq_var(offset_for_ref_frame[i]));
  277. }
  278. ue(vseq_var(max_num_ref_frames));
  279. u(1, mseq_var(gaps_in_frame_num_allowed_flag));
  280. ue(vseq->picture_width_in_mbs - 1, pic_width_in_mbs_minus1);
  281. ue(vseq->picture_height_in_mbs - 1, pic_height_in_mbs_minus1);
  282. u(1, vseq_field(frame_mbs_only_flag));
  283. if (!vseq->seq_fields.bits.frame_mbs_only_flag)
  284. u(1, vseq_field(mb_adaptive_frame_field_flag));
  285. u(1, vseq_field(direct_8x8_inference_flag));
  286. u(1, vseq_var(frame_cropping_flag));
  287. if (vseq->frame_cropping_flag) {
  288. ue(vseq_var(frame_crop_left_offset));
  289. ue(vseq_var(frame_crop_right_offset));
  290. ue(vseq_var(frame_crop_top_offset));
  291. ue(vseq_var(frame_crop_bottom_offset));
  292. }
  293. u(1, vseq_var(vui_parameters_present_flag));
  294. if (vseq->vui_parameters_present_flag)
  295. vaapi_encode_h264_write_vui(pbc, ctx);
  296. vaapi_encode_h264_write_trailing_rbsp(pbc);
  297. }
  298. static void vaapi_encode_h264_write_pps(PutBitContext *pbc,
  299. VAAPIEncodeContext *ctx)
  300. {
  301. VAEncPictureParameterBufferH264 *vpic = ctx->codec_picture_params;
  302. VAAPIEncodeH264Context *priv = ctx->priv_data;
  303. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  304. vaapi_encode_h264_write_nal_header(pbc, NAL_PPS, 3);
  305. ue(vpic_var(pic_parameter_set_id));
  306. ue(vpic_var(seq_parameter_set_id));
  307. u(1, vpic_field(entropy_coding_mode_flag));
  308. u(1, mseq_var(bottom_field_pic_order_in_frame_present_flag));
  309. ue(mseq_var(num_slice_groups_minus1));
  310. if (mseq->num_slice_groups_minus1 > 0) {
  311. ue(mseq_var(slice_group_map_type));
  312. av_assert0(0 && "slice groups not supported");
  313. }
  314. ue(vpic_var(num_ref_idx_l0_active_minus1));
  315. ue(vpic_var(num_ref_idx_l1_active_minus1));
  316. u(1, vpic_field(weighted_pred_flag));
  317. u(2, vpic_field(weighted_bipred_idc));
  318. se(vpic->pic_init_qp - 26, pic_init_qp_minus26);
  319. se(mseq_var(pic_init_qs_minus26));
  320. se(vpic_var(chroma_qp_index_offset));
  321. u(1, vpic_field(deblocking_filter_control_present_flag));
  322. u(1, vpic_field(constrained_intra_pred_flag));
  323. u(1, vpic_field(redundant_pic_cnt_present_flag));
  324. u(1, vpic_field(transform_8x8_mode_flag));
  325. u(1, vpic_field(pic_scaling_matrix_present_flag));
  326. if (vpic->pic_fields.bits.pic_scaling_matrix_present_flag) {
  327. av_assert0(0 && "scaling matrices not supported");
  328. }
  329. se(vpic_var(second_chroma_qp_index_offset));
  330. vaapi_encode_h264_write_trailing_rbsp(pbc);
  331. }
  332. static void vaapi_encode_h264_write_slice_header2(PutBitContext *pbc,
  333. VAAPIEncodeContext *ctx,
  334. VAAPIEncodePicture *pic,
  335. VAAPIEncodeSlice *slice)
  336. {
  337. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  338. VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params;
  339. VAEncSliceParameterBufferH264 *vslice = slice->codec_slice_params;
  340. VAAPIEncodeH264Context *priv = ctx->priv_data;
  341. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  342. VAAPIEncodeH264Slice *pslice = slice->priv_data;
  343. VAAPIEncodeH264MiscSliceParams *mslice = &pslice->misc_slice_params;
  344. vaapi_encode_h264_write_nal_header(pbc, mslice->nal_unit_type,
  345. mslice->nal_ref_idc);
  346. ue(vslice->macroblock_address, first_mb_in_slice);
  347. ue(vslice_var(slice_type));
  348. ue(vpic_var(pic_parameter_set_id));
  349. if (mseq->separate_colour_plane_flag) {
  350. u(2, mslice_var(colour_plane_id));
  351. }
  352. u(4 + vseq->seq_fields.bits.log2_max_frame_num_minus4,
  353. (vpic->frame_num &
  354. ((1 << (4 + vseq->seq_fields.bits.log2_max_frame_num_minus4)) - 1)),
  355. frame_num);
  356. if (!vseq->seq_fields.bits.frame_mbs_only_flag) {
  357. u(1, mslice_var(field_pic_flag));
  358. if (mslice->field_pic_flag)
  359. u(1, mslice_var(bottom_field_flag));
  360. }
  361. if (vpic->pic_fields.bits.idr_pic_flag) {
  362. ue(vslice_var(idr_pic_id));
  363. }
  364. if (vseq->seq_fields.bits.pic_order_cnt_type == 0) {
  365. u(4 + vseq->seq_fields.bits.log2_max_pic_order_cnt_lsb_minus4,
  366. vslice_var(pic_order_cnt_lsb));
  367. if (mseq->bottom_field_pic_order_in_frame_present_flag &&
  368. !mslice->field_pic_flag) {
  369. se(vslice_var(delta_pic_order_cnt_bottom));
  370. }
  371. }
  372. if (vseq->seq_fields.bits.pic_order_cnt_type == 1 &&
  373. !vseq->seq_fields.bits.delta_pic_order_always_zero_flag) {
  374. se(vslice_var(delta_pic_order_cnt[0]));
  375. if (mseq->bottom_field_pic_order_in_frame_present_flag &&
  376. !mslice->field_pic_flag) {
  377. se(vslice_var(delta_pic_order_cnt[1]));
  378. }
  379. }
  380. if (vpic->pic_fields.bits.redundant_pic_cnt_present_flag) {
  381. ue(mslice_var(redundant_pic_cnt));
  382. }
  383. if (vslice->slice_type == SLICE_TYPE_B) {
  384. u(1, vslice_var(direct_spatial_mv_pred_flag));
  385. }
  386. if (vslice->slice_type == SLICE_TYPE_P ||
  387. vslice->slice_type == SLICE_TYPE_SP ||
  388. vslice->slice_type == SLICE_TYPE_B) {
  389. u(1, vslice_var(num_ref_idx_active_override_flag));
  390. if (vslice->num_ref_idx_active_override_flag) {
  391. ue(vslice_var(num_ref_idx_l0_active_minus1));
  392. if (vslice->slice_type == SLICE_TYPE_B)
  393. ue(vslice_var(num_ref_idx_l1_active_minus1));
  394. }
  395. }
  396. if (mslice->nal_unit_type == 20 || mslice->nal_unit_type == 21) {
  397. av_assert0(0 && "no MVC support");
  398. } else {
  399. if (vslice->slice_type % 5 != 2 && vslice->slice_type % 5 != 4) {
  400. u(1, mslice_var(ref_pic_list_modification_flag_l0));
  401. if (mslice->ref_pic_list_modification_flag_l0) {
  402. av_assert0(0 && "ref pic list modification");
  403. }
  404. }
  405. if (vslice->slice_type % 5 == 1) {
  406. u(1, mslice_var(ref_pic_list_modification_flag_l1));
  407. if (mslice->ref_pic_list_modification_flag_l1) {
  408. av_assert0(0 && "ref pic list modification");
  409. }
  410. }
  411. }
  412. if ((vpic->pic_fields.bits.weighted_pred_flag &&
  413. (vslice->slice_type == SLICE_TYPE_P ||
  414. vslice->slice_type == SLICE_TYPE_SP)) ||
  415. (vpic->pic_fields.bits.weighted_bipred_idc == 1 &&
  416. vslice->slice_type == SLICE_TYPE_B)) {
  417. av_assert0(0 && "prediction weights not supported");
  418. }
  419. av_assert0(mslice->nal_ref_idc > 0 ==
  420. vpic->pic_fields.bits.reference_pic_flag);
  421. if (mslice->nal_ref_idc != 0) {
  422. if (vpic->pic_fields.bits.idr_pic_flag) {
  423. u(1, mslice_var(no_output_of_prior_pics_flag));
  424. u(1, mslice_var(long_term_reference_flag));
  425. } else {
  426. u(1, mslice_var(adaptive_ref_pic_marking_mode_flag));
  427. if (mslice->adaptive_ref_pic_marking_mode_flag) {
  428. av_assert0(0 && "MMCOs not supported");
  429. }
  430. }
  431. }
  432. if (vpic->pic_fields.bits.entropy_coding_mode_flag &&
  433. vslice->slice_type != SLICE_TYPE_I &&
  434. vslice->slice_type != SLICE_TYPE_SI) {
  435. ue(vslice_var(cabac_init_idc));
  436. }
  437. se(vslice_var(slice_qp_delta));
  438. if (vslice->slice_type == SLICE_TYPE_SP ||
  439. vslice->slice_type == SLICE_TYPE_SI) {
  440. if (vslice->slice_type == SLICE_TYPE_SP)
  441. u(1, mslice_var(sp_for_switch_flag));
  442. se(mslice_var(slice_qs_delta));
  443. }
  444. if (vpic->pic_fields.bits.deblocking_filter_control_present_flag) {
  445. ue(vslice_var(disable_deblocking_filter_idc));
  446. if (vslice->disable_deblocking_filter_idc != 1) {
  447. se(vslice_var(slice_alpha_c0_offset_div2));
  448. se(vslice_var(slice_beta_offset_div2));
  449. }
  450. }
  451. if (mseq->num_slice_groups_minus1 > 0 &&
  452. mseq->slice_group_map_type >= 3 && mseq->slice_group_map_type <= 5) {
  453. av_assert0(0 && "slice groups not supported");
  454. }
  455. // No alignment - this need not be a byte boundary.
  456. }
  457. static void vaapi_encode_h264_write_buffering_period(PutBitContext *pbc,
  458. VAAPIEncodeContext *ctx,
  459. VAAPIEncodePicture *pic)
  460. {
  461. VAAPIEncodeH264Context *priv = ctx->priv_data;
  462. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  463. VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params;
  464. int i;
  465. ue(vpic_var(seq_parameter_set_id));
  466. if (mseq->nal_hrd_parameters_present_flag) {
  467. for (i = 0; i <= mseq->cpb_cnt_minus1; i++) {
  468. u(mseq->initial_cpb_removal_delay_length_minus1 + 1,
  469. mseq_var(initial_cpb_removal_delay));
  470. u(mseq->initial_cpb_removal_delay_length_minus1 + 1,
  471. mseq_var(initial_cpb_removal_delay_offset));
  472. }
  473. }
  474. if (mseq->vcl_hrd_parameters_present_flag) {
  475. av_assert0(0 && "vcl hrd parameters not supported");
  476. }
  477. }
  478. static void vaapi_encode_h264_write_pic_timing(PutBitContext *pbc,
  479. VAAPIEncodeContext *ctx,
  480. VAAPIEncodePicture *pic)
  481. {
  482. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  483. VAAPIEncodeH264Context *priv = ctx->priv_data;
  484. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  485. int i, num_clock_ts;
  486. if (mseq->nal_hrd_parameters_present_flag ||
  487. mseq->vcl_hrd_parameters_present_flag) {
  488. u(mseq->cpb_removal_delay_length_minus1 + 1,
  489. 2 * vseq->num_units_in_tick * priv->cpb_delay,
  490. cpb_removal_delay);
  491. u(mseq->dpb_output_delay_length_minus1 + 1,
  492. 2 * vseq->num_units_in_tick * priv->dpb_delay,
  493. dpb_output_delay);
  494. }
  495. if (mseq->pic_struct_present_flag) {
  496. u(4, mseq_var(pic_struct));
  497. num_clock_ts = (mseq->pic_struct <= 2 ? 1 :
  498. mseq->pic_struct <= 4 ? 2 :
  499. mseq->pic_struct <= 8 ? 3 : 0);
  500. for (i = 0; i < num_clock_ts; i++) {
  501. u(1, 0, clock_timestamp_flag[i]);
  502. // No full timestamp information.
  503. }
  504. }
  505. }
  506. static void vaapi_encode_h264_write_sei(PutBitContext *pbc,
  507. VAAPIEncodeContext *ctx,
  508. VAAPIEncodePicture *pic)
  509. {
  510. VAAPIEncodeH264Context *priv = ctx->priv_data;
  511. PutBitContext payload_bits;
  512. char payload[256];
  513. int payload_type, payload_size, i;
  514. void (*write_payload)(PutBitContext *pbc,
  515. VAAPIEncodeContext *ctx,
  516. VAAPIEncodePicture *pic) = NULL;
  517. vaapi_encode_h264_write_nal_header(pbc, NAL_SEI, 0);
  518. for (payload_type = 0; payload_type < 64; payload_type++) {
  519. switch (payload_type) {
  520. case SEI_TYPE_BUFFERING_PERIOD:
  521. if (!priv->send_timing_sei ||
  522. pic->type != PICTURE_TYPE_IDR)
  523. continue;
  524. write_payload = &vaapi_encode_h264_write_buffering_period;
  525. break;
  526. case SEI_TYPE_PIC_TIMING:
  527. if (!priv->send_timing_sei)
  528. continue;
  529. write_payload = &vaapi_encode_h264_write_pic_timing;
  530. break;
  531. default:
  532. continue;
  533. }
  534. init_put_bits(&payload_bits, payload, sizeof(payload));
  535. write_payload(&payload_bits, ctx, pic);
  536. if (put_bits_count(&payload_bits) & 7) {
  537. write_u(&payload_bits, 1, 1, bit_equal_to_one);
  538. while (put_bits_count(&payload_bits) & 7)
  539. write_u(&payload_bits, 1, 0, bit_equal_to_zero);
  540. }
  541. payload_size = put_bits_count(&payload_bits) / 8;
  542. flush_put_bits(&payload_bits);
  543. u(8, payload_type, last_payload_type_byte);
  544. u(8, payload_size, last_payload_size_byte);
  545. for (i = 0; i < payload_size; i++)
  546. u(8, payload[i] & 0xff, sei_payload);
  547. }
  548. vaapi_encode_h264_write_trailing_rbsp(pbc);
  549. }
  550. static int vaapi_encode_h264_write_sequence_header(AVCodecContext *avctx,
  551. char *data, size_t *data_len)
  552. {
  553. VAAPIEncodeContext *ctx = avctx->priv_data;
  554. PutBitContext pbc;
  555. char tmp[256];
  556. int err;
  557. size_t nal_len, bit_len, bit_pos, next_len;
  558. bit_len = *data_len;
  559. bit_pos = 0;
  560. init_put_bits(&pbc, tmp, sizeof(tmp));
  561. vaapi_encode_h264_write_sps(&pbc, ctx);
  562. nal_len = put_bits_count(&pbc);
  563. flush_put_bits(&pbc);
  564. next_len = bit_len - bit_pos;
  565. err = ff_vaapi_encode_h26x_nal_unit_to_byte_stream(data + bit_pos / 8,
  566. &next_len,
  567. tmp, nal_len);
  568. if (err < 0)
  569. return err;
  570. bit_pos += next_len;
  571. init_put_bits(&pbc, tmp, sizeof(tmp));
  572. vaapi_encode_h264_write_pps(&pbc, ctx);
  573. nal_len = put_bits_count(&pbc);
  574. flush_put_bits(&pbc);
  575. next_len = bit_len - bit_pos;
  576. err = ff_vaapi_encode_h26x_nal_unit_to_byte_stream(data + bit_pos / 8,
  577. &next_len,
  578. tmp, nal_len);
  579. if (err < 0)
  580. return err;
  581. bit_pos += next_len;
  582. *data_len = bit_pos;
  583. return 0;
  584. }
  585. static int vaapi_encode_h264_write_slice_header(AVCodecContext *avctx,
  586. VAAPIEncodePicture *pic,
  587. VAAPIEncodeSlice *slice,
  588. char *data, size_t *data_len)
  589. {
  590. VAAPIEncodeContext *ctx = avctx->priv_data;
  591. PutBitContext pbc;
  592. char tmp[256];
  593. size_t header_len;
  594. init_put_bits(&pbc, tmp, sizeof(tmp));
  595. vaapi_encode_h264_write_slice_header2(&pbc, ctx, pic, slice);
  596. header_len = put_bits_count(&pbc);
  597. flush_put_bits(&pbc);
  598. return ff_vaapi_encode_h26x_nal_unit_to_byte_stream(data, data_len,
  599. tmp, header_len);
  600. }
  601. static int vaapi_encode_h264_write_extra_header(AVCodecContext *avctx,
  602. VAAPIEncodePicture *pic,
  603. int index, int *type,
  604. char *data, size_t *data_len)
  605. {
  606. VAAPIEncodeContext *ctx = avctx->priv_data;
  607. PutBitContext pbc;
  608. char tmp[256];
  609. size_t header_len;
  610. if (index == 0 && ctx->va_rc_mode == VA_RC_CBR) {
  611. *type = VAEncPackedHeaderH264_SEI;
  612. init_put_bits(&pbc, tmp, sizeof(tmp));
  613. vaapi_encode_h264_write_sei(&pbc, ctx, pic);
  614. header_len = put_bits_count(&pbc);
  615. flush_put_bits(&pbc);
  616. return ff_vaapi_encode_h26x_nal_unit_to_byte_stream(data, data_len,
  617. tmp, header_len);
  618. } else {
  619. return AVERROR_EOF;
  620. }
  621. }
  622. static int vaapi_encode_h264_init_sequence_params(AVCodecContext *avctx)
  623. {
  624. VAAPIEncodeContext *ctx = avctx->priv_data;
  625. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  626. VAEncPictureParameterBufferH264 *vpic = ctx->codec_picture_params;
  627. VAAPIEncodeH264Context *priv = ctx->priv_data;
  628. VAAPIEncodeH264MiscSequenceParams *mseq = &priv->misc_sequence_params;
  629. int i;
  630. {
  631. vseq->seq_parameter_set_id = 0;
  632. vseq->level_idc = avctx->level;
  633. vseq->max_num_ref_frames = 2;
  634. vseq->picture_width_in_mbs = priv->mb_width;
  635. vseq->picture_height_in_mbs = priv->mb_height;
  636. vseq->seq_fields.bits.chroma_format_idc = 1;
  637. vseq->seq_fields.bits.frame_mbs_only_flag = 1;
  638. vseq->seq_fields.bits.direct_8x8_inference_flag = 1;
  639. vseq->seq_fields.bits.log2_max_frame_num_minus4 = 4;
  640. vseq->seq_fields.bits.pic_order_cnt_type = 0;
  641. if (ctx->input_width != ctx->aligned_width ||
  642. ctx->input_height != ctx->aligned_height) {
  643. vseq->frame_cropping_flag = 1;
  644. vseq->frame_crop_left_offset = 0;
  645. vseq->frame_crop_right_offset =
  646. (ctx->aligned_width - ctx->input_width) / 2;
  647. vseq->frame_crop_top_offset = 0;
  648. vseq->frame_crop_bottom_offset =
  649. (ctx->aligned_height - ctx->input_height) / 2;
  650. } else {
  651. vseq->frame_cropping_flag = 0;
  652. }
  653. vseq->vui_parameters_present_flag = 1;
  654. if (avctx->sample_aspect_ratio.num != 0) {
  655. vseq->vui_fields.bits.aspect_ratio_info_present_flag = 1;
  656. // There is a large enum of these which we could support
  657. // individually rather than using the generic X/Y form?
  658. if (avctx->sample_aspect_ratio.num ==
  659. avctx->sample_aspect_ratio.den) {
  660. vseq->aspect_ratio_idc = 1;
  661. } else {
  662. vseq->aspect_ratio_idc = 255; // Extended SAR.
  663. vseq->sar_width = avctx->sample_aspect_ratio.num;
  664. vseq->sar_height = avctx->sample_aspect_ratio.den;
  665. }
  666. }
  667. if (avctx->color_primaries != AVCOL_PRI_UNSPECIFIED ||
  668. avctx->color_trc != AVCOL_TRC_UNSPECIFIED ||
  669. avctx->colorspace != AVCOL_SPC_UNSPECIFIED) {
  670. mseq->video_signal_type_present_flag = 1;
  671. mseq->video_format = 5; // Unspecified.
  672. mseq->video_full_range_flag = 0;
  673. mseq->colour_description_present_flag = 1;
  674. // These enums are derived from the standard and hence
  675. // we can just use the values directly.
  676. mseq->colour_primaries = avctx->color_primaries;
  677. mseq->transfer_characteristics = avctx->color_trc;
  678. mseq->matrix_coefficients = avctx->colorspace;
  679. }
  680. vseq->bits_per_second = avctx->bit_rate;
  681. vseq->vui_fields.bits.timing_info_present_flag = 1;
  682. if (avctx->framerate.num > 0 && avctx->framerate.den > 0) {
  683. vseq->num_units_in_tick = avctx->framerate.num;
  684. vseq->time_scale = 2 * avctx->framerate.den;
  685. mseq->fixed_frame_rate_flag = 1;
  686. } else {
  687. vseq->num_units_in_tick = avctx->time_base.num;
  688. vseq->time_scale = 2 * avctx->time_base.den;
  689. mseq->fixed_frame_rate_flag = 0;
  690. }
  691. if (ctx->va_rc_mode == VA_RC_CBR) {
  692. priv->send_timing_sei = 1;
  693. mseq->nal_hrd_parameters_present_flag = 1;
  694. mseq->cpb_cnt_minus1 = 0;
  695. // Try to scale these to a sensible range so that the
  696. // golomb encode of the value is not overlong.
  697. mseq->bit_rate_scale =
  698. av_clip(av_log2(avctx->bit_rate) - 15, 0, 15);
  699. mseq->bit_rate_value_minus1[0] =
  700. (avctx->bit_rate >> mseq->bit_rate_scale) - 1;
  701. mseq->cpb_size_scale =
  702. av_clip(av_log2(priv->hrd_params.hrd.buffer_size) - 15, 0, 15);
  703. mseq->cpb_size_value_minus1[0] =
  704. (priv->hrd_params.hrd.buffer_size >> mseq->cpb_size_scale) - 1;
  705. // CBR mode isn't actually available here, despite naming.
  706. mseq->cbr_flag[0] = 0;
  707. mseq->initial_cpb_removal_delay_length_minus1 = 23;
  708. mseq->cpb_removal_delay_length_minus1 = 23;
  709. mseq->dpb_output_delay_length_minus1 = 7;
  710. mseq->time_offset_length = 0;
  711. // This calculation can easily overflow 32 bits.
  712. mseq->initial_cpb_removal_delay = 90000 *
  713. (uint64_t)priv->hrd_params.hrd.initial_buffer_fullness /
  714. priv->hrd_params.hrd.buffer_size;
  715. mseq->initial_cpb_removal_delay_offset = 0;
  716. } else {
  717. priv->send_timing_sei = 0;
  718. mseq->nal_hrd_parameters_present_flag = 0;
  719. }
  720. vseq->intra_period = ctx->p_per_i * (ctx->b_per_p + 1);
  721. vseq->intra_idr_period = vseq->intra_period;
  722. vseq->ip_period = ctx->b_per_p + 1;
  723. }
  724. {
  725. vpic->CurrPic.picture_id = VA_INVALID_ID;
  726. vpic->CurrPic.flags = VA_PICTURE_H264_INVALID;
  727. for (i = 0; i < FF_ARRAY_ELEMS(vpic->ReferenceFrames); i++) {
  728. vpic->ReferenceFrames[i].picture_id = VA_INVALID_ID;
  729. vpic->ReferenceFrames[i].flags = VA_PICTURE_H264_INVALID;
  730. }
  731. vpic->coded_buf = VA_INVALID_ID;
  732. vpic->pic_parameter_set_id = 0;
  733. vpic->seq_parameter_set_id = 0;
  734. vpic->num_ref_idx_l0_active_minus1 = 0;
  735. vpic->num_ref_idx_l1_active_minus1 = 0;
  736. vpic->pic_fields.bits.entropy_coding_mode_flag =
  737. ((avctx->profile & 0xff) != 66);
  738. vpic->pic_fields.bits.weighted_pred_flag = 0;
  739. vpic->pic_fields.bits.weighted_bipred_idc = 0;
  740. vpic->pic_fields.bits.transform_8x8_mode_flag =
  741. ((avctx->profile & 0xff) >= 100);
  742. vpic->pic_init_qp = priv->fixed_qp_idr;
  743. }
  744. {
  745. mseq->profile_idc = avctx->profile & 0xff;
  746. if (avctx->profile & FF_PROFILE_H264_CONSTRAINED)
  747. mseq->constraint_set1_flag = 1;
  748. if (avctx->profile & FF_PROFILE_H264_INTRA)
  749. mseq->constraint_set3_flag = 1;
  750. }
  751. return 0;
  752. }
  753. static int vaapi_encode_h264_init_picture_params(AVCodecContext *avctx,
  754. VAAPIEncodePicture *pic)
  755. {
  756. VAAPIEncodeContext *ctx = avctx->priv_data;
  757. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  758. VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params;
  759. VAAPIEncodeH264Context *priv = ctx->priv_data;
  760. int i;
  761. if (pic->type == PICTURE_TYPE_IDR) {
  762. av_assert0(pic->display_order == pic->encode_order);
  763. vpic->frame_num = 0;
  764. priv->next_frame_num = 1;
  765. priv->cpb_delay = 0;
  766. } else {
  767. vpic->frame_num = priv->next_frame_num;
  768. if (pic->type != PICTURE_TYPE_B) {
  769. // nal_ref_idc != 0
  770. ++priv->next_frame_num;
  771. }
  772. ++priv->cpb_delay;
  773. }
  774. priv->dpb_delay = pic->display_order - pic->encode_order + 1;
  775. vpic->frame_num = vpic->frame_num &
  776. ((1 << (4 + vseq->seq_fields.bits.log2_max_frame_num_minus4)) - 1);
  777. vpic->CurrPic.picture_id = pic->recon_surface;
  778. vpic->CurrPic.frame_idx = vpic->frame_num;
  779. vpic->CurrPic.flags = 0;
  780. vpic->CurrPic.TopFieldOrderCnt = pic->display_order;
  781. vpic->CurrPic.BottomFieldOrderCnt = pic->display_order;
  782. for (i = 0; i < pic->nb_refs; i++) {
  783. VAAPIEncodePicture *ref = pic->refs[i];
  784. av_assert0(ref && ref->encode_order < pic->encode_order);
  785. vpic->ReferenceFrames[i].picture_id = ref->recon_surface;
  786. vpic->ReferenceFrames[i].frame_idx = ref->encode_order;
  787. vpic->ReferenceFrames[i].flags = VA_PICTURE_H264_SHORT_TERM_REFERENCE;
  788. vpic->ReferenceFrames[i].TopFieldOrderCnt = ref->display_order;
  789. vpic->ReferenceFrames[i].BottomFieldOrderCnt = ref->display_order;
  790. }
  791. for (; i < FF_ARRAY_ELEMS(vpic->ReferenceFrames); i++) {
  792. vpic->ReferenceFrames[i].picture_id = VA_INVALID_ID;
  793. vpic->ReferenceFrames[i].flags = VA_PICTURE_H264_INVALID;
  794. }
  795. vpic->coded_buf = pic->output_buffer;
  796. vpic->pic_fields.bits.idr_pic_flag = (pic->type == PICTURE_TYPE_IDR);
  797. vpic->pic_fields.bits.reference_pic_flag = (pic->type != PICTURE_TYPE_B);
  798. pic->nb_slices = 1;
  799. return 0;
  800. }
  801. static int vaapi_encode_h264_init_slice_params(AVCodecContext *avctx,
  802. VAAPIEncodePicture *pic,
  803. VAAPIEncodeSlice *slice)
  804. {
  805. VAAPIEncodeContext *ctx = avctx->priv_data;
  806. VAEncSequenceParameterBufferH264 *vseq = ctx->codec_sequence_params;
  807. VAEncPictureParameterBufferH264 *vpic = pic->codec_picture_params;
  808. VAEncSliceParameterBufferH264 *vslice = slice->codec_slice_params;
  809. VAAPIEncodeH264Context *priv = ctx->priv_data;
  810. VAAPIEncodeH264Slice *pslice;
  811. VAAPIEncodeH264MiscSliceParams *mslice;
  812. int i;
  813. slice->priv_data = av_mallocz(sizeof(*pslice));
  814. if (!slice->priv_data)
  815. return AVERROR(ENOMEM);
  816. pslice = slice->priv_data;
  817. mslice = &pslice->misc_slice_params;
  818. if (pic->type == PICTURE_TYPE_IDR)
  819. mslice->nal_unit_type = NAL_IDR_SLICE;
  820. else
  821. mslice->nal_unit_type = NAL_SLICE;
  822. switch (pic->type) {
  823. case PICTURE_TYPE_IDR:
  824. vslice->slice_type = SLICE_TYPE_I;
  825. mslice->nal_ref_idc = 3;
  826. break;
  827. case PICTURE_TYPE_I:
  828. vslice->slice_type = SLICE_TYPE_I;
  829. mslice->nal_ref_idc = 2;
  830. break;
  831. case PICTURE_TYPE_P:
  832. vslice->slice_type = SLICE_TYPE_P;
  833. mslice->nal_ref_idc = 1;
  834. break;
  835. case PICTURE_TYPE_B:
  836. vslice->slice_type = SLICE_TYPE_B;
  837. mslice->nal_ref_idc = 0;
  838. break;
  839. default:
  840. av_assert0(0 && "invalid picture type");
  841. }
  842. // Only one slice per frame.
  843. vslice->macroblock_address = 0;
  844. vslice->num_macroblocks = priv->mb_width * priv->mb_height;
  845. vslice->macroblock_info = VA_INVALID_ID;
  846. vslice->pic_parameter_set_id = vpic->pic_parameter_set_id;
  847. vslice->idr_pic_id = priv->idr_pic_count++;
  848. vslice->pic_order_cnt_lsb = pic->display_order &
  849. ((1 << (4 + vseq->seq_fields.bits.log2_max_pic_order_cnt_lsb_minus4)) - 1);
  850. for (i = 0; i < FF_ARRAY_ELEMS(vslice->RefPicList0); i++) {
  851. vslice->RefPicList0[i].picture_id = VA_INVALID_ID;
  852. vslice->RefPicList0[i].flags = VA_PICTURE_H264_INVALID;
  853. vslice->RefPicList1[i].picture_id = VA_INVALID_ID;
  854. vslice->RefPicList1[i].flags = VA_PICTURE_H264_INVALID;
  855. }
  856. av_assert0(pic->nb_refs <= 2);
  857. if (pic->nb_refs >= 1) {
  858. // Backward reference for P- or B-frame.
  859. av_assert0(pic->type == PICTURE_TYPE_P ||
  860. pic->type == PICTURE_TYPE_B);
  861. vslice->num_ref_idx_l0_active_minus1 = 0;
  862. vslice->RefPicList0[0] = vpic->ReferenceFrames[0];
  863. }
  864. if (pic->nb_refs >= 2) {
  865. // Forward reference for B-frame.
  866. av_assert0(pic->type == PICTURE_TYPE_B);
  867. vslice->num_ref_idx_l1_active_minus1 = 0;
  868. vslice->RefPicList1[0] = vpic->ReferenceFrames[1];
  869. }
  870. if (pic->type == PICTURE_TYPE_B)
  871. vslice->slice_qp_delta = priv->fixed_qp_b - vpic->pic_init_qp;
  872. else if (pic->type == PICTURE_TYPE_P)
  873. vslice->slice_qp_delta = priv->fixed_qp_p - vpic->pic_init_qp;
  874. else
  875. vslice->slice_qp_delta = priv->fixed_qp_idr - vpic->pic_init_qp;
  876. vslice->direct_spatial_mv_pred_flag = 1;
  877. return 0;
  878. }
  879. static av_cold int vaapi_encode_h264_init_constant_bitrate(AVCodecContext *avctx)
  880. {
  881. VAAPIEncodeContext *ctx = avctx->priv_data;
  882. VAAPIEncodeH264Context *priv = ctx->priv_data;
  883. int hrd_buffer_size;
  884. int hrd_initial_buffer_fullness;
  885. if (avctx->rc_buffer_size)
  886. hrd_buffer_size = avctx->rc_buffer_size;
  887. else
  888. hrd_buffer_size = avctx->bit_rate;
  889. if (avctx->rc_initial_buffer_occupancy)
  890. hrd_initial_buffer_fullness = avctx->rc_initial_buffer_occupancy;
  891. else
  892. hrd_initial_buffer_fullness = hrd_buffer_size * 3 / 4;
  893. priv->rc_params.misc.type = VAEncMiscParameterTypeRateControl;
  894. priv->rc_params.rc = (VAEncMiscParameterRateControl) {
  895. .bits_per_second = avctx->bit_rate,
  896. .target_percentage = 66,
  897. .window_size = 1000,
  898. .initial_qp = (avctx->qmax >= 0 ? avctx->qmax : 40),
  899. .min_qp = (avctx->qmin >= 0 ? avctx->qmin : 18),
  900. .basic_unit_size = 0,
  901. };
  902. ctx->global_params[ctx->nb_global_params] =
  903. &priv->rc_params.misc;
  904. ctx->global_params_size[ctx->nb_global_params++] =
  905. sizeof(priv->rc_params);
  906. priv->hrd_params.misc.type = VAEncMiscParameterTypeHRD;
  907. priv->hrd_params.hrd = (VAEncMiscParameterHRD) {
  908. .initial_buffer_fullness = hrd_initial_buffer_fullness,
  909. .buffer_size = hrd_buffer_size,
  910. };
  911. ctx->global_params[ctx->nb_global_params] =
  912. &priv->hrd_params.misc;
  913. ctx->global_params_size[ctx->nb_global_params++] =
  914. sizeof(priv->hrd_params);
  915. // These still need to be set for pic_init_qp/slice_qp_delta.
  916. priv->fixed_qp_idr = 26;
  917. priv->fixed_qp_p = 26;
  918. priv->fixed_qp_b = 26;
  919. av_log(avctx, AV_LOG_DEBUG, "Using constant-bitrate = %d bps.\n",
  920. avctx->bit_rate);
  921. return 0;
  922. }
  923. static av_cold int vaapi_encode_h264_init_fixed_qp(AVCodecContext *avctx)
  924. {
  925. VAAPIEncodeContext *ctx = avctx->priv_data;
  926. VAAPIEncodeH264Context *priv = ctx->priv_data;
  927. VAAPIEncodeH264Options *opt = ctx->codec_options;
  928. priv->fixed_qp_p = opt->qp;
  929. if (avctx->i_quant_factor > 0.0)
  930. priv->fixed_qp_idr = (int)((priv->fixed_qp_p * avctx->i_quant_factor +
  931. avctx->i_quant_offset) + 0.5);
  932. else
  933. priv->fixed_qp_idr = priv->fixed_qp_p;
  934. if (avctx->b_quant_factor > 0.0)
  935. priv->fixed_qp_b = (int)((priv->fixed_qp_p * avctx->b_quant_factor +
  936. avctx->b_quant_offset) + 0.5);
  937. else
  938. priv->fixed_qp_b = priv->fixed_qp_p;
  939. av_log(avctx, AV_LOG_DEBUG, "Using fixed QP = "
  940. "%d / %d / %d for IDR- / P- / B-frames.\n",
  941. priv->fixed_qp_idr, priv->fixed_qp_p, priv->fixed_qp_b);
  942. return 0;
  943. }
  944. static av_cold int vaapi_encode_h264_init_internal(AVCodecContext *avctx)
  945. {
  946. static const VAConfigAttrib default_config_attributes[] = {
  947. { .type = VAConfigAttribRTFormat,
  948. .value = VA_RT_FORMAT_YUV420 },
  949. { .type = VAConfigAttribEncPackedHeaders,
  950. .value = (VA_ENC_PACKED_HEADER_SEQUENCE |
  951. VA_ENC_PACKED_HEADER_SLICE) },
  952. };
  953. VAAPIEncodeContext *ctx = avctx->priv_data;
  954. VAAPIEncodeH264Context *priv = ctx->priv_data;
  955. VAAPIEncodeH264Options *opt = ctx->codec_options;
  956. int i, err;
  957. switch (avctx->profile) {
  958. case FF_PROFILE_H264_CONSTRAINED_BASELINE:
  959. ctx->va_profile = VAProfileH264ConstrainedBaseline;
  960. break;
  961. case FF_PROFILE_H264_BASELINE:
  962. ctx->va_profile = VAProfileH264Baseline;
  963. break;
  964. case FF_PROFILE_H264_MAIN:
  965. ctx->va_profile = VAProfileH264Main;
  966. break;
  967. case FF_PROFILE_H264_EXTENDED:
  968. av_log(avctx, AV_LOG_ERROR, "H.264 extended profile "
  969. "is not supported.\n");
  970. return AVERROR_PATCHWELCOME;
  971. case FF_PROFILE_UNKNOWN:
  972. case FF_PROFILE_H264_HIGH:
  973. ctx->va_profile = VAProfileH264High;
  974. break;
  975. case FF_PROFILE_H264_HIGH_10:
  976. case FF_PROFILE_H264_HIGH_10_INTRA:
  977. av_log(avctx, AV_LOG_ERROR, "H.264 10-bit profiles "
  978. "are not supported.\n");
  979. return AVERROR_PATCHWELCOME;
  980. case FF_PROFILE_H264_HIGH_422:
  981. case FF_PROFILE_H264_HIGH_422_INTRA:
  982. case FF_PROFILE_H264_HIGH_444:
  983. case FF_PROFILE_H264_HIGH_444_PREDICTIVE:
  984. case FF_PROFILE_H264_HIGH_444_INTRA:
  985. case FF_PROFILE_H264_CAVLC_444:
  986. av_log(avctx, AV_LOG_ERROR, "H.264 non-4:2:0 profiles "
  987. "are not supported.\n");
  988. return AVERROR_PATCHWELCOME;
  989. default:
  990. av_log(avctx, AV_LOG_ERROR, "Unknown H.264 profile %d.\n",
  991. avctx->profile);
  992. return AVERROR(EINVAL);
  993. }
  994. if (opt->low_power) {
  995. #if VA_CHECK_VERSION(0, 39, 1)
  996. ctx->va_entrypoint = VAEntrypointEncSliceLP;
  997. #else
  998. av_log(avctx, AV_LOG_ERROR, "Low-power encoding is not "
  999. "supported with this VAAPI version.\n");
  1000. return AVERROR(EINVAL);
  1001. #endif
  1002. } else {
  1003. ctx->va_entrypoint = VAEntrypointEncSlice;
  1004. }
  1005. ctx->input_width = avctx->width;
  1006. ctx->input_height = avctx->height;
  1007. ctx->aligned_width = FFALIGN(ctx->input_width, 16);
  1008. ctx->aligned_height = FFALIGN(ctx->input_height, 16);
  1009. priv->mb_width = ctx->aligned_width / 16;
  1010. priv->mb_height = ctx->aligned_height / 16;
  1011. for (i = 0; i < FF_ARRAY_ELEMS(default_config_attributes); i++) {
  1012. ctx->config_attributes[ctx->nb_config_attributes++] =
  1013. default_config_attributes[i];
  1014. }
  1015. if (avctx->bit_rate > 0) {
  1016. ctx->va_rc_mode = VA_RC_CBR;
  1017. err = vaapi_encode_h264_init_constant_bitrate(avctx);
  1018. } else {
  1019. ctx->va_rc_mode = VA_RC_CQP;
  1020. err = vaapi_encode_h264_init_fixed_qp(avctx);
  1021. }
  1022. if (err < 0)
  1023. return err;
  1024. ctx->config_attributes[ctx->nb_config_attributes++] = (VAConfigAttrib) {
  1025. .type = VAConfigAttribRateControl,
  1026. .value = ctx->va_rc_mode,
  1027. };
  1028. if (opt->quality > 0) {
  1029. #if VA_CHECK_VERSION(0, 36, 0)
  1030. priv->quality_params.misc.type =
  1031. VAEncMiscParameterTypeQualityLevel;
  1032. priv->quality_params.quality.quality_level = opt->quality;
  1033. ctx->global_params[ctx->nb_global_params] =
  1034. &priv->quality_params.misc;
  1035. ctx->global_params_size[ctx->nb_global_params++] =
  1036. sizeof(priv->quality_params);
  1037. #else
  1038. av_log(avctx, AV_LOG_WARNING, "The encode quality option is not "
  1039. "supported with this VAAPI version.\n");
  1040. #endif
  1041. }
  1042. ctx->nb_recon_frames = 20;
  1043. return 0;
  1044. }
  1045. static VAAPIEncodeType vaapi_encode_type_h264 = {
  1046. .priv_data_size = sizeof(VAAPIEncodeH264Context),
  1047. .init = &vaapi_encode_h264_init_internal,
  1048. .sequence_params_size = sizeof(VAEncSequenceParameterBufferH264),
  1049. .init_sequence_params = &vaapi_encode_h264_init_sequence_params,
  1050. .picture_params_size = sizeof(VAEncPictureParameterBufferH264),
  1051. .init_picture_params = &vaapi_encode_h264_init_picture_params,
  1052. .slice_params_size = sizeof(VAEncSliceParameterBufferH264),
  1053. .init_slice_params = &vaapi_encode_h264_init_slice_params,
  1054. .sequence_header_type = VAEncPackedHeaderSequence,
  1055. .write_sequence_header = &vaapi_encode_h264_write_sequence_header,
  1056. .slice_header_type = VAEncPackedHeaderH264_Slice,
  1057. .write_slice_header = &vaapi_encode_h264_write_slice_header,
  1058. .write_extra_header = &vaapi_encode_h264_write_extra_header,
  1059. };
  1060. static av_cold int vaapi_encode_h264_init(AVCodecContext *avctx)
  1061. {
  1062. return ff_vaapi_encode_init(avctx, &vaapi_encode_type_h264);
  1063. }
  1064. #define OFFSET(x) (offsetof(VAAPIEncodeContext, codec_options_data) + \
  1065. offsetof(VAAPIEncodeH264Options, x))
  1066. #define FLAGS (AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM)
  1067. static const AVOption vaapi_encode_h264_options[] = {
  1068. { "qp", "Constant QP (for P-frames; scaled by qfactor/qoffset for I/B)",
  1069. OFFSET(qp), AV_OPT_TYPE_INT, { .i64 = 20 }, 0, 52, FLAGS },
  1070. { "quality", "Set encode quality (trades off against speed, higher is faster)",
  1071. OFFSET(quality), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 8, FLAGS },
  1072. { "low_power", "Use low-power encoding mode (experimental: only supported "
  1073. "on some platforms, does not support all features)",
  1074. OFFSET(low_power), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, 1, FLAGS },
  1075. { NULL },
  1076. };
  1077. static const AVCodecDefault vaapi_encode_h264_defaults[] = {
  1078. { "profile", "100" },
  1079. { "level", "51" },
  1080. { "b", "0" },
  1081. { "bf", "2" },
  1082. { "g", "120" },
  1083. { "i_qfactor", "1.0" },
  1084. { "i_qoffset", "0.0" },
  1085. { "b_qfactor", "1.2" },
  1086. { "b_qoffset", "0.0" },
  1087. { NULL },
  1088. };
  1089. static const AVClass vaapi_encode_h264_class = {
  1090. .class_name = "h264_vaapi",
  1091. .item_name = av_default_item_name,
  1092. .option = vaapi_encode_h264_options,
  1093. .version = LIBAVUTIL_VERSION_INT,
  1094. };
  1095. AVCodec ff_h264_vaapi_encoder = {
  1096. .name = "h264_vaapi",
  1097. .long_name = NULL_IF_CONFIG_SMALL("H.264/AVC (VAAPI)"),
  1098. .type = AVMEDIA_TYPE_VIDEO,
  1099. .id = AV_CODEC_ID_H264,
  1100. .priv_data_size = (sizeof(VAAPIEncodeContext) +
  1101. sizeof(VAAPIEncodeH264Options)),
  1102. .init = &vaapi_encode_h264_init,
  1103. .encode2 = &ff_vaapi_encode2,
  1104. .close = &ff_vaapi_encode_close,
  1105. .priv_class = &vaapi_encode_h264_class,
  1106. .capabilities = AV_CODEC_CAP_DELAY,
  1107. .defaults = vaapi_encode_h264_defaults,
  1108. .pix_fmts = (const enum AVPixelFormat[]) {
  1109. AV_PIX_FMT_VAAPI,
  1110. AV_PIX_FMT_NONE,
  1111. },
  1112. };