You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1758 lines
66KB

  1. /*
  2. * HEVC Parameter Set decoding
  3. *
  4. * Copyright (C) 2012 - 2103 Guillaume Martres
  5. * Copyright (C) 2012 - 2103 Mickael Raulet
  6. * Copyright (C) 2012 - 2013 Gildas Cocherel
  7. * Copyright (C) 2013 Vittorio Giovara
  8. *
  9. * This file is part of FFmpeg.
  10. *
  11. * FFmpeg is free software; you can redistribute it and/or
  12. * modify it under the terms of the GNU Lesser General Public
  13. * License as published by the Free Software Foundation; either
  14. * version 2.1 of the License, or (at your option) any later version.
  15. *
  16. * FFmpeg is distributed in the hope that it will be useful,
  17. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  18. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  19. * Lesser General Public License for more details.
  20. *
  21. * You should have received a copy of the GNU Lesser General Public
  22. * License along with FFmpeg; if not, write to the Free Software
  23. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  24. */
  25. #include "libavutil/imgutils.h"
  26. #include "golomb.h"
  27. #include "hevc_data.h"
  28. #include "hevc_ps.h"
  29. static const uint8_t default_scaling_list_intra[] = {
  30. 16, 16, 16, 16, 17, 18, 21, 24,
  31. 16, 16, 16, 16, 17, 19, 22, 25,
  32. 16, 16, 17, 18, 20, 22, 25, 29,
  33. 16, 16, 18, 21, 24, 27, 31, 36,
  34. 17, 17, 20, 24, 30, 35, 41, 47,
  35. 18, 19, 22, 27, 35, 44, 54, 65,
  36. 21, 22, 25, 31, 41, 54, 70, 88,
  37. 24, 25, 29, 36, 47, 65, 88, 115
  38. };
  39. static const uint8_t default_scaling_list_inter[] = {
  40. 16, 16, 16, 16, 17, 18, 20, 24,
  41. 16, 16, 16, 17, 18, 20, 24, 25,
  42. 16, 16, 17, 18, 20, 24, 25, 28,
  43. 16, 17, 18, 20, 24, 25, 28, 33,
  44. 17, 18, 20, 24, 25, 28, 33, 41,
  45. 18, 20, 24, 25, 28, 33, 41, 54,
  46. 20, 24, 25, 28, 33, 41, 54, 71,
  47. 24, 25, 28, 33, 41, 54, 71, 91
  48. };
  49. static const AVRational vui_sar[] = {
  50. { 0, 1 },
  51. { 1, 1 },
  52. { 12, 11 },
  53. { 10, 11 },
  54. { 16, 11 },
  55. { 40, 33 },
  56. { 24, 11 },
  57. { 20, 11 },
  58. { 32, 11 },
  59. { 80, 33 },
  60. { 18, 11 },
  61. { 15, 11 },
  62. { 64, 33 },
  63. { 160, 99 },
  64. { 4, 3 },
  65. { 3, 2 },
  66. { 2, 1 },
  67. };
  68. static const uint8_t hevc_sub_width_c[] = {
  69. 1, 2, 2, 1
  70. };
  71. static const uint8_t hevc_sub_height_c[] = {
  72. 1, 2, 1, 1
  73. };
  74. static void remove_pps(HEVCParamSets *s, int id)
  75. {
  76. if (s->pps_list[id] && s->pps == (const HEVCPPS*)s->pps_list[id]->data)
  77. s->pps = NULL;
  78. av_buffer_unref(&s->pps_list[id]);
  79. }
  80. static void remove_sps(HEVCParamSets *s, int id)
  81. {
  82. int i;
  83. if (s->sps_list[id]) {
  84. if (s->sps == (const HEVCSPS*)s->sps_list[id]->data)
  85. s->sps = NULL;
  86. /* drop all PPS that depend on this SPS */
  87. for (i = 0; i < FF_ARRAY_ELEMS(s->pps_list); i++)
  88. if (s->pps_list[i] && ((HEVCPPS*)s->pps_list[i]->data)->sps_id == id)
  89. remove_pps(s, i);
  90. av_assert0(!(s->sps_list[id] && s->sps == (HEVCSPS*)s->sps_list[id]->data));
  91. }
  92. av_buffer_unref(&s->sps_list[id]);
  93. }
  94. static void remove_vps(HEVCParamSets *s, int id)
  95. {
  96. int i;
  97. if (s->vps_list[id]) {
  98. if (s->vps == (const HEVCVPS*)s->vps_list[id]->data)
  99. s->vps = NULL;
  100. for (i = 0; i < FF_ARRAY_ELEMS(s->sps_list); i++)
  101. if (s->sps_list[i] && ((HEVCSPS*)s->sps_list[i]->data)->vps_id == id)
  102. remove_sps(s, i);
  103. }
  104. av_buffer_unref(&s->vps_list[id]);
  105. }
  106. int ff_hevc_decode_short_term_rps(GetBitContext *gb, AVCodecContext *avctx,
  107. ShortTermRPS *rps, const HEVCSPS *sps, int is_slice_header)
  108. {
  109. uint8_t rps_predict = 0;
  110. int delta_poc;
  111. int k0 = 0;
  112. int k1 = 0;
  113. int k = 0;
  114. int i;
  115. if (rps != sps->st_rps && sps->nb_st_rps)
  116. rps_predict = get_bits1(gb);
  117. if (rps_predict) {
  118. const ShortTermRPS *rps_ridx;
  119. int delta_rps;
  120. unsigned abs_delta_rps;
  121. uint8_t use_delta_flag = 0;
  122. uint8_t delta_rps_sign;
  123. if (is_slice_header) {
  124. unsigned int delta_idx = get_ue_golomb_long(gb) + 1;
  125. if (delta_idx > sps->nb_st_rps) {
  126. av_log(avctx, AV_LOG_ERROR,
  127. "Invalid value of delta_idx in slice header RPS: %d > %d.\n",
  128. delta_idx, sps->nb_st_rps);
  129. return AVERROR_INVALIDDATA;
  130. }
  131. rps_ridx = &sps->st_rps[sps->nb_st_rps - delta_idx];
  132. rps->rps_idx_num_delta_pocs = rps_ridx->num_delta_pocs;
  133. } else
  134. rps_ridx = &sps->st_rps[rps - sps->st_rps - 1];
  135. delta_rps_sign = get_bits1(gb);
  136. abs_delta_rps = get_ue_golomb_long(gb) + 1;
  137. if (abs_delta_rps < 1 || abs_delta_rps > 32768) {
  138. av_log(avctx, AV_LOG_ERROR,
  139. "Invalid value of abs_delta_rps: %d\n",
  140. abs_delta_rps);
  141. return AVERROR_INVALIDDATA;
  142. }
  143. delta_rps = (1 - (delta_rps_sign << 1)) * abs_delta_rps;
  144. for (i = 0; i <= rps_ridx->num_delta_pocs; i++) {
  145. int used = rps->used[k] = get_bits1(gb);
  146. if (!used)
  147. use_delta_flag = get_bits1(gb);
  148. if (used || use_delta_flag) {
  149. if (i < rps_ridx->num_delta_pocs)
  150. delta_poc = delta_rps + rps_ridx->delta_poc[i];
  151. else
  152. delta_poc = delta_rps;
  153. rps->delta_poc[k] = delta_poc;
  154. if (delta_poc < 0)
  155. k0++;
  156. else
  157. k1++;
  158. k++;
  159. }
  160. }
  161. if (k >= FF_ARRAY_ELEMS(rps->used)) {
  162. av_log(avctx, AV_LOG_ERROR,
  163. "Invalid num_delta_pocs: %d\n", k);
  164. return AVERROR_INVALIDDATA;
  165. }
  166. rps->num_delta_pocs = k;
  167. rps->num_negative_pics = k0;
  168. // sort in increasing order (smallest first)
  169. if (rps->num_delta_pocs != 0) {
  170. int used, tmp;
  171. for (i = 1; i < rps->num_delta_pocs; i++) {
  172. delta_poc = rps->delta_poc[i];
  173. used = rps->used[i];
  174. for (k = i - 1; k >= 0; k--) {
  175. tmp = rps->delta_poc[k];
  176. if (delta_poc < tmp) {
  177. rps->delta_poc[k + 1] = tmp;
  178. rps->used[k + 1] = rps->used[k];
  179. rps->delta_poc[k] = delta_poc;
  180. rps->used[k] = used;
  181. }
  182. }
  183. }
  184. }
  185. if ((rps->num_negative_pics >> 1) != 0) {
  186. int used;
  187. k = rps->num_negative_pics - 1;
  188. // flip the negative values to largest first
  189. for (i = 0; i < rps->num_negative_pics >> 1; i++) {
  190. delta_poc = rps->delta_poc[i];
  191. used = rps->used[i];
  192. rps->delta_poc[i] = rps->delta_poc[k];
  193. rps->used[i] = rps->used[k];
  194. rps->delta_poc[k] = delta_poc;
  195. rps->used[k] = used;
  196. k--;
  197. }
  198. }
  199. } else {
  200. unsigned int prev, nb_positive_pics;
  201. rps->num_negative_pics = get_ue_golomb_long(gb);
  202. nb_positive_pics = get_ue_golomb_long(gb);
  203. if (rps->num_negative_pics >= HEVC_MAX_REFS ||
  204. nb_positive_pics >= HEVC_MAX_REFS) {
  205. av_log(avctx, AV_LOG_ERROR, "Too many refs in a short term RPS.\n");
  206. return AVERROR_INVALIDDATA;
  207. }
  208. rps->num_delta_pocs = rps->num_negative_pics + nb_positive_pics;
  209. if (rps->num_delta_pocs) {
  210. prev = 0;
  211. for (i = 0; i < rps->num_negative_pics; i++) {
  212. delta_poc = get_ue_golomb_long(gb) + 1;
  213. if (delta_poc < 1 || delta_poc > 32768) {
  214. av_log(avctx, AV_LOG_ERROR,
  215. "Invalid value of delta_poc: %d\n",
  216. delta_poc);
  217. return AVERROR_INVALIDDATA;
  218. }
  219. prev -= delta_poc;
  220. rps->delta_poc[i] = prev;
  221. rps->used[i] = get_bits1(gb);
  222. }
  223. prev = 0;
  224. for (i = 0; i < nb_positive_pics; i++) {
  225. delta_poc = get_ue_golomb_long(gb) + 1;
  226. if (delta_poc < 1 || delta_poc > 32768) {
  227. av_log(avctx, AV_LOG_ERROR,
  228. "Invalid value of delta_poc: %d\n",
  229. delta_poc);
  230. return AVERROR_INVALIDDATA;
  231. }
  232. prev += delta_poc;
  233. rps->delta_poc[rps->num_negative_pics + i] = prev;
  234. rps->used[rps->num_negative_pics + i] = get_bits1(gb);
  235. }
  236. }
  237. }
  238. return 0;
  239. }
  240. static int decode_profile_tier_level(GetBitContext *gb, AVCodecContext *avctx,
  241. PTLCommon *ptl)
  242. {
  243. int i;
  244. if (get_bits_left(gb) < 2+1+5 + 32 + 4 + 16 + 16 + 12)
  245. return -1;
  246. ptl->profile_space = get_bits(gb, 2);
  247. ptl->tier_flag = get_bits1(gb);
  248. ptl->profile_idc = get_bits(gb, 5);
  249. if (ptl->profile_idc == FF_PROFILE_HEVC_MAIN)
  250. av_log(avctx, AV_LOG_DEBUG, "Main profile bitstream\n");
  251. else if (ptl->profile_idc == FF_PROFILE_HEVC_MAIN_10)
  252. av_log(avctx, AV_LOG_DEBUG, "Main 10 profile bitstream\n");
  253. else if (ptl->profile_idc == FF_PROFILE_HEVC_MAIN_STILL_PICTURE)
  254. av_log(avctx, AV_LOG_DEBUG, "Main Still Picture profile bitstream\n");
  255. else if (ptl->profile_idc == FF_PROFILE_HEVC_REXT)
  256. av_log(avctx, AV_LOG_DEBUG, "Range Extension profile bitstream\n");
  257. else
  258. av_log(avctx, AV_LOG_WARNING, "Unknown HEVC profile: %d\n", ptl->profile_idc);
  259. for (i = 0; i < 32; i++) {
  260. ptl->profile_compatibility_flag[i] = get_bits1(gb);
  261. if (ptl->profile_idc == 0 && i > 0 && ptl->profile_compatibility_flag[i])
  262. ptl->profile_idc = i;
  263. }
  264. ptl->progressive_source_flag = get_bits1(gb);
  265. ptl->interlaced_source_flag = get_bits1(gb);
  266. ptl->non_packed_constraint_flag = get_bits1(gb);
  267. ptl->frame_only_constraint_flag = get_bits1(gb);
  268. skip_bits(gb, 16); // XXX_reserved_zero_44bits[0..15]
  269. skip_bits(gb, 16); // XXX_reserved_zero_44bits[16..31]
  270. skip_bits(gb, 12); // XXX_reserved_zero_44bits[32..43]
  271. return 0;
  272. }
  273. static int parse_ptl(GetBitContext *gb, AVCodecContext *avctx,
  274. PTL *ptl, int max_num_sub_layers)
  275. {
  276. int i;
  277. if (decode_profile_tier_level(gb, avctx, &ptl->general_ptl) < 0 ||
  278. get_bits_left(gb) < 8 + (8*2 * (max_num_sub_layers - 1 > 0))) {
  279. av_log(avctx, AV_LOG_ERROR, "PTL information too short\n");
  280. return -1;
  281. }
  282. ptl->general_ptl.level_idc = get_bits(gb, 8);
  283. for (i = 0; i < max_num_sub_layers - 1; i++) {
  284. ptl->sub_layer_profile_present_flag[i] = get_bits1(gb);
  285. ptl->sub_layer_level_present_flag[i] = get_bits1(gb);
  286. }
  287. if (max_num_sub_layers - 1> 0)
  288. for (i = max_num_sub_layers - 1; i < 8; i++)
  289. skip_bits(gb, 2); // reserved_zero_2bits[i]
  290. for (i = 0; i < max_num_sub_layers - 1; i++) {
  291. if (ptl->sub_layer_profile_present_flag[i] &&
  292. decode_profile_tier_level(gb, avctx, &ptl->sub_layer_ptl[i]) < 0) {
  293. av_log(avctx, AV_LOG_ERROR,
  294. "PTL information for sublayer %i too short\n", i);
  295. return -1;
  296. }
  297. if (ptl->sub_layer_level_present_flag[i]) {
  298. if (get_bits_left(gb) < 8) {
  299. av_log(avctx, AV_LOG_ERROR,
  300. "Not enough data for sublayer %i level_idc\n", i);
  301. return -1;
  302. } else
  303. ptl->sub_layer_ptl[i].level_idc = get_bits(gb, 8);
  304. }
  305. }
  306. return 0;
  307. }
  308. static void decode_sublayer_hrd(GetBitContext *gb, unsigned int nb_cpb,
  309. int subpic_params_present)
  310. {
  311. int i;
  312. for (i = 0; i < nb_cpb; i++) {
  313. get_ue_golomb_long(gb); // bit_rate_value_minus1
  314. get_ue_golomb_long(gb); // cpb_size_value_minus1
  315. if (subpic_params_present) {
  316. get_ue_golomb_long(gb); // cpb_size_du_value_minus1
  317. get_ue_golomb_long(gb); // bit_rate_du_value_minus1
  318. }
  319. skip_bits1(gb); // cbr_flag
  320. }
  321. }
  322. static int decode_hrd(GetBitContext *gb, int common_inf_present,
  323. int max_sublayers)
  324. {
  325. int nal_params_present = 0, vcl_params_present = 0;
  326. int subpic_params_present = 0;
  327. int i;
  328. if (common_inf_present) {
  329. nal_params_present = get_bits1(gb);
  330. vcl_params_present = get_bits1(gb);
  331. if (nal_params_present || vcl_params_present) {
  332. subpic_params_present = get_bits1(gb);
  333. if (subpic_params_present) {
  334. skip_bits(gb, 8); // tick_divisor_minus2
  335. skip_bits(gb, 5); // du_cpb_removal_delay_increment_length_minus1
  336. skip_bits(gb, 1); // sub_pic_cpb_params_in_pic_timing_sei_flag
  337. skip_bits(gb, 5); // dpb_output_delay_du_length_minus1
  338. }
  339. skip_bits(gb, 4); // bit_rate_scale
  340. skip_bits(gb, 4); // cpb_size_scale
  341. if (subpic_params_present)
  342. skip_bits(gb, 4); // cpb_size_du_scale
  343. skip_bits(gb, 5); // initial_cpb_removal_delay_length_minus1
  344. skip_bits(gb, 5); // au_cpb_removal_delay_length_minus1
  345. skip_bits(gb, 5); // dpb_output_delay_length_minus1
  346. }
  347. }
  348. for (i = 0; i < max_sublayers; i++) {
  349. int low_delay = 0;
  350. unsigned int nb_cpb = 1;
  351. int fixed_rate = get_bits1(gb);
  352. if (!fixed_rate)
  353. fixed_rate = get_bits1(gb);
  354. if (fixed_rate)
  355. get_ue_golomb_long(gb); // elemental_duration_in_tc_minus1
  356. else
  357. low_delay = get_bits1(gb);
  358. if (!low_delay) {
  359. nb_cpb = get_ue_golomb_long(gb) + 1;
  360. if (nb_cpb < 1 || nb_cpb > 32) {
  361. av_log(NULL, AV_LOG_ERROR, "nb_cpb %d invalid\n", nb_cpb);
  362. return AVERROR_INVALIDDATA;
  363. }
  364. }
  365. if (nal_params_present)
  366. decode_sublayer_hrd(gb, nb_cpb, subpic_params_present);
  367. if (vcl_params_present)
  368. decode_sublayer_hrd(gb, nb_cpb, subpic_params_present);
  369. }
  370. return 0;
  371. }
  372. int ff_hevc_decode_nal_vps(GetBitContext *gb, AVCodecContext *avctx,
  373. HEVCParamSets *ps)
  374. {
  375. int i,j;
  376. int vps_id = 0;
  377. ptrdiff_t nal_size;
  378. HEVCVPS *vps;
  379. AVBufferRef *vps_buf = av_buffer_allocz(sizeof(*vps));
  380. if (!vps_buf)
  381. return AVERROR(ENOMEM);
  382. vps = (HEVCVPS*)vps_buf->data;
  383. av_log(avctx, AV_LOG_DEBUG, "Decoding VPS\n");
  384. nal_size = gb->buffer_end - gb->buffer;
  385. if (nal_size > sizeof(vps->data)) {
  386. av_log(avctx, AV_LOG_WARNING, "Truncating likely oversized VPS "
  387. "(%"PTRDIFF_SPECIFIER" > %"SIZE_SPECIFIER")\n",
  388. nal_size, sizeof(vps->data));
  389. vps->data_size = sizeof(vps->data);
  390. } else {
  391. vps->data_size = nal_size;
  392. }
  393. memcpy(vps->data, gb->buffer, vps->data_size);
  394. vps_id = get_bits(gb, 4);
  395. if (vps_id >= HEVC_MAX_VPS_COUNT) {
  396. av_log(avctx, AV_LOG_ERROR, "VPS id out of range: %d\n", vps_id);
  397. goto err;
  398. }
  399. if (get_bits(gb, 2) != 3) { // vps_reserved_three_2bits
  400. av_log(avctx, AV_LOG_ERROR, "vps_reserved_three_2bits is not three\n");
  401. goto err;
  402. }
  403. vps->vps_max_layers = get_bits(gb, 6) + 1;
  404. vps->vps_max_sub_layers = get_bits(gb, 3) + 1;
  405. vps->vps_temporal_id_nesting_flag = get_bits1(gb);
  406. if (get_bits(gb, 16) != 0xffff) { // vps_reserved_ffff_16bits
  407. av_log(avctx, AV_LOG_ERROR, "vps_reserved_ffff_16bits is not 0xffff\n");
  408. goto err;
  409. }
  410. if (vps->vps_max_sub_layers > HEVC_MAX_SUB_LAYERS) {
  411. av_log(avctx, AV_LOG_ERROR, "vps_max_sub_layers out of range: %d\n",
  412. vps->vps_max_sub_layers);
  413. goto err;
  414. }
  415. if (parse_ptl(gb, avctx, &vps->ptl, vps->vps_max_sub_layers) < 0)
  416. goto err;
  417. vps->vps_sub_layer_ordering_info_present_flag = get_bits1(gb);
  418. i = vps->vps_sub_layer_ordering_info_present_flag ? 0 : vps->vps_max_sub_layers - 1;
  419. for (; i < vps->vps_max_sub_layers; i++) {
  420. vps->vps_max_dec_pic_buffering[i] = get_ue_golomb_long(gb) + 1;
  421. vps->vps_num_reorder_pics[i] = get_ue_golomb_long(gb);
  422. vps->vps_max_latency_increase[i] = get_ue_golomb_long(gb) - 1;
  423. if (vps->vps_max_dec_pic_buffering[i] > HEVC_MAX_DPB_SIZE || !vps->vps_max_dec_pic_buffering[i]) {
  424. av_log(avctx, AV_LOG_ERROR, "vps_max_dec_pic_buffering_minus1 out of range: %d\n",
  425. vps->vps_max_dec_pic_buffering[i] - 1);
  426. goto err;
  427. }
  428. if (vps->vps_num_reorder_pics[i] > vps->vps_max_dec_pic_buffering[i] - 1) {
  429. av_log(avctx, AV_LOG_WARNING, "vps_max_num_reorder_pics out of range: %d\n",
  430. vps->vps_num_reorder_pics[i]);
  431. if (avctx->err_recognition & AV_EF_EXPLODE)
  432. goto err;
  433. }
  434. }
  435. vps->vps_max_layer_id = get_bits(gb, 6);
  436. vps->vps_num_layer_sets = get_ue_golomb_long(gb) + 1;
  437. if (vps->vps_num_layer_sets < 1 || vps->vps_num_layer_sets > 1024 ||
  438. (vps->vps_num_layer_sets - 1LL) * (vps->vps_max_layer_id + 1LL) > get_bits_left(gb)) {
  439. av_log(avctx, AV_LOG_ERROR, "too many layer_id_included_flags\n");
  440. goto err;
  441. }
  442. for (i = 1; i < vps->vps_num_layer_sets; i++)
  443. for (j = 0; j <= vps->vps_max_layer_id; j++)
  444. skip_bits(gb, 1); // layer_id_included_flag[i][j]
  445. vps->vps_timing_info_present_flag = get_bits1(gb);
  446. if (vps->vps_timing_info_present_flag) {
  447. vps->vps_num_units_in_tick = get_bits_long(gb, 32);
  448. vps->vps_time_scale = get_bits_long(gb, 32);
  449. vps->vps_poc_proportional_to_timing_flag = get_bits1(gb);
  450. if (vps->vps_poc_proportional_to_timing_flag)
  451. vps->vps_num_ticks_poc_diff_one = get_ue_golomb_long(gb) + 1;
  452. vps->vps_num_hrd_parameters = get_ue_golomb_long(gb);
  453. if (vps->vps_num_hrd_parameters > (unsigned)vps->vps_num_layer_sets) {
  454. av_log(avctx, AV_LOG_ERROR,
  455. "vps_num_hrd_parameters %d is invalid\n", vps->vps_num_hrd_parameters);
  456. goto err;
  457. }
  458. for (i = 0; i < vps->vps_num_hrd_parameters; i++) {
  459. int common_inf_present = 1;
  460. get_ue_golomb_long(gb); // hrd_layer_set_idx
  461. if (i)
  462. common_inf_present = get_bits1(gb);
  463. decode_hrd(gb, common_inf_present, vps->vps_max_sub_layers);
  464. }
  465. }
  466. get_bits1(gb); /* vps_extension_flag */
  467. if (get_bits_left(gb) < 0) {
  468. av_log(avctx, AV_LOG_ERROR,
  469. "Overread VPS by %d bits\n", -get_bits_left(gb));
  470. if (ps->vps_list[vps_id])
  471. goto err;
  472. }
  473. if (ps->vps_list[vps_id] &&
  474. !memcmp(ps->vps_list[vps_id]->data, vps_buf->data, vps_buf->size)) {
  475. av_buffer_unref(&vps_buf);
  476. } else {
  477. remove_vps(ps, vps_id);
  478. ps->vps_list[vps_id] = vps_buf;
  479. }
  480. return 0;
  481. err:
  482. av_buffer_unref(&vps_buf);
  483. return AVERROR_INVALIDDATA;
  484. }
  485. static void decode_vui(GetBitContext *gb, AVCodecContext *avctx,
  486. int apply_defdispwin, HEVCSPS *sps)
  487. {
  488. VUI backup_vui, *vui = &sps->vui;
  489. GetBitContext backup;
  490. int sar_present, alt = 0;
  491. av_log(avctx, AV_LOG_DEBUG, "Decoding VUI\n");
  492. sar_present = get_bits1(gb);
  493. if (sar_present) {
  494. uint8_t sar_idx = get_bits(gb, 8);
  495. if (sar_idx < FF_ARRAY_ELEMS(vui_sar))
  496. vui->sar = vui_sar[sar_idx];
  497. else if (sar_idx == 255) {
  498. vui->sar.num = get_bits(gb, 16);
  499. vui->sar.den = get_bits(gb, 16);
  500. } else
  501. av_log(avctx, AV_LOG_WARNING,
  502. "Unknown SAR index: %u.\n", sar_idx);
  503. }
  504. vui->overscan_info_present_flag = get_bits1(gb);
  505. if (vui->overscan_info_present_flag)
  506. vui->overscan_appropriate_flag = get_bits1(gb);
  507. vui->video_signal_type_present_flag = get_bits1(gb);
  508. if (vui->video_signal_type_present_flag) {
  509. vui->video_format = get_bits(gb, 3);
  510. vui->video_full_range_flag = get_bits1(gb);
  511. vui->colour_description_present_flag = get_bits1(gb);
  512. if (vui->video_full_range_flag && sps->pix_fmt == AV_PIX_FMT_YUV420P)
  513. sps->pix_fmt = AV_PIX_FMT_YUVJ420P;
  514. if (vui->colour_description_present_flag) {
  515. vui->colour_primaries = get_bits(gb, 8);
  516. vui->transfer_characteristic = get_bits(gb, 8);
  517. vui->matrix_coeffs = get_bits(gb, 8);
  518. // Set invalid values to "unspecified"
  519. if (!av_color_primaries_name(vui->colour_primaries))
  520. vui->colour_primaries = AVCOL_PRI_UNSPECIFIED;
  521. if (!av_color_transfer_name(vui->transfer_characteristic))
  522. vui->transfer_characteristic = AVCOL_TRC_UNSPECIFIED;
  523. if (!av_color_space_name(vui->matrix_coeffs))
  524. vui->matrix_coeffs = AVCOL_SPC_UNSPECIFIED;
  525. if (vui->matrix_coeffs == AVCOL_SPC_RGB) {
  526. switch (sps->pix_fmt) {
  527. case AV_PIX_FMT_YUV444P:
  528. sps->pix_fmt = AV_PIX_FMT_GBRP;
  529. break;
  530. case AV_PIX_FMT_YUV444P10:
  531. sps->pix_fmt = AV_PIX_FMT_GBRP10;
  532. break;
  533. case AV_PIX_FMT_YUV444P12:
  534. sps->pix_fmt = AV_PIX_FMT_GBRP12;
  535. break;
  536. }
  537. }
  538. }
  539. }
  540. vui->chroma_loc_info_present_flag = get_bits1(gb);
  541. if (vui->chroma_loc_info_present_flag) {
  542. vui->chroma_sample_loc_type_top_field = get_ue_golomb_long(gb);
  543. vui->chroma_sample_loc_type_bottom_field = get_ue_golomb_long(gb);
  544. }
  545. vui->neutra_chroma_indication_flag = get_bits1(gb);
  546. vui->field_seq_flag = get_bits1(gb);
  547. vui->frame_field_info_present_flag = get_bits1(gb);
  548. // Backup context in case an alternate header is detected
  549. memcpy(&backup, gb, sizeof(backup));
  550. memcpy(&backup_vui, vui, sizeof(backup_vui));
  551. if (get_bits_left(gb) >= 68 && show_bits_long(gb, 21) == 0x100000) {
  552. vui->default_display_window_flag = 0;
  553. av_log(avctx, AV_LOG_WARNING, "Invalid default display window\n");
  554. } else
  555. vui->default_display_window_flag = get_bits1(gb);
  556. if (vui->default_display_window_flag) {
  557. int vert_mult = hevc_sub_height_c[sps->chroma_format_idc];
  558. int horiz_mult = hevc_sub_width_c[sps->chroma_format_idc];
  559. vui->def_disp_win.left_offset = get_ue_golomb_long(gb) * horiz_mult;
  560. vui->def_disp_win.right_offset = get_ue_golomb_long(gb) * horiz_mult;
  561. vui->def_disp_win.top_offset = get_ue_golomb_long(gb) * vert_mult;
  562. vui->def_disp_win.bottom_offset = get_ue_golomb_long(gb) * vert_mult;
  563. if (apply_defdispwin &&
  564. avctx->flags2 & AV_CODEC_FLAG2_IGNORE_CROP) {
  565. av_log(avctx, AV_LOG_DEBUG,
  566. "discarding vui default display window, "
  567. "original values are l:%u r:%u t:%u b:%u\n",
  568. vui->def_disp_win.left_offset,
  569. vui->def_disp_win.right_offset,
  570. vui->def_disp_win.top_offset,
  571. vui->def_disp_win.bottom_offset);
  572. vui->def_disp_win.left_offset =
  573. vui->def_disp_win.right_offset =
  574. vui->def_disp_win.top_offset =
  575. vui->def_disp_win.bottom_offset = 0;
  576. }
  577. }
  578. timing_info:
  579. vui->vui_timing_info_present_flag = get_bits1(gb);
  580. if (vui->vui_timing_info_present_flag) {
  581. if( get_bits_left(gb) < 66 && !alt) {
  582. // The alternate syntax seem to have timing info located
  583. // at where def_disp_win is normally located
  584. av_log(avctx, AV_LOG_WARNING,
  585. "Strange VUI timing information, retrying...\n");
  586. memcpy(vui, &backup_vui, sizeof(backup_vui));
  587. memcpy(gb, &backup, sizeof(backup));
  588. alt = 1;
  589. goto timing_info;
  590. }
  591. vui->vui_num_units_in_tick = get_bits_long(gb, 32);
  592. vui->vui_time_scale = get_bits_long(gb, 32);
  593. if (alt) {
  594. av_log(avctx, AV_LOG_INFO, "Retry got %"PRIu32"/%"PRIu32" fps\n",
  595. vui->vui_time_scale, vui->vui_num_units_in_tick);
  596. }
  597. vui->vui_poc_proportional_to_timing_flag = get_bits1(gb);
  598. if (vui->vui_poc_proportional_to_timing_flag)
  599. vui->vui_num_ticks_poc_diff_one_minus1 = get_ue_golomb_long(gb);
  600. vui->vui_hrd_parameters_present_flag = get_bits1(gb);
  601. if (vui->vui_hrd_parameters_present_flag)
  602. decode_hrd(gb, 1, sps->max_sub_layers);
  603. }
  604. vui->bitstream_restriction_flag = get_bits1(gb);
  605. if (vui->bitstream_restriction_flag) {
  606. if (get_bits_left(gb) < 8 && !alt) {
  607. av_log(avctx, AV_LOG_WARNING,
  608. "Strange VUI bitstream restriction information, retrying"
  609. " from timing information...\n");
  610. memcpy(vui, &backup_vui, sizeof(backup_vui));
  611. memcpy(gb, &backup, sizeof(backup));
  612. alt = 1;
  613. goto timing_info;
  614. }
  615. vui->tiles_fixed_structure_flag = get_bits1(gb);
  616. vui->motion_vectors_over_pic_boundaries_flag = get_bits1(gb);
  617. vui->restricted_ref_pic_lists_flag = get_bits1(gb);
  618. vui->min_spatial_segmentation_idc = get_ue_golomb_long(gb);
  619. vui->max_bytes_per_pic_denom = get_ue_golomb_long(gb);
  620. vui->max_bits_per_min_cu_denom = get_ue_golomb_long(gb);
  621. vui->log2_max_mv_length_horizontal = get_ue_golomb_long(gb);
  622. vui->log2_max_mv_length_vertical = get_ue_golomb_long(gb);
  623. }
  624. if (get_bits_left(gb) < 1 && !alt) {
  625. // XXX: Alternate syntax when sps_range_extension_flag != 0?
  626. av_log(avctx, AV_LOG_WARNING,
  627. "Overread in VUI, retrying from timing information...\n");
  628. memcpy(vui, &backup_vui, sizeof(backup_vui));
  629. memcpy(gb, &backup, sizeof(backup));
  630. alt = 1;
  631. goto timing_info;
  632. }
  633. }
  634. static void set_default_scaling_list_data(ScalingList *sl)
  635. {
  636. int matrixId;
  637. for (matrixId = 0; matrixId < 6; matrixId++) {
  638. // 4x4 default is 16
  639. memset(sl->sl[0][matrixId], 16, 16);
  640. sl->sl_dc[0][matrixId] = 16; // default for 16x16
  641. sl->sl_dc[1][matrixId] = 16; // default for 32x32
  642. }
  643. memcpy(sl->sl[1][0], default_scaling_list_intra, 64);
  644. memcpy(sl->sl[1][1], default_scaling_list_intra, 64);
  645. memcpy(sl->sl[1][2], default_scaling_list_intra, 64);
  646. memcpy(sl->sl[1][3], default_scaling_list_inter, 64);
  647. memcpy(sl->sl[1][4], default_scaling_list_inter, 64);
  648. memcpy(sl->sl[1][5], default_scaling_list_inter, 64);
  649. memcpy(sl->sl[2][0], default_scaling_list_intra, 64);
  650. memcpy(sl->sl[2][1], default_scaling_list_intra, 64);
  651. memcpy(sl->sl[2][2], default_scaling_list_intra, 64);
  652. memcpy(sl->sl[2][3], default_scaling_list_inter, 64);
  653. memcpy(sl->sl[2][4], default_scaling_list_inter, 64);
  654. memcpy(sl->sl[2][5], default_scaling_list_inter, 64);
  655. memcpy(sl->sl[3][0], default_scaling_list_intra, 64);
  656. memcpy(sl->sl[3][1], default_scaling_list_intra, 64);
  657. memcpy(sl->sl[3][2], default_scaling_list_intra, 64);
  658. memcpy(sl->sl[3][3], default_scaling_list_inter, 64);
  659. memcpy(sl->sl[3][4], default_scaling_list_inter, 64);
  660. memcpy(sl->sl[3][5], default_scaling_list_inter, 64);
  661. }
  662. static int scaling_list_data(GetBitContext *gb, AVCodecContext *avctx, ScalingList *sl, HEVCSPS *sps)
  663. {
  664. uint8_t scaling_list_pred_mode_flag;
  665. int32_t scaling_list_dc_coef[2][6];
  666. int size_id, matrix_id, pos;
  667. int i;
  668. for (size_id = 0; size_id < 4; size_id++)
  669. for (matrix_id = 0; matrix_id < 6; matrix_id += ((size_id == 3) ? 3 : 1)) {
  670. scaling_list_pred_mode_flag = get_bits1(gb);
  671. if (!scaling_list_pred_mode_flag) {
  672. unsigned int delta = get_ue_golomb_long(gb);
  673. /* Only need to handle non-zero delta. Zero means default,
  674. * which should already be in the arrays. */
  675. if (delta) {
  676. // Copy from previous array.
  677. delta *= (size_id == 3) ? 3 : 1;
  678. if (matrix_id < delta) {
  679. av_log(avctx, AV_LOG_ERROR,
  680. "Invalid delta in scaling list data: %d.\n", delta);
  681. return AVERROR_INVALIDDATA;
  682. }
  683. memcpy(sl->sl[size_id][matrix_id],
  684. sl->sl[size_id][matrix_id - delta],
  685. size_id > 0 ? 64 : 16);
  686. if (size_id > 1)
  687. sl->sl_dc[size_id - 2][matrix_id] = sl->sl_dc[size_id - 2][matrix_id - delta];
  688. }
  689. } else {
  690. int next_coef, coef_num;
  691. int32_t scaling_list_delta_coef;
  692. next_coef = 8;
  693. coef_num = FFMIN(64, 1 << (4 + (size_id << 1)));
  694. if (size_id > 1) {
  695. scaling_list_dc_coef[size_id - 2][matrix_id] = get_se_golomb(gb) + 8;
  696. next_coef = scaling_list_dc_coef[size_id - 2][matrix_id];
  697. sl->sl_dc[size_id - 2][matrix_id] = next_coef;
  698. }
  699. for (i = 0; i < coef_num; i++) {
  700. if (size_id == 0)
  701. pos = 4 * ff_hevc_diag_scan4x4_y[i] +
  702. ff_hevc_diag_scan4x4_x[i];
  703. else
  704. pos = 8 * ff_hevc_diag_scan8x8_y[i] +
  705. ff_hevc_diag_scan8x8_x[i];
  706. scaling_list_delta_coef = get_se_golomb(gb);
  707. next_coef = (next_coef + 256U + scaling_list_delta_coef) % 256;
  708. sl->sl[size_id][matrix_id][pos] = next_coef;
  709. }
  710. }
  711. }
  712. if (sps->chroma_format_idc == 3) {
  713. for (i = 0; i < 64; i++) {
  714. sl->sl[3][1][i] = sl->sl[2][1][i];
  715. sl->sl[3][2][i] = sl->sl[2][2][i];
  716. sl->sl[3][4][i] = sl->sl[2][4][i];
  717. sl->sl[3][5][i] = sl->sl[2][5][i];
  718. }
  719. sl->sl_dc[1][1] = sl->sl_dc[0][1];
  720. sl->sl_dc[1][2] = sl->sl_dc[0][2];
  721. sl->sl_dc[1][4] = sl->sl_dc[0][4];
  722. sl->sl_dc[1][5] = sl->sl_dc[0][5];
  723. }
  724. return 0;
  725. }
  726. static int map_pixel_format(AVCodecContext *avctx, HEVCSPS *sps)
  727. {
  728. const AVPixFmtDescriptor *desc;
  729. switch (sps->bit_depth) {
  730. case 8:
  731. if (sps->chroma_format_idc == 0) sps->pix_fmt = AV_PIX_FMT_GRAY8;
  732. if (sps->chroma_format_idc == 1) sps->pix_fmt = AV_PIX_FMT_YUV420P;
  733. if (sps->chroma_format_idc == 2) sps->pix_fmt = AV_PIX_FMT_YUV422P;
  734. if (sps->chroma_format_idc == 3) sps->pix_fmt = AV_PIX_FMT_YUV444P;
  735. break;
  736. case 9:
  737. if (sps->chroma_format_idc == 0) sps->pix_fmt = AV_PIX_FMT_GRAY9;
  738. if (sps->chroma_format_idc == 1) sps->pix_fmt = AV_PIX_FMT_YUV420P9;
  739. if (sps->chroma_format_idc == 2) sps->pix_fmt = AV_PIX_FMT_YUV422P9;
  740. if (sps->chroma_format_idc == 3) sps->pix_fmt = AV_PIX_FMT_YUV444P9;
  741. break;
  742. case 10:
  743. if (sps->chroma_format_idc == 0) sps->pix_fmt = AV_PIX_FMT_GRAY10;
  744. if (sps->chroma_format_idc == 1) sps->pix_fmt = AV_PIX_FMT_YUV420P10;
  745. if (sps->chroma_format_idc == 2) sps->pix_fmt = AV_PIX_FMT_YUV422P10;
  746. if (sps->chroma_format_idc == 3) sps->pix_fmt = AV_PIX_FMT_YUV444P10;
  747. break;
  748. case 12:
  749. if (sps->chroma_format_idc == 0) sps->pix_fmt = AV_PIX_FMT_GRAY12;
  750. if (sps->chroma_format_idc == 1) sps->pix_fmt = AV_PIX_FMT_YUV420P12;
  751. if (sps->chroma_format_idc == 2) sps->pix_fmt = AV_PIX_FMT_YUV422P12;
  752. if (sps->chroma_format_idc == 3) sps->pix_fmt = AV_PIX_FMT_YUV444P12;
  753. break;
  754. default:
  755. av_log(avctx, AV_LOG_ERROR,
  756. "The following bit-depths are currently specified: 8, 9, 10 and 12 bits, "
  757. "chroma_format_idc is %d, depth is %d\n",
  758. sps->chroma_format_idc, sps->bit_depth);
  759. return AVERROR_INVALIDDATA;
  760. }
  761. desc = av_pix_fmt_desc_get(sps->pix_fmt);
  762. if (!desc)
  763. return AVERROR(EINVAL);
  764. sps->hshift[0] = sps->vshift[0] = 0;
  765. sps->hshift[2] = sps->hshift[1] = desc->log2_chroma_w;
  766. sps->vshift[2] = sps->vshift[1] = desc->log2_chroma_h;
  767. sps->pixel_shift = sps->bit_depth > 8;
  768. return 0;
  769. }
  770. int ff_hevc_parse_sps(HEVCSPS *sps, GetBitContext *gb, unsigned int *sps_id,
  771. int apply_defdispwin, AVBufferRef **vps_list, AVCodecContext *avctx)
  772. {
  773. HEVCWindow *ow;
  774. int ret = 0;
  775. int log2_diff_max_min_transform_block_size;
  776. int bit_depth_chroma, start, vui_present, sublayer_ordering_info;
  777. int i;
  778. // Coded parameters
  779. sps->vps_id = get_bits(gb, 4);
  780. if (sps->vps_id >= HEVC_MAX_VPS_COUNT) {
  781. av_log(avctx, AV_LOG_ERROR, "VPS id out of range: %d\n", sps->vps_id);
  782. return AVERROR_INVALIDDATA;
  783. }
  784. if (vps_list && !vps_list[sps->vps_id]) {
  785. av_log(avctx, AV_LOG_ERROR, "VPS %d does not exist\n",
  786. sps->vps_id);
  787. return AVERROR_INVALIDDATA;
  788. }
  789. sps->max_sub_layers = get_bits(gb, 3) + 1;
  790. if (sps->max_sub_layers > HEVC_MAX_SUB_LAYERS) {
  791. av_log(avctx, AV_LOG_ERROR, "sps_max_sub_layers out of range: %d\n",
  792. sps->max_sub_layers);
  793. return AVERROR_INVALIDDATA;
  794. }
  795. sps->temporal_id_nesting_flag = get_bits(gb, 1);
  796. if ((ret = parse_ptl(gb, avctx, &sps->ptl, sps->max_sub_layers)) < 0)
  797. return ret;
  798. *sps_id = get_ue_golomb_long(gb);
  799. if (*sps_id >= HEVC_MAX_SPS_COUNT) {
  800. av_log(avctx, AV_LOG_ERROR, "SPS id out of range: %d\n", *sps_id);
  801. return AVERROR_INVALIDDATA;
  802. }
  803. sps->chroma_format_idc = get_ue_golomb_long(gb);
  804. if (sps->chroma_format_idc > 3U) {
  805. av_log(avctx, AV_LOG_ERROR, "chroma_format_idc %d is invalid\n", sps->chroma_format_idc);
  806. return AVERROR_INVALIDDATA;
  807. }
  808. if (sps->chroma_format_idc == 3)
  809. sps->separate_colour_plane_flag = get_bits1(gb);
  810. if (sps->separate_colour_plane_flag)
  811. sps->chroma_format_idc = 0;
  812. sps->width = get_ue_golomb_long(gb);
  813. sps->height = get_ue_golomb_long(gb);
  814. if ((ret = av_image_check_size(sps->width,
  815. sps->height, 0, avctx)) < 0)
  816. return ret;
  817. if (get_bits1(gb)) { // pic_conformance_flag
  818. int vert_mult = hevc_sub_height_c[sps->chroma_format_idc];
  819. int horiz_mult = hevc_sub_width_c[sps->chroma_format_idc];
  820. sps->pic_conf_win.left_offset = get_ue_golomb_long(gb) * horiz_mult;
  821. sps->pic_conf_win.right_offset = get_ue_golomb_long(gb) * horiz_mult;
  822. sps->pic_conf_win.top_offset = get_ue_golomb_long(gb) * vert_mult;
  823. sps->pic_conf_win.bottom_offset = get_ue_golomb_long(gb) * vert_mult;
  824. if (avctx->flags2 & AV_CODEC_FLAG2_IGNORE_CROP) {
  825. av_log(avctx, AV_LOG_DEBUG,
  826. "discarding sps conformance window, "
  827. "original values are l:%u r:%u t:%u b:%u\n",
  828. sps->pic_conf_win.left_offset,
  829. sps->pic_conf_win.right_offset,
  830. sps->pic_conf_win.top_offset,
  831. sps->pic_conf_win.bottom_offset);
  832. sps->pic_conf_win.left_offset =
  833. sps->pic_conf_win.right_offset =
  834. sps->pic_conf_win.top_offset =
  835. sps->pic_conf_win.bottom_offset = 0;
  836. }
  837. sps->output_window = sps->pic_conf_win;
  838. }
  839. sps->bit_depth = get_ue_golomb_long(gb) + 8;
  840. bit_depth_chroma = get_ue_golomb_long(gb) + 8;
  841. if (sps->chroma_format_idc && bit_depth_chroma != sps->bit_depth) {
  842. av_log(avctx, AV_LOG_ERROR,
  843. "Luma bit depth (%d) is different from chroma bit depth (%d), "
  844. "this is unsupported.\n",
  845. sps->bit_depth, bit_depth_chroma);
  846. return AVERROR_INVALIDDATA;
  847. }
  848. sps->bit_depth_chroma = bit_depth_chroma;
  849. ret = map_pixel_format(avctx, sps);
  850. if (ret < 0)
  851. return ret;
  852. sps->log2_max_poc_lsb = get_ue_golomb_long(gb) + 4;
  853. if (sps->log2_max_poc_lsb > 16) {
  854. av_log(avctx, AV_LOG_ERROR, "log2_max_pic_order_cnt_lsb_minus4 out range: %d\n",
  855. sps->log2_max_poc_lsb - 4);
  856. return AVERROR_INVALIDDATA;
  857. }
  858. sublayer_ordering_info = get_bits1(gb);
  859. start = sublayer_ordering_info ? 0 : sps->max_sub_layers - 1;
  860. for (i = start; i < sps->max_sub_layers; i++) {
  861. sps->temporal_layer[i].max_dec_pic_buffering = get_ue_golomb_long(gb) + 1;
  862. sps->temporal_layer[i].num_reorder_pics = get_ue_golomb_long(gb);
  863. sps->temporal_layer[i].max_latency_increase = get_ue_golomb_long(gb) - 1;
  864. if (sps->temporal_layer[i].max_dec_pic_buffering > (unsigned)HEVC_MAX_DPB_SIZE) {
  865. av_log(avctx, AV_LOG_ERROR, "sps_max_dec_pic_buffering_minus1 out of range: %d\n",
  866. sps->temporal_layer[i].max_dec_pic_buffering - 1U);
  867. return AVERROR_INVALIDDATA;
  868. }
  869. if (sps->temporal_layer[i].num_reorder_pics > sps->temporal_layer[i].max_dec_pic_buffering - 1) {
  870. av_log(avctx, AV_LOG_WARNING, "sps_max_num_reorder_pics out of range: %d\n",
  871. sps->temporal_layer[i].num_reorder_pics);
  872. if (avctx->err_recognition & AV_EF_EXPLODE ||
  873. sps->temporal_layer[i].num_reorder_pics > HEVC_MAX_DPB_SIZE - 1) {
  874. return AVERROR_INVALIDDATA;
  875. }
  876. sps->temporal_layer[i].max_dec_pic_buffering = sps->temporal_layer[i].num_reorder_pics + 1;
  877. }
  878. }
  879. if (!sublayer_ordering_info) {
  880. for (i = 0; i < start; i++) {
  881. sps->temporal_layer[i].max_dec_pic_buffering = sps->temporal_layer[start].max_dec_pic_buffering;
  882. sps->temporal_layer[i].num_reorder_pics = sps->temporal_layer[start].num_reorder_pics;
  883. sps->temporal_layer[i].max_latency_increase = sps->temporal_layer[start].max_latency_increase;
  884. }
  885. }
  886. sps->log2_min_cb_size = get_ue_golomb_long(gb) + 3;
  887. sps->log2_diff_max_min_coding_block_size = get_ue_golomb_long(gb);
  888. sps->log2_min_tb_size = get_ue_golomb_long(gb) + 2;
  889. log2_diff_max_min_transform_block_size = get_ue_golomb_long(gb);
  890. sps->log2_max_trafo_size = log2_diff_max_min_transform_block_size +
  891. sps->log2_min_tb_size;
  892. if (sps->log2_min_cb_size < 3 || sps->log2_min_cb_size > 30) {
  893. av_log(avctx, AV_LOG_ERROR, "Invalid value %d for log2_min_cb_size", sps->log2_min_cb_size);
  894. return AVERROR_INVALIDDATA;
  895. }
  896. if (sps->log2_diff_max_min_coding_block_size > 30) {
  897. av_log(avctx, AV_LOG_ERROR, "Invalid value %d for log2_diff_max_min_coding_block_size", sps->log2_diff_max_min_coding_block_size);
  898. return AVERROR_INVALIDDATA;
  899. }
  900. if (sps->log2_min_tb_size >= sps->log2_min_cb_size || sps->log2_min_tb_size < 2) {
  901. av_log(avctx, AV_LOG_ERROR, "Invalid value for log2_min_tb_size");
  902. return AVERROR_INVALIDDATA;
  903. }
  904. if (log2_diff_max_min_transform_block_size < 0 || log2_diff_max_min_transform_block_size > 30) {
  905. av_log(avctx, AV_LOG_ERROR, "Invalid value %d for log2_diff_max_min_transform_block_size", log2_diff_max_min_transform_block_size);
  906. return AVERROR_INVALIDDATA;
  907. }
  908. sps->max_transform_hierarchy_depth_inter = get_ue_golomb_long(gb);
  909. sps->max_transform_hierarchy_depth_intra = get_ue_golomb_long(gb);
  910. sps->scaling_list_enable_flag = get_bits1(gb);
  911. if (sps->scaling_list_enable_flag) {
  912. set_default_scaling_list_data(&sps->scaling_list);
  913. if (get_bits1(gb)) {
  914. ret = scaling_list_data(gb, avctx, &sps->scaling_list, sps);
  915. if (ret < 0)
  916. return ret;
  917. }
  918. }
  919. sps->amp_enabled_flag = get_bits1(gb);
  920. sps->sao_enabled = get_bits1(gb);
  921. sps->pcm_enabled_flag = get_bits1(gb);
  922. if (sps->pcm_enabled_flag) {
  923. sps->pcm.bit_depth = get_bits(gb, 4) + 1;
  924. sps->pcm.bit_depth_chroma = get_bits(gb, 4) + 1;
  925. sps->pcm.log2_min_pcm_cb_size = get_ue_golomb_long(gb) + 3;
  926. sps->pcm.log2_max_pcm_cb_size = sps->pcm.log2_min_pcm_cb_size +
  927. get_ue_golomb_long(gb);
  928. if (FFMAX(sps->pcm.bit_depth, sps->pcm.bit_depth_chroma) > sps->bit_depth) {
  929. av_log(avctx, AV_LOG_ERROR,
  930. "PCM bit depth (%d, %d) is greater than normal bit depth (%d)\n",
  931. sps->pcm.bit_depth, sps->pcm.bit_depth_chroma, sps->bit_depth);
  932. return AVERROR_INVALIDDATA;
  933. }
  934. sps->pcm.loop_filter_disable_flag = get_bits1(gb);
  935. }
  936. sps->nb_st_rps = get_ue_golomb_long(gb);
  937. if (sps->nb_st_rps > HEVC_MAX_SHORT_TERM_REF_PIC_SETS) {
  938. av_log(avctx, AV_LOG_ERROR, "Too many short term RPS: %d.\n",
  939. sps->nb_st_rps);
  940. return AVERROR_INVALIDDATA;
  941. }
  942. for (i = 0; i < sps->nb_st_rps; i++) {
  943. if ((ret = ff_hevc_decode_short_term_rps(gb, avctx, &sps->st_rps[i],
  944. sps, 0)) < 0)
  945. return ret;
  946. }
  947. sps->long_term_ref_pics_present_flag = get_bits1(gb);
  948. if (sps->long_term_ref_pics_present_flag) {
  949. sps->num_long_term_ref_pics_sps = get_ue_golomb_long(gb);
  950. if (sps->num_long_term_ref_pics_sps > HEVC_MAX_LONG_TERM_REF_PICS) {
  951. av_log(avctx, AV_LOG_ERROR, "Too many long term ref pics: %d.\n",
  952. sps->num_long_term_ref_pics_sps);
  953. return AVERROR_INVALIDDATA;
  954. }
  955. for (i = 0; i < sps->num_long_term_ref_pics_sps; i++) {
  956. sps->lt_ref_pic_poc_lsb_sps[i] = get_bits(gb, sps->log2_max_poc_lsb);
  957. sps->used_by_curr_pic_lt_sps_flag[i] = get_bits1(gb);
  958. }
  959. }
  960. sps->sps_temporal_mvp_enabled_flag = get_bits1(gb);
  961. sps->sps_strong_intra_smoothing_enable_flag = get_bits1(gb);
  962. sps->vui.sar = (AVRational){0, 1};
  963. vui_present = get_bits1(gb);
  964. if (vui_present)
  965. decode_vui(gb, avctx, apply_defdispwin, sps);
  966. if (get_bits1(gb)) { // sps_extension_flag
  967. int sps_range_extension_flag = get_bits1(gb);
  968. skip_bits(gb, 7); //sps_extension_7bits = get_bits(gb, 7);
  969. if (sps_range_extension_flag) {
  970. int extended_precision_processing_flag;
  971. int cabac_bypass_alignment_enabled_flag;
  972. sps->transform_skip_rotation_enabled_flag = get_bits1(gb);
  973. sps->transform_skip_context_enabled_flag = get_bits1(gb);
  974. sps->implicit_rdpcm_enabled_flag = get_bits1(gb);
  975. sps->explicit_rdpcm_enabled_flag = get_bits1(gb);
  976. extended_precision_processing_flag = get_bits1(gb);
  977. if (extended_precision_processing_flag)
  978. av_log(avctx, AV_LOG_WARNING,
  979. "extended_precision_processing_flag not yet implemented\n");
  980. sps->intra_smoothing_disabled_flag = get_bits1(gb);
  981. sps->high_precision_offsets_enabled_flag = get_bits1(gb);
  982. if (sps->high_precision_offsets_enabled_flag)
  983. av_log(avctx, AV_LOG_WARNING,
  984. "high_precision_offsets_enabled_flag not yet implemented\n");
  985. sps->persistent_rice_adaptation_enabled_flag = get_bits1(gb);
  986. cabac_bypass_alignment_enabled_flag = get_bits1(gb);
  987. if (cabac_bypass_alignment_enabled_flag)
  988. av_log(avctx, AV_LOG_WARNING,
  989. "cabac_bypass_alignment_enabled_flag not yet implemented\n");
  990. }
  991. }
  992. if (apply_defdispwin) {
  993. sps->output_window.left_offset += sps->vui.def_disp_win.left_offset;
  994. sps->output_window.right_offset += sps->vui.def_disp_win.right_offset;
  995. sps->output_window.top_offset += sps->vui.def_disp_win.top_offset;
  996. sps->output_window.bottom_offset += sps->vui.def_disp_win.bottom_offset;
  997. }
  998. ow = &sps->output_window;
  999. if (ow->left_offset >= INT_MAX - ow->right_offset ||
  1000. ow->top_offset >= INT_MAX - ow->bottom_offset ||
  1001. ow->left_offset + ow->right_offset >= sps->width ||
  1002. ow->top_offset + ow->bottom_offset >= sps->height) {
  1003. av_log(avctx, AV_LOG_WARNING, "Invalid cropping offsets: %u/%u/%u/%u\n",
  1004. ow->left_offset, ow->right_offset, ow->top_offset, ow->bottom_offset);
  1005. if (avctx->err_recognition & AV_EF_EXPLODE) {
  1006. return AVERROR_INVALIDDATA;
  1007. }
  1008. av_log(avctx, AV_LOG_WARNING,
  1009. "Displaying the whole video surface.\n");
  1010. memset(ow, 0, sizeof(*ow));
  1011. memset(&sps->pic_conf_win, 0, sizeof(sps->pic_conf_win));
  1012. }
  1013. // Inferred parameters
  1014. sps->log2_ctb_size = sps->log2_min_cb_size +
  1015. sps->log2_diff_max_min_coding_block_size;
  1016. sps->log2_min_pu_size = sps->log2_min_cb_size - 1;
  1017. if (sps->log2_ctb_size > HEVC_MAX_LOG2_CTB_SIZE) {
  1018. av_log(avctx, AV_LOG_ERROR, "CTB size out of range: 2^%d\n", sps->log2_ctb_size);
  1019. return AVERROR_INVALIDDATA;
  1020. }
  1021. if (sps->log2_ctb_size < 4) {
  1022. av_log(avctx,
  1023. AV_LOG_ERROR,
  1024. "log2_ctb_size %d differs from the bounds of any known profile\n",
  1025. sps->log2_ctb_size);
  1026. avpriv_request_sample(avctx, "log2_ctb_size %d", sps->log2_ctb_size);
  1027. return AVERROR_INVALIDDATA;
  1028. }
  1029. sps->ctb_width = (sps->width + (1 << sps->log2_ctb_size) - 1) >> sps->log2_ctb_size;
  1030. sps->ctb_height = (sps->height + (1 << sps->log2_ctb_size) - 1) >> sps->log2_ctb_size;
  1031. sps->ctb_size = sps->ctb_width * sps->ctb_height;
  1032. sps->min_cb_width = sps->width >> sps->log2_min_cb_size;
  1033. sps->min_cb_height = sps->height >> sps->log2_min_cb_size;
  1034. sps->min_tb_width = sps->width >> sps->log2_min_tb_size;
  1035. sps->min_tb_height = sps->height >> sps->log2_min_tb_size;
  1036. sps->min_pu_width = sps->width >> sps->log2_min_pu_size;
  1037. sps->min_pu_height = sps->height >> sps->log2_min_pu_size;
  1038. sps->tb_mask = (1 << (sps->log2_ctb_size - sps->log2_min_tb_size)) - 1;
  1039. sps->qp_bd_offset = 6 * (sps->bit_depth - 8);
  1040. if (av_mod_uintp2(sps->width, sps->log2_min_cb_size) ||
  1041. av_mod_uintp2(sps->height, sps->log2_min_cb_size)) {
  1042. av_log(avctx, AV_LOG_ERROR, "Invalid coded frame dimensions.\n");
  1043. return AVERROR_INVALIDDATA;
  1044. }
  1045. if (sps->max_transform_hierarchy_depth_inter > sps->log2_ctb_size - sps->log2_min_tb_size) {
  1046. av_log(avctx, AV_LOG_ERROR, "max_transform_hierarchy_depth_inter out of range: %d\n",
  1047. sps->max_transform_hierarchy_depth_inter);
  1048. return AVERROR_INVALIDDATA;
  1049. }
  1050. if (sps->max_transform_hierarchy_depth_intra > sps->log2_ctb_size - sps->log2_min_tb_size) {
  1051. av_log(avctx, AV_LOG_ERROR, "max_transform_hierarchy_depth_intra out of range: %d\n",
  1052. sps->max_transform_hierarchy_depth_intra);
  1053. return AVERROR_INVALIDDATA;
  1054. }
  1055. if (sps->log2_max_trafo_size > FFMIN(sps->log2_ctb_size, 5)) {
  1056. av_log(avctx, AV_LOG_ERROR,
  1057. "max transform block size out of range: %d\n",
  1058. sps->log2_max_trafo_size);
  1059. return AVERROR_INVALIDDATA;
  1060. }
  1061. if (get_bits_left(gb) < 0) {
  1062. av_log(avctx, AV_LOG_ERROR,
  1063. "Overread SPS by %d bits\n", -get_bits_left(gb));
  1064. return AVERROR_INVALIDDATA;
  1065. }
  1066. return 0;
  1067. }
  1068. int ff_hevc_decode_nal_sps(GetBitContext *gb, AVCodecContext *avctx,
  1069. HEVCParamSets *ps, int apply_defdispwin)
  1070. {
  1071. HEVCSPS *sps;
  1072. AVBufferRef *sps_buf = av_buffer_allocz(sizeof(*sps));
  1073. unsigned int sps_id;
  1074. int ret;
  1075. ptrdiff_t nal_size;
  1076. if (!sps_buf)
  1077. return AVERROR(ENOMEM);
  1078. sps = (HEVCSPS*)sps_buf->data;
  1079. av_log(avctx, AV_LOG_DEBUG, "Decoding SPS\n");
  1080. nal_size = gb->buffer_end - gb->buffer;
  1081. if (nal_size > sizeof(sps->data)) {
  1082. av_log(avctx, AV_LOG_WARNING, "Truncating likely oversized SPS "
  1083. "(%"PTRDIFF_SPECIFIER" > %"SIZE_SPECIFIER")\n",
  1084. nal_size, sizeof(sps->data));
  1085. sps->data_size = sizeof(sps->data);
  1086. } else {
  1087. sps->data_size = nal_size;
  1088. }
  1089. memcpy(sps->data, gb->buffer, sps->data_size);
  1090. ret = ff_hevc_parse_sps(sps, gb, &sps_id,
  1091. apply_defdispwin,
  1092. ps->vps_list, avctx);
  1093. if (ret < 0) {
  1094. av_buffer_unref(&sps_buf);
  1095. return ret;
  1096. }
  1097. if (avctx->debug & FF_DEBUG_BITSTREAM) {
  1098. av_log(avctx, AV_LOG_DEBUG,
  1099. "Parsed SPS: id %d; coded wxh: %dx%d; "
  1100. "cropped wxh: %dx%d; pix_fmt: %s.\n",
  1101. sps_id, sps->width, sps->height,
  1102. sps->width - (sps->output_window.left_offset + sps->output_window.right_offset),
  1103. sps->height - (sps->output_window.top_offset + sps->output_window.bottom_offset),
  1104. av_get_pix_fmt_name(sps->pix_fmt));
  1105. }
  1106. /* check if this is a repeat of an already parsed SPS, then keep the
  1107. * original one.
  1108. * otherwise drop all PPSes that depend on it */
  1109. if (ps->sps_list[sps_id] &&
  1110. !memcmp(ps->sps_list[sps_id]->data, sps_buf->data, sps_buf->size)) {
  1111. av_buffer_unref(&sps_buf);
  1112. } else {
  1113. remove_sps(ps, sps_id);
  1114. ps->sps_list[sps_id] = sps_buf;
  1115. }
  1116. return 0;
  1117. }
  1118. static void hevc_pps_free(void *opaque, uint8_t *data)
  1119. {
  1120. HEVCPPS *pps = (HEVCPPS*)data;
  1121. av_freep(&pps->column_width);
  1122. av_freep(&pps->row_height);
  1123. av_freep(&pps->col_bd);
  1124. av_freep(&pps->row_bd);
  1125. av_freep(&pps->col_idxX);
  1126. av_freep(&pps->ctb_addr_rs_to_ts);
  1127. av_freep(&pps->ctb_addr_ts_to_rs);
  1128. av_freep(&pps->tile_pos_rs);
  1129. av_freep(&pps->tile_id);
  1130. av_freep(&pps->min_tb_addr_zs_tab);
  1131. av_freep(&pps);
  1132. }
  1133. static int pps_range_extensions(GetBitContext *gb, AVCodecContext *avctx,
  1134. HEVCPPS *pps, HEVCSPS *sps) {
  1135. int i;
  1136. if (pps->transform_skip_enabled_flag) {
  1137. pps->log2_max_transform_skip_block_size = get_ue_golomb_long(gb) + 2;
  1138. }
  1139. pps->cross_component_prediction_enabled_flag = get_bits1(gb);
  1140. pps->chroma_qp_offset_list_enabled_flag = get_bits1(gb);
  1141. if (pps->chroma_qp_offset_list_enabled_flag) {
  1142. pps->diff_cu_chroma_qp_offset_depth = get_ue_golomb_long(gb);
  1143. pps->chroma_qp_offset_list_len_minus1 = get_ue_golomb_long(gb);
  1144. if (pps->chroma_qp_offset_list_len_minus1 > 5) {
  1145. av_log(avctx, AV_LOG_ERROR,
  1146. "chroma_qp_offset_list_len_minus1 shall be in the range [0, 5].\n");
  1147. return AVERROR_INVALIDDATA;
  1148. }
  1149. for (i = 0; i <= pps->chroma_qp_offset_list_len_minus1; i++) {
  1150. pps->cb_qp_offset_list[i] = get_se_golomb_long(gb);
  1151. if (pps->cb_qp_offset_list[i]) {
  1152. av_log(avctx, AV_LOG_WARNING,
  1153. "cb_qp_offset_list not tested yet.\n");
  1154. }
  1155. pps->cr_qp_offset_list[i] = get_se_golomb_long(gb);
  1156. if (pps->cr_qp_offset_list[i]) {
  1157. av_log(avctx, AV_LOG_WARNING,
  1158. "cb_qp_offset_list not tested yet.\n");
  1159. }
  1160. }
  1161. }
  1162. pps->log2_sao_offset_scale_luma = get_ue_golomb_long(gb);
  1163. pps->log2_sao_offset_scale_chroma = get_ue_golomb_long(gb);
  1164. if ( pps->log2_sao_offset_scale_luma > FFMAX(sps->bit_depth - 10, 0)
  1165. || pps->log2_sao_offset_scale_chroma > FFMAX(sps->bit_depth_chroma - 10, 0)
  1166. )
  1167. return AVERROR_INVALIDDATA;
  1168. return(0);
  1169. }
  1170. static inline int setup_pps(AVCodecContext *avctx, GetBitContext *gb,
  1171. HEVCPPS *pps, HEVCSPS *sps)
  1172. {
  1173. int log2_diff;
  1174. int pic_area_in_ctbs;
  1175. int i, j, x, y, ctb_addr_rs, tile_id;
  1176. // Inferred parameters
  1177. pps->col_bd = av_malloc_array(pps->num_tile_columns + 1, sizeof(*pps->col_bd));
  1178. pps->row_bd = av_malloc_array(pps->num_tile_rows + 1, sizeof(*pps->row_bd));
  1179. pps->col_idxX = av_malloc_array(sps->ctb_width, sizeof(*pps->col_idxX));
  1180. if (!pps->col_bd || !pps->row_bd || !pps->col_idxX)
  1181. return AVERROR(ENOMEM);
  1182. if (pps->uniform_spacing_flag) {
  1183. if (!pps->column_width) {
  1184. pps->column_width = av_malloc_array(pps->num_tile_columns, sizeof(*pps->column_width));
  1185. pps->row_height = av_malloc_array(pps->num_tile_rows, sizeof(*pps->row_height));
  1186. }
  1187. if (!pps->column_width || !pps->row_height)
  1188. return AVERROR(ENOMEM);
  1189. for (i = 0; i < pps->num_tile_columns; i++) {
  1190. pps->column_width[i] = ((i + 1) * sps->ctb_width) / pps->num_tile_columns -
  1191. (i * sps->ctb_width) / pps->num_tile_columns;
  1192. }
  1193. for (i = 0; i < pps->num_tile_rows; i++) {
  1194. pps->row_height[i] = ((i + 1) * sps->ctb_height) / pps->num_tile_rows -
  1195. (i * sps->ctb_height) / pps->num_tile_rows;
  1196. }
  1197. }
  1198. pps->col_bd[0] = 0;
  1199. for (i = 0; i < pps->num_tile_columns; i++)
  1200. pps->col_bd[i + 1] = pps->col_bd[i] + pps->column_width[i];
  1201. pps->row_bd[0] = 0;
  1202. for (i = 0; i < pps->num_tile_rows; i++)
  1203. pps->row_bd[i + 1] = pps->row_bd[i] + pps->row_height[i];
  1204. for (i = 0, j = 0; i < sps->ctb_width; i++) {
  1205. if (i > pps->col_bd[j])
  1206. j++;
  1207. pps->col_idxX[i] = j;
  1208. }
  1209. /**
  1210. * 6.5
  1211. */
  1212. pic_area_in_ctbs = sps->ctb_width * sps->ctb_height;
  1213. pps->ctb_addr_rs_to_ts = av_malloc_array(pic_area_in_ctbs, sizeof(*pps->ctb_addr_rs_to_ts));
  1214. pps->ctb_addr_ts_to_rs = av_malloc_array(pic_area_in_ctbs, sizeof(*pps->ctb_addr_ts_to_rs));
  1215. pps->tile_id = av_malloc_array(pic_area_in_ctbs, sizeof(*pps->tile_id));
  1216. pps->min_tb_addr_zs_tab = av_malloc_array((sps->tb_mask+2) * (sps->tb_mask+2), sizeof(*pps->min_tb_addr_zs_tab));
  1217. if (!pps->ctb_addr_rs_to_ts || !pps->ctb_addr_ts_to_rs ||
  1218. !pps->tile_id || !pps->min_tb_addr_zs_tab) {
  1219. return AVERROR(ENOMEM);
  1220. }
  1221. for (ctb_addr_rs = 0; ctb_addr_rs < pic_area_in_ctbs; ctb_addr_rs++) {
  1222. int tb_x = ctb_addr_rs % sps->ctb_width;
  1223. int tb_y = ctb_addr_rs / sps->ctb_width;
  1224. int tile_x = 0;
  1225. int tile_y = 0;
  1226. int val = 0;
  1227. for (i = 0; i < pps->num_tile_columns; i++) {
  1228. if (tb_x < pps->col_bd[i + 1]) {
  1229. tile_x = i;
  1230. break;
  1231. }
  1232. }
  1233. for (i = 0; i < pps->num_tile_rows; i++) {
  1234. if (tb_y < pps->row_bd[i + 1]) {
  1235. tile_y = i;
  1236. break;
  1237. }
  1238. }
  1239. for (i = 0; i < tile_x; i++)
  1240. val += pps->row_height[tile_y] * pps->column_width[i];
  1241. for (i = 0; i < tile_y; i++)
  1242. val += sps->ctb_width * pps->row_height[i];
  1243. val += (tb_y - pps->row_bd[tile_y]) * pps->column_width[tile_x] +
  1244. tb_x - pps->col_bd[tile_x];
  1245. pps->ctb_addr_rs_to_ts[ctb_addr_rs] = val;
  1246. pps->ctb_addr_ts_to_rs[val] = ctb_addr_rs;
  1247. }
  1248. for (j = 0, tile_id = 0; j < pps->num_tile_rows; j++)
  1249. for (i = 0; i < pps->num_tile_columns; i++, tile_id++)
  1250. for (y = pps->row_bd[j]; y < pps->row_bd[j + 1]; y++)
  1251. for (x = pps->col_bd[i]; x < pps->col_bd[i + 1]; x++)
  1252. pps->tile_id[pps->ctb_addr_rs_to_ts[y * sps->ctb_width + x]] = tile_id;
  1253. pps->tile_pos_rs = av_malloc_array(tile_id, sizeof(*pps->tile_pos_rs));
  1254. if (!pps->tile_pos_rs)
  1255. return AVERROR(ENOMEM);
  1256. for (j = 0; j < pps->num_tile_rows; j++)
  1257. for (i = 0; i < pps->num_tile_columns; i++)
  1258. pps->tile_pos_rs[j * pps->num_tile_columns + i] =
  1259. pps->row_bd[j] * sps->ctb_width + pps->col_bd[i];
  1260. log2_diff = sps->log2_ctb_size - sps->log2_min_tb_size;
  1261. pps->min_tb_addr_zs = &pps->min_tb_addr_zs_tab[1*(sps->tb_mask+2)+1];
  1262. for (y = 0; y < sps->tb_mask+2; y++) {
  1263. pps->min_tb_addr_zs_tab[y*(sps->tb_mask+2)] = -1;
  1264. pps->min_tb_addr_zs_tab[y] = -1;
  1265. }
  1266. for (y = 0; y < sps->tb_mask+1; y++) {
  1267. for (x = 0; x < sps->tb_mask+1; x++) {
  1268. int tb_x = x >> log2_diff;
  1269. int tb_y = y >> log2_diff;
  1270. int rs = sps->ctb_width * tb_y + tb_x;
  1271. int val = pps->ctb_addr_rs_to_ts[rs] << (log2_diff * 2);
  1272. for (i = 0; i < log2_diff; i++) {
  1273. int m = 1 << i;
  1274. val += (m & x ? m * m : 0) + (m & y ? 2 * m * m : 0);
  1275. }
  1276. pps->min_tb_addr_zs[y * (sps->tb_mask+2) + x] = val;
  1277. }
  1278. }
  1279. return 0;
  1280. }
  1281. int ff_hevc_decode_nal_pps(GetBitContext *gb, AVCodecContext *avctx,
  1282. HEVCParamSets *ps)
  1283. {
  1284. HEVCSPS *sps = NULL;
  1285. int i, ret = 0;
  1286. unsigned int pps_id = 0;
  1287. ptrdiff_t nal_size;
  1288. unsigned log2_parallel_merge_level_minus2;
  1289. AVBufferRef *pps_buf;
  1290. HEVCPPS *pps = av_mallocz(sizeof(*pps));
  1291. if (!pps)
  1292. return AVERROR(ENOMEM);
  1293. pps_buf = av_buffer_create((uint8_t *)pps, sizeof(*pps),
  1294. hevc_pps_free, NULL, 0);
  1295. if (!pps_buf) {
  1296. av_freep(&pps);
  1297. return AVERROR(ENOMEM);
  1298. }
  1299. av_log(avctx, AV_LOG_DEBUG, "Decoding PPS\n");
  1300. nal_size = gb->buffer_end - gb->buffer;
  1301. if (nal_size > sizeof(pps->data)) {
  1302. av_log(avctx, AV_LOG_WARNING, "Truncating likely oversized PPS "
  1303. "(%"PTRDIFF_SPECIFIER" > %"SIZE_SPECIFIER")\n",
  1304. nal_size, sizeof(pps->data));
  1305. pps->data_size = sizeof(pps->data);
  1306. } else {
  1307. pps->data_size = nal_size;
  1308. }
  1309. memcpy(pps->data, gb->buffer, pps->data_size);
  1310. // Default values
  1311. pps->loop_filter_across_tiles_enabled_flag = 1;
  1312. pps->num_tile_columns = 1;
  1313. pps->num_tile_rows = 1;
  1314. pps->uniform_spacing_flag = 1;
  1315. pps->disable_dbf = 0;
  1316. pps->beta_offset = 0;
  1317. pps->tc_offset = 0;
  1318. pps->log2_max_transform_skip_block_size = 2;
  1319. // Coded parameters
  1320. pps_id = get_ue_golomb_long(gb);
  1321. if (pps_id >= HEVC_MAX_PPS_COUNT) {
  1322. av_log(avctx, AV_LOG_ERROR, "PPS id out of range: %d\n", pps_id);
  1323. ret = AVERROR_INVALIDDATA;
  1324. goto err;
  1325. }
  1326. pps->sps_id = get_ue_golomb_long(gb);
  1327. if (pps->sps_id >= HEVC_MAX_SPS_COUNT) {
  1328. av_log(avctx, AV_LOG_ERROR, "SPS id out of range: %d\n", pps->sps_id);
  1329. ret = AVERROR_INVALIDDATA;
  1330. goto err;
  1331. }
  1332. if (!ps->sps_list[pps->sps_id]) {
  1333. av_log(avctx, AV_LOG_ERROR, "SPS %u does not exist.\n", pps->sps_id);
  1334. ret = AVERROR_INVALIDDATA;
  1335. goto err;
  1336. }
  1337. sps = (HEVCSPS *)ps->sps_list[pps->sps_id]->data;
  1338. pps->dependent_slice_segments_enabled_flag = get_bits1(gb);
  1339. pps->output_flag_present_flag = get_bits1(gb);
  1340. pps->num_extra_slice_header_bits = get_bits(gb, 3);
  1341. pps->sign_data_hiding_flag = get_bits1(gb);
  1342. pps->cabac_init_present_flag = get_bits1(gb);
  1343. pps->num_ref_idx_l0_default_active = get_ue_golomb_long(gb) + 1;
  1344. pps->num_ref_idx_l1_default_active = get_ue_golomb_long(gb) + 1;
  1345. pps->pic_init_qp_minus26 = get_se_golomb(gb);
  1346. pps->constrained_intra_pred_flag = get_bits1(gb);
  1347. pps->transform_skip_enabled_flag = get_bits1(gb);
  1348. pps->cu_qp_delta_enabled_flag = get_bits1(gb);
  1349. pps->diff_cu_qp_delta_depth = 0;
  1350. if (pps->cu_qp_delta_enabled_flag)
  1351. pps->diff_cu_qp_delta_depth = get_ue_golomb_long(gb);
  1352. if (pps->diff_cu_qp_delta_depth < 0 ||
  1353. pps->diff_cu_qp_delta_depth > sps->log2_diff_max_min_coding_block_size) {
  1354. av_log(avctx, AV_LOG_ERROR, "diff_cu_qp_delta_depth %d is invalid\n",
  1355. pps->diff_cu_qp_delta_depth);
  1356. ret = AVERROR_INVALIDDATA;
  1357. goto err;
  1358. }
  1359. pps->cb_qp_offset = get_se_golomb(gb);
  1360. if (pps->cb_qp_offset < -12 || pps->cb_qp_offset > 12) {
  1361. av_log(avctx, AV_LOG_ERROR, "pps_cb_qp_offset out of range: %d\n",
  1362. pps->cb_qp_offset);
  1363. ret = AVERROR_INVALIDDATA;
  1364. goto err;
  1365. }
  1366. pps->cr_qp_offset = get_se_golomb(gb);
  1367. if (pps->cr_qp_offset < -12 || pps->cr_qp_offset > 12) {
  1368. av_log(avctx, AV_LOG_ERROR, "pps_cr_qp_offset out of range: %d\n",
  1369. pps->cr_qp_offset);
  1370. ret = AVERROR_INVALIDDATA;
  1371. goto err;
  1372. }
  1373. pps->pic_slice_level_chroma_qp_offsets_present_flag = get_bits1(gb);
  1374. pps->weighted_pred_flag = get_bits1(gb);
  1375. pps->weighted_bipred_flag = get_bits1(gb);
  1376. pps->transquant_bypass_enable_flag = get_bits1(gb);
  1377. pps->tiles_enabled_flag = get_bits1(gb);
  1378. pps->entropy_coding_sync_enabled_flag = get_bits1(gb);
  1379. if (pps->tiles_enabled_flag) {
  1380. int num_tile_columns_minus1 = get_ue_golomb(gb);
  1381. int num_tile_rows_minus1 = get_ue_golomb(gb);
  1382. if (num_tile_columns_minus1 < 0 ||
  1383. num_tile_columns_minus1 >= sps->ctb_width - 1) {
  1384. av_log(avctx, AV_LOG_ERROR, "num_tile_columns_minus1 out of range: %d\n",
  1385. num_tile_columns_minus1);
  1386. ret = num_tile_columns_minus1 < 0 ? num_tile_columns_minus1 : AVERROR_INVALIDDATA;
  1387. goto err;
  1388. }
  1389. if (num_tile_rows_minus1 < 0 ||
  1390. num_tile_rows_minus1 >= sps->ctb_height - 1) {
  1391. av_log(avctx, AV_LOG_ERROR, "num_tile_rows_minus1 out of range: %d\n",
  1392. num_tile_rows_minus1);
  1393. ret = num_tile_rows_minus1 < 0 ? num_tile_rows_minus1 : AVERROR_INVALIDDATA;
  1394. goto err;
  1395. }
  1396. pps->num_tile_columns = num_tile_columns_minus1 + 1;
  1397. pps->num_tile_rows = num_tile_rows_minus1 + 1;
  1398. pps->column_width = av_malloc_array(pps->num_tile_columns, sizeof(*pps->column_width));
  1399. pps->row_height = av_malloc_array(pps->num_tile_rows, sizeof(*pps->row_height));
  1400. if (!pps->column_width || !pps->row_height) {
  1401. ret = AVERROR(ENOMEM);
  1402. goto err;
  1403. }
  1404. pps->uniform_spacing_flag = get_bits1(gb);
  1405. if (!pps->uniform_spacing_flag) {
  1406. uint64_t sum = 0;
  1407. for (i = 0; i < pps->num_tile_columns - 1; i++) {
  1408. pps->column_width[i] = get_ue_golomb_long(gb) + 1;
  1409. sum += pps->column_width[i];
  1410. }
  1411. if (sum >= sps->ctb_width) {
  1412. av_log(avctx, AV_LOG_ERROR, "Invalid tile widths.\n");
  1413. ret = AVERROR_INVALIDDATA;
  1414. goto err;
  1415. }
  1416. pps->column_width[pps->num_tile_columns - 1] = sps->ctb_width - sum;
  1417. sum = 0;
  1418. for (i = 0; i < pps->num_tile_rows - 1; i++) {
  1419. pps->row_height[i] = get_ue_golomb_long(gb) + 1;
  1420. sum += pps->row_height[i];
  1421. }
  1422. if (sum >= sps->ctb_height) {
  1423. av_log(avctx, AV_LOG_ERROR, "Invalid tile heights.\n");
  1424. ret = AVERROR_INVALIDDATA;
  1425. goto err;
  1426. }
  1427. pps->row_height[pps->num_tile_rows - 1] = sps->ctb_height - sum;
  1428. }
  1429. pps->loop_filter_across_tiles_enabled_flag = get_bits1(gb);
  1430. }
  1431. pps->seq_loop_filter_across_slices_enabled_flag = get_bits1(gb);
  1432. pps->deblocking_filter_control_present_flag = get_bits1(gb);
  1433. if (pps->deblocking_filter_control_present_flag) {
  1434. pps->deblocking_filter_override_enabled_flag = get_bits1(gb);
  1435. pps->disable_dbf = get_bits1(gb);
  1436. if (!pps->disable_dbf) {
  1437. int beta_offset_div2 = get_se_golomb(gb);
  1438. int tc_offset_div2 = get_se_golomb(gb) ;
  1439. if (beta_offset_div2 < -6 || beta_offset_div2 > 6) {
  1440. av_log(avctx, AV_LOG_ERROR, "pps_beta_offset_div2 out of range: %d\n",
  1441. beta_offset_div2);
  1442. ret = AVERROR_INVALIDDATA;
  1443. goto err;
  1444. }
  1445. if (tc_offset_div2 < -6 || tc_offset_div2 > 6) {
  1446. av_log(avctx, AV_LOG_ERROR, "pps_tc_offset_div2 out of range: %d\n",
  1447. tc_offset_div2);
  1448. ret = AVERROR_INVALIDDATA;
  1449. goto err;
  1450. }
  1451. pps->beta_offset = 2 * beta_offset_div2;
  1452. pps->tc_offset = 2 * tc_offset_div2;
  1453. }
  1454. }
  1455. pps->scaling_list_data_present_flag = get_bits1(gb);
  1456. if (pps->scaling_list_data_present_flag) {
  1457. set_default_scaling_list_data(&pps->scaling_list);
  1458. ret = scaling_list_data(gb, avctx, &pps->scaling_list, sps);
  1459. if (ret < 0)
  1460. goto err;
  1461. }
  1462. pps->lists_modification_present_flag = get_bits1(gb);
  1463. log2_parallel_merge_level_minus2 = get_ue_golomb_long(gb);
  1464. if (log2_parallel_merge_level_minus2 > sps->log2_ctb_size) {
  1465. av_log(avctx, AV_LOG_ERROR, "log2_parallel_merge_level_minus2 out of range: %d\n",
  1466. log2_parallel_merge_level_minus2);
  1467. ret = AVERROR_INVALIDDATA;
  1468. goto err;
  1469. }
  1470. pps->log2_parallel_merge_level = log2_parallel_merge_level_minus2 + 2;
  1471. pps->slice_header_extension_present_flag = get_bits1(gb);
  1472. if (get_bits1(gb)) { // pps_extension_present_flag
  1473. int pps_range_extensions_flag = get_bits1(gb);
  1474. skip_bits(gb, 7); // pps_extension_7bits
  1475. if (sps->ptl.general_ptl.profile_idc == FF_PROFILE_HEVC_REXT && pps_range_extensions_flag) {
  1476. if ((ret = pps_range_extensions(gb, avctx, pps, sps)) < 0)
  1477. goto err;
  1478. }
  1479. }
  1480. ret = setup_pps(avctx, gb, pps, sps);
  1481. if (ret < 0)
  1482. goto err;
  1483. if (get_bits_left(gb) < 0) {
  1484. av_log(avctx, AV_LOG_ERROR,
  1485. "Overread PPS by %d bits\n", -get_bits_left(gb));
  1486. goto err;
  1487. }
  1488. remove_pps(ps, pps_id);
  1489. ps->pps_list[pps_id] = pps_buf;
  1490. return 0;
  1491. err:
  1492. av_buffer_unref(&pps_buf);
  1493. return ret;
  1494. }
  1495. void ff_hevc_ps_uninit(HEVCParamSets *ps)
  1496. {
  1497. int i;
  1498. for (i = 0; i < FF_ARRAY_ELEMS(ps->vps_list); i++)
  1499. av_buffer_unref(&ps->vps_list[i]);
  1500. for (i = 0; i < FF_ARRAY_ELEMS(ps->sps_list); i++)
  1501. av_buffer_unref(&ps->sps_list[i]);
  1502. for (i = 0; i < FF_ARRAY_ELEMS(ps->pps_list); i++)
  1503. av_buffer_unref(&ps->pps_list[i]);
  1504. ps->sps = NULL;
  1505. ps->pps = NULL;
  1506. ps->vps = NULL;
  1507. }
  1508. int ff_hevc_compute_poc(const HEVCSPS *sps, int pocTid0, int poc_lsb, int nal_unit_type)
  1509. {
  1510. int max_poc_lsb = 1 << sps->log2_max_poc_lsb;
  1511. int prev_poc_lsb = pocTid0 % max_poc_lsb;
  1512. int prev_poc_msb = pocTid0 - prev_poc_lsb;
  1513. int poc_msb;
  1514. if (poc_lsb < prev_poc_lsb && prev_poc_lsb - poc_lsb >= max_poc_lsb / 2)
  1515. poc_msb = prev_poc_msb + max_poc_lsb;
  1516. else if (poc_lsb > prev_poc_lsb && poc_lsb - prev_poc_lsb > max_poc_lsb / 2)
  1517. poc_msb = prev_poc_msb - max_poc_lsb;
  1518. else
  1519. poc_msb = prev_poc_msb;
  1520. // For BLA picture types, POCmsb is set to 0.
  1521. if (nal_unit_type == HEVC_NAL_BLA_W_LP ||
  1522. nal_unit_type == HEVC_NAL_BLA_W_RADL ||
  1523. nal_unit_type == HEVC_NAL_BLA_N_LP)
  1524. poc_msb = 0;
  1525. return poc_msb + poc_lsb;
  1526. }