You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1126 lines
43KB

  1. /*
  2. * Copyright (c) 2010, Google, Inc.
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * AV1 encoder support via libaom
  23. */
  24. #define AOM_DISABLE_CTRL_TYPECHECKS 1
  25. #include <aom/aom_encoder.h>
  26. #include <aom/aomcx.h>
  27. #include "libavutil/avassert.h"
  28. #include "libavutil/base64.h"
  29. #include "libavutil/common.h"
  30. #include "libavutil/mathematics.h"
  31. #include "libavutil/opt.h"
  32. #include "libavutil/pixdesc.h"
  33. #include "av1.h"
  34. #include "avcodec.h"
  35. #include "internal.h"
  36. #include "profiles.h"
  37. /*
  38. * Portion of struct aom_codec_cx_pkt from aom_encoder.h.
  39. * One encoded frame returned from the library.
  40. */
  41. struct FrameListData {
  42. void *buf; /**< compressed data buffer */
  43. size_t sz; /**< length of compressed data */
  44. int64_t pts; /**< time stamp to show frame
  45. (in timebase units) */
  46. unsigned long duration; /**< duration to show frame
  47. (in timebase units) */
  48. uint32_t flags; /**< flags for this frame */
  49. uint64_t sse[4];
  50. int have_sse; /**< true if we have pending sse[] */
  51. uint64_t frame_number;
  52. struct FrameListData *next;
  53. };
  54. typedef struct AOMEncoderContext {
  55. AVClass *class;
  56. AVBSFContext *bsf;
  57. struct aom_codec_ctx encoder;
  58. struct aom_image rawimg;
  59. struct aom_fixed_buf twopass_stats;
  60. struct FrameListData *coded_frame_list;
  61. int cpu_used;
  62. int auto_alt_ref;
  63. int arnr_max_frames;
  64. int arnr_strength;
  65. int aq_mode;
  66. int lag_in_frames;
  67. int error_resilient;
  68. int crf;
  69. int static_thresh;
  70. int drop_threshold;
  71. int denoise_noise_level;
  72. int denoise_block_size;
  73. uint64_t sse[4];
  74. int have_sse; /**< true if we have pending sse[] */
  75. uint64_t frame_number;
  76. int rc_undershoot_pct;
  77. int rc_overshoot_pct;
  78. int minsection_pct;
  79. int maxsection_pct;
  80. int frame_parallel;
  81. int tile_cols, tile_rows;
  82. int tile_cols_log2, tile_rows_log2;
  83. aom_superblock_size_t superblock_size;
  84. int uniform_tiles;
  85. int row_mt;
  86. int enable_cdef;
  87. int enable_global_motion;
  88. int enable_intrabc;
  89. } AOMContext;
  90. static const char *const ctlidstr[] = {
  91. [AOME_SET_CPUUSED] = "AOME_SET_CPUUSED",
  92. [AOME_SET_CQ_LEVEL] = "AOME_SET_CQ_LEVEL",
  93. [AOME_SET_ENABLEAUTOALTREF] = "AOME_SET_ENABLEAUTOALTREF",
  94. [AOME_SET_ARNR_MAXFRAMES] = "AOME_SET_ARNR_MAXFRAMES",
  95. [AOME_SET_ARNR_STRENGTH] = "AOME_SET_ARNR_STRENGTH",
  96. [AOME_SET_STATIC_THRESHOLD] = "AOME_SET_STATIC_THRESHOLD",
  97. [AV1E_SET_COLOR_RANGE] = "AV1E_SET_COLOR_RANGE",
  98. [AV1E_SET_COLOR_PRIMARIES] = "AV1E_SET_COLOR_PRIMARIES",
  99. [AV1E_SET_MATRIX_COEFFICIENTS] = "AV1E_SET_MATRIX_COEFFICIENTS",
  100. [AV1E_SET_TRANSFER_CHARACTERISTICS] = "AV1E_SET_TRANSFER_CHARACTERISTICS",
  101. [AV1E_SET_AQ_MODE] = "AV1E_SET_AQ_MODE",
  102. [AV1E_SET_FRAME_PARALLEL_DECODING] = "AV1E_SET_FRAME_PARALLEL_DECODING",
  103. [AV1E_SET_SUPERBLOCK_SIZE] = "AV1E_SET_SUPERBLOCK_SIZE",
  104. [AV1E_SET_TILE_COLUMNS] = "AV1E_SET_TILE_COLUMNS",
  105. [AV1E_SET_TILE_ROWS] = "AV1E_SET_TILE_ROWS",
  106. #ifdef AOM_CTRL_AV1E_SET_ROW_MT
  107. [AV1E_SET_ROW_MT] = "AV1E_SET_ROW_MT",
  108. #endif
  109. #ifdef AOM_CTRL_AV1E_SET_DENOISE_NOISE_LEVEL
  110. [AV1E_SET_DENOISE_NOISE_LEVEL] = "AV1E_SET_DENOISE_NOISE_LEVEL",
  111. #endif
  112. #ifdef AOM_CTRL_AV1E_SET_DENOISE_BLOCK_SIZE
  113. [AV1E_SET_DENOISE_BLOCK_SIZE] = "AV1E_SET_DENOISE_BLOCK_SIZE",
  114. #endif
  115. #ifdef AOM_CTRL_AV1E_SET_MAX_REFERENCE_FRAMES
  116. [AV1E_SET_MAX_REFERENCE_FRAMES] = "AV1E_SET_MAX_REFERENCE_FRAMES",
  117. #endif
  118. #ifdef AOM_CTRL_AV1E_SET_ENABLE_GLOBAL_MOTION
  119. [AV1E_SET_ENABLE_GLOBAL_MOTION] = "AV1E_SET_ENABLE_GLOBAL_MOTION",
  120. #endif
  121. #ifdef AOM_CTRL_AV1E_SET_ENABLE_INTRABC
  122. [AV1E_SET_ENABLE_INTRABC] = "AV1E_SET_ENABLE_INTRABC",
  123. #endif
  124. [AV1E_SET_ENABLE_CDEF] = "AV1E_SET_ENABLE_CDEF",
  125. };
  126. static av_cold void log_encoder_error(AVCodecContext *avctx, const char *desc)
  127. {
  128. AOMContext *ctx = avctx->priv_data;
  129. const char *error = aom_codec_error(&ctx->encoder);
  130. const char *detail = aom_codec_error_detail(&ctx->encoder);
  131. av_log(avctx, AV_LOG_ERROR, "%s: %s\n", desc, error);
  132. if (detail)
  133. av_log(avctx, AV_LOG_ERROR, " Additional information: %s\n", detail);
  134. }
  135. static av_cold void dump_enc_cfg(AVCodecContext *avctx,
  136. const struct aom_codec_enc_cfg *cfg)
  137. {
  138. int width = -30;
  139. int level = AV_LOG_DEBUG;
  140. av_log(avctx, level, "aom_codec_enc_cfg\n");
  141. av_log(avctx, level, "generic settings\n"
  142. " %*s%u\n %*s%u\n %*s%u\n %*s%u\n %*s%u\n"
  143. " %*s%u\n %*s%u\n"
  144. " %*s{%u/%u}\n %*s%u\n %*s%d\n %*s%u\n",
  145. width, "g_usage:", cfg->g_usage,
  146. width, "g_threads:", cfg->g_threads,
  147. width, "g_profile:", cfg->g_profile,
  148. width, "g_w:", cfg->g_w,
  149. width, "g_h:", cfg->g_h,
  150. width, "g_bit_depth:", cfg->g_bit_depth,
  151. width, "g_input_bit_depth:", cfg->g_input_bit_depth,
  152. width, "g_timebase:", cfg->g_timebase.num, cfg->g_timebase.den,
  153. width, "g_error_resilient:", cfg->g_error_resilient,
  154. width, "g_pass:", cfg->g_pass,
  155. width, "g_lag_in_frames:", cfg->g_lag_in_frames);
  156. av_log(avctx, level, "rate control settings\n"
  157. " %*s%u\n %*s%d\n %*s%p(%"SIZE_SPECIFIER")\n %*s%u\n",
  158. width, "rc_dropframe_thresh:", cfg->rc_dropframe_thresh,
  159. width, "rc_end_usage:", cfg->rc_end_usage,
  160. width, "rc_twopass_stats_in:", cfg->rc_twopass_stats_in.buf, cfg->rc_twopass_stats_in.sz,
  161. width, "rc_target_bitrate:", cfg->rc_target_bitrate);
  162. av_log(avctx, level, "quantizer settings\n"
  163. " %*s%u\n %*s%u\n",
  164. width, "rc_min_quantizer:", cfg->rc_min_quantizer,
  165. width, "rc_max_quantizer:", cfg->rc_max_quantizer);
  166. av_log(avctx, level, "bitrate tolerance\n"
  167. " %*s%u\n %*s%u\n",
  168. width, "rc_undershoot_pct:", cfg->rc_undershoot_pct,
  169. width, "rc_overshoot_pct:", cfg->rc_overshoot_pct);
  170. av_log(avctx, level, "decoder buffer model\n"
  171. " %*s%u\n %*s%u\n %*s%u\n",
  172. width, "rc_buf_sz:", cfg->rc_buf_sz,
  173. width, "rc_buf_initial_sz:", cfg->rc_buf_initial_sz,
  174. width, "rc_buf_optimal_sz:", cfg->rc_buf_optimal_sz);
  175. av_log(avctx, level, "2 pass rate control settings\n"
  176. " %*s%u\n %*s%u\n %*s%u\n",
  177. width, "rc_2pass_vbr_bias_pct:", cfg->rc_2pass_vbr_bias_pct,
  178. width, "rc_2pass_vbr_minsection_pct:", cfg->rc_2pass_vbr_minsection_pct,
  179. width, "rc_2pass_vbr_maxsection_pct:", cfg->rc_2pass_vbr_maxsection_pct);
  180. av_log(avctx, level, "keyframing settings\n"
  181. " %*s%d\n %*s%u\n %*s%u\n",
  182. width, "kf_mode:", cfg->kf_mode,
  183. width, "kf_min_dist:", cfg->kf_min_dist,
  184. width, "kf_max_dist:", cfg->kf_max_dist);
  185. av_log(avctx, level, "tile settings\n"
  186. " %*s%d\n %*s%d\n",
  187. width, "tile_width_count:", cfg->tile_width_count,
  188. width, "tile_height_count:", cfg->tile_height_count);
  189. av_log(avctx, level, "\n");
  190. }
  191. static void coded_frame_add(void *list, struct FrameListData *cx_frame)
  192. {
  193. struct FrameListData **p = list;
  194. while (*p)
  195. p = &(*p)->next;
  196. *p = cx_frame;
  197. cx_frame->next = NULL;
  198. }
  199. static av_cold void free_coded_frame(struct FrameListData *cx_frame)
  200. {
  201. av_freep(&cx_frame->buf);
  202. av_freep(&cx_frame);
  203. }
  204. static av_cold void free_frame_list(struct FrameListData *list)
  205. {
  206. struct FrameListData *p = list;
  207. while (p) {
  208. list = list->next;
  209. free_coded_frame(p);
  210. p = list;
  211. }
  212. }
  213. static av_cold int codecctl_int(AVCodecContext *avctx,
  214. #ifdef UENUM1BYTE
  215. aome_enc_control_id id,
  216. #else
  217. enum aome_enc_control_id id,
  218. #endif
  219. int val)
  220. {
  221. AOMContext *ctx = avctx->priv_data;
  222. char buf[80];
  223. int width = -30;
  224. int res;
  225. snprintf(buf, sizeof(buf), "%s:", ctlidstr[id]);
  226. av_log(avctx, AV_LOG_DEBUG, " %*s%d\n", width, buf, val);
  227. res = aom_codec_control(&ctx->encoder, id, val);
  228. if (res != AOM_CODEC_OK) {
  229. snprintf(buf, sizeof(buf), "Failed to set %s codec control",
  230. ctlidstr[id]);
  231. log_encoder_error(avctx, buf);
  232. return AVERROR(EINVAL);
  233. }
  234. return 0;
  235. }
  236. static av_cold int aom_free(AVCodecContext *avctx)
  237. {
  238. AOMContext *ctx = avctx->priv_data;
  239. aom_codec_destroy(&ctx->encoder);
  240. av_freep(&ctx->twopass_stats.buf);
  241. av_freep(&avctx->stats_out);
  242. free_frame_list(ctx->coded_frame_list);
  243. av_bsf_free(&ctx->bsf);
  244. return 0;
  245. }
  246. static int set_pix_fmt(AVCodecContext *avctx, aom_codec_caps_t codec_caps,
  247. struct aom_codec_enc_cfg *enccfg, aom_codec_flags_t *flags,
  248. aom_img_fmt_t *img_fmt)
  249. {
  250. AOMContext av_unused *ctx = avctx->priv_data;
  251. enccfg->g_bit_depth = enccfg->g_input_bit_depth = 8;
  252. switch (avctx->pix_fmt) {
  253. case AV_PIX_FMT_YUV420P:
  254. enccfg->g_profile = FF_PROFILE_AV1_MAIN;
  255. *img_fmt = AOM_IMG_FMT_I420;
  256. return 0;
  257. case AV_PIX_FMT_YUV422P:
  258. enccfg->g_profile = FF_PROFILE_AV1_PROFESSIONAL;
  259. *img_fmt = AOM_IMG_FMT_I422;
  260. return 0;
  261. case AV_PIX_FMT_YUV444P:
  262. enccfg->g_profile = FF_PROFILE_AV1_HIGH;
  263. *img_fmt = AOM_IMG_FMT_I444;
  264. return 0;
  265. case AV_PIX_FMT_YUV420P10:
  266. case AV_PIX_FMT_YUV420P12:
  267. if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
  268. enccfg->g_bit_depth = enccfg->g_input_bit_depth =
  269. avctx->pix_fmt == AV_PIX_FMT_YUV420P10 ? 10 : 12;
  270. enccfg->g_profile =
  271. enccfg->g_bit_depth == 10 ? FF_PROFILE_AV1_MAIN : FF_PROFILE_AV1_PROFESSIONAL;
  272. *img_fmt = AOM_IMG_FMT_I42016;
  273. *flags |= AOM_CODEC_USE_HIGHBITDEPTH;
  274. return 0;
  275. }
  276. break;
  277. case AV_PIX_FMT_YUV422P10:
  278. case AV_PIX_FMT_YUV422P12:
  279. if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
  280. enccfg->g_bit_depth = enccfg->g_input_bit_depth =
  281. avctx->pix_fmt == AV_PIX_FMT_YUV422P10 ? 10 : 12;
  282. enccfg->g_profile = FF_PROFILE_AV1_PROFESSIONAL;
  283. *img_fmt = AOM_IMG_FMT_I42216;
  284. *flags |= AOM_CODEC_USE_HIGHBITDEPTH;
  285. return 0;
  286. }
  287. break;
  288. case AV_PIX_FMT_YUV444P10:
  289. case AV_PIX_FMT_YUV444P12:
  290. if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH) {
  291. enccfg->g_bit_depth = enccfg->g_input_bit_depth =
  292. avctx->pix_fmt == AV_PIX_FMT_YUV444P10 ? 10 : 12;
  293. enccfg->g_profile =
  294. enccfg->g_bit_depth == 10 ? FF_PROFILE_AV1_HIGH : FF_PROFILE_AV1_PROFESSIONAL;
  295. *img_fmt = AOM_IMG_FMT_I44416;
  296. *flags |= AOM_CODEC_USE_HIGHBITDEPTH;
  297. return 0;
  298. }
  299. break;
  300. default:
  301. break;
  302. }
  303. av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format.\n");
  304. return AVERROR_INVALIDDATA;
  305. }
  306. static void set_color_range(AVCodecContext *avctx)
  307. {
  308. aom_color_range_t aom_cr;
  309. switch (avctx->color_range) {
  310. case AVCOL_RANGE_UNSPECIFIED:
  311. case AVCOL_RANGE_MPEG: aom_cr = AOM_CR_STUDIO_RANGE; break;
  312. case AVCOL_RANGE_JPEG: aom_cr = AOM_CR_FULL_RANGE; break;
  313. default:
  314. av_log(avctx, AV_LOG_WARNING, "Unsupported color range (%d)\n",
  315. avctx->color_range);
  316. return;
  317. }
  318. codecctl_int(avctx, AV1E_SET_COLOR_RANGE, aom_cr);
  319. }
  320. static int count_uniform_tiling(int dim, int sb_size, int tiles_log2)
  321. {
  322. int sb_dim = (dim + sb_size - 1) / sb_size;
  323. int tile_dim = (sb_dim + (1 << tiles_log2) - 1) >> tiles_log2;
  324. av_assert0(tile_dim > 0);
  325. return (sb_dim + tile_dim - 1) / tile_dim;
  326. }
  327. static int choose_tiling(AVCodecContext *avctx,
  328. struct aom_codec_enc_cfg *enccfg)
  329. {
  330. AOMContext *ctx = avctx->priv_data;
  331. int sb_128x128_possible, sb_size, sb_width, sb_height;
  332. int uniform_rows, uniform_cols;
  333. int uniform_64x64_possible, uniform_128x128_possible;
  334. int tile_size, rounding, i;
  335. if (ctx->tile_cols_log2 >= 0)
  336. ctx->tile_cols = 1 << ctx->tile_cols_log2;
  337. if (ctx->tile_rows_log2 >= 0)
  338. ctx->tile_rows = 1 << ctx->tile_rows_log2;
  339. if (ctx->tile_cols == 0) {
  340. ctx->tile_cols = (avctx->width + AV1_MAX_TILE_WIDTH - 1) /
  341. AV1_MAX_TILE_WIDTH;
  342. if (ctx->tile_cols > 1) {
  343. av_log(avctx, AV_LOG_DEBUG, "Automatically using %d tile "
  344. "columns to fill width.\n", ctx->tile_cols);
  345. }
  346. }
  347. av_assert0(ctx->tile_cols > 0);
  348. if (ctx->tile_rows == 0) {
  349. int max_tile_width =
  350. FFALIGN((FFALIGN(avctx->width, 128) +
  351. ctx->tile_cols - 1) / ctx->tile_cols, 128);
  352. ctx->tile_rows =
  353. (max_tile_width * FFALIGN(avctx->height, 128) +
  354. AV1_MAX_TILE_AREA - 1) / AV1_MAX_TILE_AREA;
  355. if (ctx->tile_rows > 1) {
  356. av_log(avctx, AV_LOG_DEBUG, "Automatically using %d tile "
  357. "rows to fill area.\n", ctx->tile_rows);
  358. }
  359. }
  360. av_assert0(ctx->tile_rows > 0);
  361. if ((avctx->width + 63) / 64 < ctx->tile_cols ||
  362. (avctx->height + 63) / 64 < ctx->tile_rows) {
  363. av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: frame not "
  364. "large enough to fit specified tile arrangement.\n");
  365. return AVERROR(EINVAL);
  366. }
  367. if (ctx->tile_cols > AV1_MAX_TILE_COLS ||
  368. ctx->tile_rows > AV1_MAX_TILE_ROWS) {
  369. av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: AV1 does "
  370. "not allow more than %dx%d tiles.\n",
  371. AV1_MAX_TILE_COLS, AV1_MAX_TILE_ROWS);
  372. return AVERROR(EINVAL);
  373. }
  374. if (avctx->width / ctx->tile_cols > AV1_MAX_TILE_WIDTH) {
  375. av_log(avctx, AV_LOG_ERROR, "Invalid tile sizing: AV1 does "
  376. "not allow tiles of width greater than %d.\n",
  377. AV1_MAX_TILE_WIDTH);
  378. return AVERROR(EINVAL);
  379. }
  380. ctx->superblock_size = AOM_SUPERBLOCK_SIZE_DYNAMIC;
  381. if (ctx->tile_cols == 1 && ctx->tile_rows == 1) {
  382. av_log(avctx, AV_LOG_DEBUG, "Using a single tile.\n");
  383. return 0;
  384. }
  385. sb_128x128_possible =
  386. (avctx->width + 127) / 128 >= ctx->tile_cols &&
  387. (avctx->height + 127) / 128 >= ctx->tile_rows;
  388. ctx->tile_cols_log2 = ctx->tile_cols == 1 ? 0 :
  389. av_log2(ctx->tile_cols - 1) + 1;
  390. ctx->tile_rows_log2 = ctx->tile_rows == 1 ? 0 :
  391. av_log2(ctx->tile_rows - 1) + 1;
  392. uniform_cols = count_uniform_tiling(avctx->width,
  393. 64, ctx->tile_cols_log2);
  394. uniform_rows = count_uniform_tiling(avctx->height,
  395. 64, ctx->tile_rows_log2);
  396. av_log(avctx, AV_LOG_DEBUG, "Uniform with 64x64 superblocks "
  397. "-> %dx%d tiles.\n", uniform_cols, uniform_rows);
  398. uniform_64x64_possible = uniform_cols == ctx->tile_cols &&
  399. uniform_rows == ctx->tile_rows;
  400. if (sb_128x128_possible) {
  401. uniform_cols = count_uniform_tiling(avctx->width,
  402. 128, ctx->tile_cols_log2);
  403. uniform_rows = count_uniform_tiling(avctx->height,
  404. 128, ctx->tile_rows_log2);
  405. av_log(avctx, AV_LOG_DEBUG, "Uniform with 128x128 superblocks "
  406. "-> %dx%d tiles.\n", uniform_cols, uniform_rows);
  407. uniform_128x128_possible = uniform_cols == ctx->tile_cols &&
  408. uniform_rows == ctx->tile_rows;
  409. } else {
  410. av_log(avctx, AV_LOG_DEBUG, "128x128 superblocks not possible.\n");
  411. uniform_128x128_possible = 0;
  412. }
  413. ctx->uniform_tiles = 1;
  414. if (uniform_64x64_possible && uniform_128x128_possible) {
  415. av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with dynamic "
  416. "superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
  417. ctx->tile_cols_log2, ctx->tile_rows_log2);
  418. return 0;
  419. }
  420. if (uniform_64x64_possible && !sb_128x128_possible) {
  421. av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with 64x64 "
  422. "superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
  423. ctx->tile_cols_log2, ctx->tile_rows_log2);
  424. ctx->superblock_size = AOM_SUPERBLOCK_SIZE_64X64;
  425. return 0;
  426. }
  427. if (uniform_128x128_possible) {
  428. av_log(avctx, AV_LOG_DEBUG, "Using uniform tiling with 128x128 "
  429. "superblocks (tile_cols_log2 = %d, tile_rows_log2 = %d).\n",
  430. ctx->tile_cols_log2, ctx->tile_rows_log2);
  431. ctx->superblock_size = AOM_SUPERBLOCK_SIZE_128X128;
  432. return 0;
  433. }
  434. ctx->uniform_tiles = 0;
  435. if (sb_128x128_possible) {
  436. sb_size = 128;
  437. ctx->superblock_size = AOM_SUPERBLOCK_SIZE_128X128;
  438. } else {
  439. sb_size = 64;
  440. ctx->superblock_size = AOM_SUPERBLOCK_SIZE_64X64;
  441. }
  442. av_log(avctx, AV_LOG_DEBUG, "Using fixed tiling with %dx%d "
  443. "superblocks (tile_cols = %d, tile_rows = %d).\n",
  444. sb_size, sb_size, ctx->tile_cols, ctx->tile_rows);
  445. enccfg->tile_width_count = ctx->tile_cols;
  446. enccfg->tile_height_count = ctx->tile_rows;
  447. sb_width = (avctx->width + sb_size - 1) / sb_size;
  448. sb_height = (avctx->height + sb_size - 1) / sb_size;
  449. tile_size = sb_width / ctx->tile_cols;
  450. rounding = sb_width % ctx->tile_cols;
  451. for (i = 0; i < ctx->tile_cols; i++) {
  452. enccfg->tile_widths[i] = tile_size +
  453. (i < rounding / 2 ||
  454. i > ctx->tile_cols - 1 - (rounding + 1) / 2);
  455. }
  456. tile_size = sb_height / ctx->tile_rows;
  457. rounding = sb_height % ctx->tile_rows;
  458. for (i = 0; i < ctx->tile_rows; i++) {
  459. enccfg->tile_heights[i] = tile_size +
  460. (i < rounding / 2 ||
  461. i > ctx->tile_rows - 1 - (rounding + 1) / 2);
  462. }
  463. return 0;
  464. }
  465. static av_cold int aom_init(AVCodecContext *avctx,
  466. const struct aom_codec_iface *iface)
  467. {
  468. AOMContext *ctx = avctx->priv_data;
  469. struct aom_codec_enc_cfg enccfg = { 0 };
  470. #ifdef AOM_FRAME_IS_INTRAONLY
  471. aom_codec_flags_t flags =
  472. (avctx->flags & AV_CODEC_FLAG_PSNR) ? AOM_CODEC_USE_PSNR : 0;
  473. #else
  474. aom_codec_flags_t flags = 0;
  475. #endif
  476. AVCPBProperties *cpb_props;
  477. int res;
  478. aom_img_fmt_t img_fmt;
  479. aom_codec_caps_t codec_caps = aom_codec_get_caps(iface);
  480. av_log(avctx, AV_LOG_INFO, "%s\n", aom_codec_version_str());
  481. av_log(avctx, AV_LOG_VERBOSE, "%s\n", aom_codec_build_config());
  482. if ((res = aom_codec_enc_config_default(iface, &enccfg, 0)) != AOM_CODEC_OK) {
  483. av_log(avctx, AV_LOG_ERROR, "Failed to get config: %s\n",
  484. aom_codec_err_to_string(res));
  485. return AVERROR(EINVAL);
  486. }
  487. if (set_pix_fmt(avctx, codec_caps, &enccfg, &flags, &img_fmt))
  488. return AVERROR(EINVAL);
  489. if(!avctx->bit_rate)
  490. if(avctx->rc_max_rate || avctx->rc_buffer_size || avctx->rc_initial_buffer_occupancy) {
  491. av_log( avctx, AV_LOG_ERROR, "Rate control parameters set without a bitrate\n");
  492. return AVERROR(EINVAL);
  493. }
  494. dump_enc_cfg(avctx, &enccfg);
  495. enccfg.g_w = avctx->width;
  496. enccfg.g_h = avctx->height;
  497. enccfg.g_timebase.num = avctx->time_base.num;
  498. enccfg.g_timebase.den = avctx->time_base.den;
  499. enccfg.g_threads =
  500. FFMIN(avctx->thread_count ? avctx->thread_count : av_cpu_count(), 64);
  501. if (ctx->lag_in_frames >= 0)
  502. enccfg.g_lag_in_frames = ctx->lag_in_frames;
  503. if (avctx->flags & AV_CODEC_FLAG_PASS1)
  504. enccfg.g_pass = AOM_RC_FIRST_PASS;
  505. else if (avctx->flags & AV_CODEC_FLAG_PASS2)
  506. enccfg.g_pass = AOM_RC_LAST_PASS;
  507. else
  508. enccfg.g_pass = AOM_RC_ONE_PASS;
  509. if (avctx->rc_min_rate == avctx->rc_max_rate &&
  510. avctx->rc_min_rate == avctx->bit_rate && avctx->bit_rate) {
  511. enccfg.rc_end_usage = AOM_CBR;
  512. } else if (ctx->crf >= 0) {
  513. enccfg.rc_end_usage = AOM_CQ;
  514. if (!avctx->bit_rate)
  515. enccfg.rc_end_usage = AOM_Q;
  516. }
  517. if (avctx->bit_rate) {
  518. enccfg.rc_target_bitrate = av_rescale_rnd(avctx->bit_rate, 1, 1000,
  519. AV_ROUND_NEAR_INF);
  520. } else if (enccfg.rc_end_usage != AOM_Q) {
  521. if (enccfg.rc_end_usage == AOM_CQ) {
  522. enccfg.rc_target_bitrate = 1000000;
  523. } else {
  524. enccfg.rc_end_usage = AOM_Q;
  525. ctx->crf = 32;
  526. av_log(avctx, AV_LOG_WARNING,
  527. "Neither bitrate nor constrained quality specified, using default CRF of %d\n",
  528. ctx->crf);
  529. }
  530. }
  531. if (avctx->qmin >= 0)
  532. enccfg.rc_min_quantizer = avctx->qmin;
  533. if (avctx->qmax >= 0)
  534. enccfg.rc_max_quantizer = avctx->qmax;
  535. if (enccfg.rc_end_usage == AOM_CQ || enccfg.rc_end_usage == AOM_Q) {
  536. if (ctx->crf < enccfg.rc_min_quantizer || ctx->crf > enccfg.rc_max_quantizer) {
  537. av_log(avctx, AV_LOG_ERROR,
  538. "CQ level %d must be between minimum and maximum quantizer value (%d-%d)\n",
  539. ctx->crf, enccfg.rc_min_quantizer, enccfg.rc_max_quantizer);
  540. return AVERROR(EINVAL);
  541. }
  542. }
  543. enccfg.rc_dropframe_thresh = ctx->drop_threshold;
  544. // 0-100 (0 => CBR, 100 => VBR)
  545. enccfg.rc_2pass_vbr_bias_pct = round(avctx->qcompress * 100);
  546. if (ctx->minsection_pct >= 0)
  547. enccfg.rc_2pass_vbr_minsection_pct = ctx->minsection_pct;
  548. else if (avctx->bit_rate)
  549. enccfg.rc_2pass_vbr_minsection_pct =
  550. avctx->rc_min_rate * 100LL / avctx->bit_rate;
  551. if (ctx->maxsection_pct >= 0)
  552. enccfg.rc_2pass_vbr_maxsection_pct = ctx->maxsection_pct;
  553. else if (avctx->rc_max_rate)
  554. enccfg.rc_2pass_vbr_maxsection_pct =
  555. avctx->rc_max_rate * 100LL / avctx->bit_rate;
  556. if (avctx->rc_buffer_size)
  557. enccfg.rc_buf_sz =
  558. avctx->rc_buffer_size * 1000LL / avctx->bit_rate;
  559. if (avctx->rc_initial_buffer_occupancy)
  560. enccfg.rc_buf_initial_sz =
  561. avctx->rc_initial_buffer_occupancy * 1000LL / avctx->bit_rate;
  562. enccfg.rc_buf_optimal_sz = enccfg.rc_buf_sz * 5 / 6;
  563. if (ctx->rc_undershoot_pct >= 0)
  564. enccfg.rc_undershoot_pct = ctx->rc_undershoot_pct;
  565. if (ctx->rc_overshoot_pct >= 0)
  566. enccfg.rc_overshoot_pct = ctx->rc_overshoot_pct;
  567. // _enc_init() will balk if kf_min_dist differs from max w/AOM_KF_AUTO
  568. if (avctx->keyint_min >= 0 && avctx->keyint_min == avctx->gop_size)
  569. enccfg.kf_min_dist = avctx->keyint_min;
  570. if (avctx->gop_size >= 0)
  571. enccfg.kf_max_dist = avctx->gop_size;
  572. if (enccfg.g_pass == AOM_RC_FIRST_PASS)
  573. enccfg.g_lag_in_frames = 0;
  574. else if (enccfg.g_pass == AOM_RC_LAST_PASS) {
  575. int decode_size, ret;
  576. if (!avctx->stats_in) {
  577. av_log(avctx, AV_LOG_ERROR, "No stats file for second pass\n");
  578. return AVERROR_INVALIDDATA;
  579. }
  580. ctx->twopass_stats.sz = strlen(avctx->stats_in) * 3 / 4;
  581. ret = av_reallocp(&ctx->twopass_stats.buf, ctx->twopass_stats.sz);
  582. if (ret < 0) {
  583. av_log(avctx, AV_LOG_ERROR,
  584. "Stat buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
  585. ctx->twopass_stats.sz);
  586. ctx->twopass_stats.sz = 0;
  587. return ret;
  588. }
  589. decode_size = av_base64_decode(ctx->twopass_stats.buf, avctx->stats_in,
  590. ctx->twopass_stats.sz);
  591. if (decode_size < 0) {
  592. av_log(avctx, AV_LOG_ERROR, "Stat buffer decode failed\n");
  593. return AVERROR_INVALIDDATA;
  594. }
  595. ctx->twopass_stats.sz = decode_size;
  596. enccfg.rc_twopass_stats_in = ctx->twopass_stats;
  597. }
  598. /* 0-3: For non-zero values the encoder increasingly optimizes for reduced
  599. * complexity playback on low powered devices at the expense of encode
  600. * quality. */
  601. if (avctx->profile != FF_PROFILE_UNKNOWN)
  602. enccfg.g_profile = avctx->profile;
  603. enccfg.g_error_resilient = ctx->error_resilient;
  604. res = choose_tiling(avctx, &enccfg);
  605. if (res < 0)
  606. return res;
  607. dump_enc_cfg(avctx, &enccfg);
  608. /* Construct Encoder Context */
  609. res = aom_codec_enc_init(&ctx->encoder, iface, &enccfg, flags);
  610. if (res != AOM_CODEC_OK) {
  611. log_encoder_error(avctx, "Failed to initialize encoder");
  612. return AVERROR(EINVAL);
  613. }
  614. // codec control failures are currently treated only as warnings
  615. av_log(avctx, AV_LOG_DEBUG, "aom_codec_control\n");
  616. codecctl_int(avctx, AOME_SET_CPUUSED, ctx->cpu_used);
  617. if (ctx->auto_alt_ref >= 0)
  618. codecctl_int(avctx, AOME_SET_ENABLEAUTOALTREF, ctx->auto_alt_ref);
  619. if (ctx->arnr_max_frames >= 0)
  620. codecctl_int(avctx, AOME_SET_ARNR_MAXFRAMES, ctx->arnr_max_frames);
  621. if (ctx->arnr_strength >= 0)
  622. codecctl_int(avctx, AOME_SET_ARNR_STRENGTH, ctx->arnr_strength);
  623. if (ctx->enable_cdef >= 0)
  624. codecctl_int(avctx, AV1E_SET_ENABLE_CDEF, ctx->enable_cdef);
  625. codecctl_int(avctx, AOME_SET_STATIC_THRESHOLD, ctx->static_thresh);
  626. if (ctx->crf >= 0)
  627. codecctl_int(avctx, AOME_SET_CQ_LEVEL, ctx->crf);
  628. codecctl_int(avctx, AV1E_SET_COLOR_PRIMARIES, avctx->color_primaries);
  629. codecctl_int(avctx, AV1E_SET_MATRIX_COEFFICIENTS, avctx->colorspace);
  630. codecctl_int(avctx, AV1E_SET_TRANSFER_CHARACTERISTICS, avctx->color_trc);
  631. if (ctx->aq_mode >= 0)
  632. codecctl_int(avctx, AV1E_SET_AQ_MODE, ctx->aq_mode);
  633. if (ctx->frame_parallel >= 0)
  634. codecctl_int(avctx, AV1E_SET_FRAME_PARALLEL_DECODING, ctx->frame_parallel);
  635. set_color_range(avctx);
  636. codecctl_int(avctx, AV1E_SET_SUPERBLOCK_SIZE, ctx->superblock_size);
  637. if (ctx->uniform_tiles) {
  638. codecctl_int(avctx, AV1E_SET_TILE_COLUMNS, ctx->tile_cols_log2);
  639. codecctl_int(avctx, AV1E_SET_TILE_ROWS, ctx->tile_rows_log2);
  640. }
  641. #ifdef AOM_CTRL_AV1E_SET_DENOISE_NOISE_LEVEL
  642. if (ctx->denoise_noise_level >= 0)
  643. codecctl_int(avctx, AV1E_SET_DENOISE_NOISE_LEVEL, ctx->denoise_noise_level);
  644. #endif
  645. #ifdef AOM_CTRL_AV1E_SET_DENOISE_BLOCK_SIZE
  646. if (ctx->denoise_block_size >= 0)
  647. codecctl_int(avctx, AV1E_SET_DENOISE_BLOCK_SIZE, ctx->denoise_block_size);
  648. #endif
  649. #ifdef AOM_CTRL_AV1E_SET_ENABLE_GLOBAL_MOTION
  650. if (ctx->enable_global_motion >= 0)
  651. codecctl_int(avctx, AV1E_SET_ENABLE_GLOBAL_MOTION, ctx->enable_global_motion);
  652. #endif
  653. #ifdef AOM_CTRL_AV1E_SET_MAX_REFERENCE_FRAMES
  654. if (avctx->refs >= 3) {
  655. codecctl_int(avctx, AV1E_SET_MAX_REFERENCE_FRAMES, avctx->refs);
  656. }
  657. #endif
  658. #ifdef AOM_CTRL_AV1E_SET_ROW_MT
  659. if (ctx->row_mt >= 0)
  660. codecctl_int(avctx, AV1E_SET_ROW_MT, ctx->row_mt);
  661. #endif
  662. #ifdef AOM_CTRL_AV1E_SET_ENABLE_INTRABC
  663. if (ctx->enable_intrabc >= 0)
  664. codecctl_int(avctx, AV1E_SET_ENABLE_INTRABC, ctx->enable_intrabc);
  665. #endif
  666. // provide dummy value to initialize wrapper, values will be updated each _encode()
  667. aom_img_wrap(&ctx->rawimg, img_fmt, avctx->width, avctx->height, 1,
  668. (unsigned char*)1);
  669. if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH)
  670. ctx->rawimg.bit_depth = enccfg.g_bit_depth;
  671. cpb_props = ff_add_cpb_side_data(avctx);
  672. if (!cpb_props)
  673. return AVERROR(ENOMEM);
  674. if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
  675. const AVBitStreamFilter *filter = av_bsf_get_by_name("extract_extradata");
  676. int ret;
  677. if (!filter) {
  678. av_log(avctx, AV_LOG_ERROR, "extract_extradata bitstream filter "
  679. "not found. This is a bug, please report it.\n");
  680. return AVERROR_BUG;
  681. }
  682. ret = av_bsf_alloc(filter, &ctx->bsf);
  683. if (ret < 0)
  684. return ret;
  685. ret = avcodec_parameters_from_context(ctx->bsf->par_in, avctx);
  686. if (ret < 0)
  687. return ret;
  688. ret = av_bsf_init(ctx->bsf);
  689. if (ret < 0)
  690. return ret;
  691. }
  692. if (enccfg.rc_end_usage == AOM_CBR ||
  693. enccfg.g_pass != AOM_RC_ONE_PASS) {
  694. cpb_props->max_bitrate = avctx->rc_max_rate;
  695. cpb_props->min_bitrate = avctx->rc_min_rate;
  696. cpb_props->avg_bitrate = avctx->bit_rate;
  697. }
  698. cpb_props->buffer_size = avctx->rc_buffer_size;
  699. return 0;
  700. }
  701. static inline void cx_pktcpy(AOMContext *ctx,
  702. struct FrameListData *dst,
  703. const struct aom_codec_cx_pkt *src)
  704. {
  705. dst->pts = src->data.frame.pts;
  706. dst->duration = src->data.frame.duration;
  707. dst->flags = src->data.frame.flags;
  708. dst->sz = src->data.frame.sz;
  709. dst->buf = src->data.frame.buf;
  710. #ifdef AOM_FRAME_IS_INTRAONLY
  711. dst->have_sse = 0;
  712. dst->frame_number = ++ctx->frame_number;
  713. dst->have_sse = ctx->have_sse;
  714. if (ctx->have_sse) {
  715. /* associate last-seen SSE to the frame. */
  716. /* Transfers ownership from ctx to dst. */
  717. memcpy(dst->sse, ctx->sse, sizeof(dst->sse));
  718. ctx->have_sse = 0;
  719. }
  720. #endif
  721. }
  722. /**
  723. * Store coded frame information in format suitable for return from encode2().
  724. *
  725. * Write information from @a cx_frame to @a pkt
  726. * @return packet data size on success
  727. * @return a negative AVERROR on error
  728. */
  729. static int storeframe(AVCodecContext *avctx, struct FrameListData *cx_frame,
  730. AVPacket *pkt)
  731. {
  732. AOMContext *ctx = avctx->priv_data;
  733. int av_unused pict_type;
  734. int ret = ff_alloc_packet2(avctx, pkt, cx_frame->sz, 0);
  735. if (ret < 0) {
  736. av_log(avctx, AV_LOG_ERROR,
  737. "Error getting output packet of size %"SIZE_SPECIFIER".\n", cx_frame->sz);
  738. return ret;
  739. }
  740. memcpy(pkt->data, cx_frame->buf, pkt->size);
  741. pkt->pts = pkt->dts = cx_frame->pts;
  742. if (!!(cx_frame->flags & AOM_FRAME_IS_KEY)) {
  743. pkt->flags |= AV_PKT_FLAG_KEY;
  744. #ifdef AOM_FRAME_IS_INTRAONLY
  745. pict_type = AV_PICTURE_TYPE_I;
  746. } else if (cx_frame->flags & AOM_FRAME_IS_INTRAONLY) {
  747. pict_type = AV_PICTURE_TYPE_I;
  748. } else {
  749. pict_type = AV_PICTURE_TYPE_P;
  750. }
  751. ff_side_data_set_encoder_stats(pkt, 0, cx_frame->sse + 1,
  752. cx_frame->have_sse ? 3 : 0, pict_type);
  753. if (cx_frame->have_sse) {
  754. int i;
  755. for (i = 0; i < 3; ++i) {
  756. avctx->error[i] += cx_frame->sse[i + 1];
  757. }
  758. cx_frame->have_sse = 0;
  759. #endif
  760. }
  761. if (avctx->flags & AV_CODEC_FLAG_GLOBAL_HEADER) {
  762. ret = av_bsf_send_packet(ctx->bsf, pkt);
  763. if (ret < 0) {
  764. av_log(avctx, AV_LOG_ERROR, "extract_extradata filter "
  765. "failed to send input packet\n");
  766. return ret;
  767. }
  768. ret = av_bsf_receive_packet(ctx->bsf, pkt);
  769. if (ret < 0) {
  770. av_log(avctx, AV_LOG_ERROR, "extract_extradata filter "
  771. "failed to receive output packet\n");
  772. return ret;
  773. }
  774. }
  775. return pkt->size;
  776. }
  777. /**
  778. * Queue multiple output frames from the encoder, returning the front-most.
  779. * In cases where aom_codec_get_cx_data() returns more than 1 frame append
  780. * the frame queue. Return the head frame if available.
  781. * @return Stored frame size
  782. * @return AVERROR(EINVAL) on output size error
  783. * @return AVERROR(ENOMEM) on coded frame queue data allocation error
  784. */
  785. static int queue_frames(AVCodecContext *avctx, AVPacket *pkt_out)
  786. {
  787. AOMContext *ctx = avctx->priv_data;
  788. const struct aom_codec_cx_pkt *pkt;
  789. const void *iter = NULL;
  790. int size = 0;
  791. if (ctx->coded_frame_list) {
  792. struct FrameListData *cx_frame = ctx->coded_frame_list;
  793. /* return the leading frame if we've already begun queueing */
  794. size = storeframe(avctx, cx_frame, pkt_out);
  795. if (size < 0)
  796. return size;
  797. ctx->coded_frame_list = cx_frame->next;
  798. free_coded_frame(cx_frame);
  799. }
  800. /* consume all available output from the encoder before returning. buffers
  801. * are only good through the next aom_codec call */
  802. while ((pkt = aom_codec_get_cx_data(&ctx->encoder, &iter))) {
  803. switch (pkt->kind) {
  804. case AOM_CODEC_CX_FRAME_PKT:
  805. if (!size) {
  806. struct FrameListData cx_frame;
  807. /* avoid storing the frame when the list is empty and we haven't yet
  808. * provided a frame for output */
  809. av_assert0(!ctx->coded_frame_list);
  810. cx_pktcpy(ctx, &cx_frame, pkt);
  811. size = storeframe(avctx, &cx_frame, pkt_out);
  812. if (size < 0)
  813. return size;
  814. } else {
  815. struct FrameListData *cx_frame =
  816. av_malloc(sizeof(struct FrameListData));
  817. if (!cx_frame) {
  818. av_log(avctx, AV_LOG_ERROR,
  819. "Frame queue element alloc failed\n");
  820. return AVERROR(ENOMEM);
  821. }
  822. cx_pktcpy(ctx, cx_frame, pkt);
  823. cx_frame->buf = av_malloc(cx_frame->sz);
  824. if (!cx_frame->buf) {
  825. av_log(avctx, AV_LOG_ERROR,
  826. "Data buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
  827. cx_frame->sz);
  828. av_freep(&cx_frame);
  829. return AVERROR(ENOMEM);
  830. }
  831. memcpy(cx_frame->buf, pkt->data.frame.buf, pkt->data.frame.sz);
  832. coded_frame_add(&ctx->coded_frame_list, cx_frame);
  833. }
  834. break;
  835. case AOM_CODEC_STATS_PKT:
  836. {
  837. struct aom_fixed_buf *stats = &ctx->twopass_stats;
  838. int err;
  839. if ((err = av_reallocp(&stats->buf,
  840. stats->sz +
  841. pkt->data.twopass_stats.sz)) < 0) {
  842. stats->sz = 0;
  843. av_log(avctx, AV_LOG_ERROR, "Stat buffer realloc failed\n");
  844. return err;
  845. }
  846. memcpy((uint8_t *)stats->buf + stats->sz,
  847. pkt->data.twopass_stats.buf, pkt->data.twopass_stats.sz);
  848. stats->sz += pkt->data.twopass_stats.sz;
  849. break;
  850. }
  851. #ifdef AOM_FRAME_IS_INTRAONLY
  852. case AOM_CODEC_PSNR_PKT:
  853. {
  854. av_assert0(!ctx->have_sse);
  855. ctx->sse[0] = pkt->data.psnr.sse[0];
  856. ctx->sse[1] = pkt->data.psnr.sse[1];
  857. ctx->sse[2] = pkt->data.psnr.sse[2];
  858. ctx->sse[3] = pkt->data.psnr.sse[3];
  859. ctx->have_sse = 1;
  860. break;
  861. }
  862. #endif
  863. case AOM_CODEC_CUSTOM_PKT:
  864. // ignore unsupported/unrecognized packet types
  865. break;
  866. }
  867. }
  868. return size;
  869. }
  870. static int aom_encode(AVCodecContext *avctx, AVPacket *pkt,
  871. const AVFrame *frame, int *got_packet)
  872. {
  873. AOMContext *ctx = avctx->priv_data;
  874. struct aom_image *rawimg = NULL;
  875. int64_t timestamp = 0;
  876. int res, coded_size;
  877. aom_enc_frame_flags_t flags = 0;
  878. if (frame) {
  879. rawimg = &ctx->rawimg;
  880. rawimg->planes[AOM_PLANE_Y] = frame->data[0];
  881. rawimg->planes[AOM_PLANE_U] = frame->data[1];
  882. rawimg->planes[AOM_PLANE_V] = frame->data[2];
  883. rawimg->stride[AOM_PLANE_Y] = frame->linesize[0];
  884. rawimg->stride[AOM_PLANE_U] = frame->linesize[1];
  885. rawimg->stride[AOM_PLANE_V] = frame->linesize[2];
  886. timestamp = frame->pts;
  887. switch (frame->color_range) {
  888. case AVCOL_RANGE_MPEG:
  889. rawimg->range = AOM_CR_STUDIO_RANGE;
  890. break;
  891. case AVCOL_RANGE_JPEG:
  892. rawimg->range = AOM_CR_FULL_RANGE;
  893. break;
  894. }
  895. if (frame->pict_type == AV_PICTURE_TYPE_I)
  896. flags |= AOM_EFLAG_FORCE_KF;
  897. }
  898. res = aom_codec_encode(&ctx->encoder, rawimg, timestamp,
  899. avctx->ticks_per_frame, flags);
  900. if (res != AOM_CODEC_OK) {
  901. log_encoder_error(avctx, "Error encoding frame");
  902. return AVERROR_INVALIDDATA;
  903. }
  904. coded_size = queue_frames(avctx, pkt);
  905. if (!frame && avctx->flags & AV_CODEC_FLAG_PASS1) {
  906. size_t b64_size = AV_BASE64_SIZE(ctx->twopass_stats.sz);
  907. avctx->stats_out = av_malloc(b64_size);
  908. if (!avctx->stats_out) {
  909. av_log(avctx, AV_LOG_ERROR, "Stat buffer alloc (%"SIZE_SPECIFIER" bytes) failed\n",
  910. b64_size);
  911. return AVERROR(ENOMEM);
  912. }
  913. av_base64_encode(avctx->stats_out, b64_size, ctx->twopass_stats.buf,
  914. ctx->twopass_stats.sz);
  915. }
  916. *got_packet = !!coded_size;
  917. return 0;
  918. }
  919. static const enum AVPixelFormat av1_pix_fmts[] = {
  920. AV_PIX_FMT_YUV420P,
  921. AV_PIX_FMT_YUV422P,
  922. AV_PIX_FMT_YUV444P,
  923. AV_PIX_FMT_NONE
  924. };
  925. static const enum AVPixelFormat av1_pix_fmts_highbd[] = {
  926. AV_PIX_FMT_YUV420P,
  927. AV_PIX_FMT_YUV422P,
  928. AV_PIX_FMT_YUV444P,
  929. AV_PIX_FMT_YUV420P10,
  930. AV_PIX_FMT_YUV422P10,
  931. AV_PIX_FMT_YUV444P10,
  932. AV_PIX_FMT_YUV420P12,
  933. AV_PIX_FMT_YUV422P12,
  934. AV_PIX_FMT_YUV444P12,
  935. AV_PIX_FMT_NONE
  936. };
  937. static av_cold void av1_init_static(AVCodec *codec)
  938. {
  939. aom_codec_caps_t codec_caps = aom_codec_get_caps(aom_codec_av1_cx());
  940. if (codec_caps & AOM_CODEC_CAP_HIGHBITDEPTH)
  941. codec->pix_fmts = av1_pix_fmts_highbd;
  942. else
  943. codec->pix_fmts = av1_pix_fmts;
  944. }
  945. static av_cold int av1_init(AVCodecContext *avctx)
  946. {
  947. return aom_init(avctx, aom_codec_av1_cx());
  948. }
  949. #define OFFSET(x) offsetof(AOMContext, x)
  950. #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
  951. static const AVOption options[] = {
  952. { "cpu-used", "Quality/Speed ratio modifier", OFFSET(cpu_used), AV_OPT_TYPE_INT, {.i64 = 1}, 0, 8, VE},
  953. { "auto-alt-ref", "Enable use of alternate reference "
  954. "frames (2-pass only)", OFFSET(auto_alt_ref), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 2, VE},
  955. { "lag-in-frames", "Number of frames to look ahead at for "
  956. "alternate reference frame selection", OFFSET(lag_in_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
  957. { "arnr-max-frames", "altref noise reduction max frame count", OFFSET(arnr_max_frames), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
  958. { "arnr-strength", "altref noise reduction filter strength", OFFSET(arnr_strength), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
  959. { "aq-mode", "adaptive quantization mode", OFFSET(aq_mode), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 4, VE, "aq_mode"},
  960. { "none", "Aq not used", 0, AV_OPT_TYPE_CONST, {.i64 = 0}, 0, 0, VE, "aq_mode"},
  961. { "variance", "Variance based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 1}, 0, 0, VE, "aq_mode"},
  962. { "complexity", "Complexity based Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 2}, 0, 0, VE, "aq_mode"},
  963. { "cyclic", "Cyclic Refresh Aq", 0, AV_OPT_TYPE_CONST, {.i64 = 3}, 0, 0, VE, "aq_mode"},
  964. { "error-resilience", "Error resilience configuration", OFFSET(error_resilient), AV_OPT_TYPE_FLAGS, {.i64 = 0}, INT_MIN, INT_MAX, VE, "er"},
  965. { "default", "Improve resiliency against losses of whole frames", 0, AV_OPT_TYPE_CONST, {.i64 = AOM_ERROR_RESILIENT_DEFAULT}, 0, 0, VE, "er"},
  966. { "crf", "Select the quality for constant quality mode", offsetof(AOMContext, crf), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 63, VE },
  967. { "static-thresh", "A change threshold on blocks below which they will be skipped by the encoder", OFFSET(static_thresh), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, VE },
  968. { "drop-threshold", "Frame drop threshold", offsetof(AOMContext, drop_threshold), AV_OPT_TYPE_INT, {.i64 = 0 }, INT_MIN, INT_MAX, VE },
  969. { "denoise-noise-level", "Amount of noise to be removed", OFFSET(denoise_noise_level), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
  970. { "denoise-block-size", "Denoise block size ", OFFSET(denoise_block_size), AV_OPT_TYPE_INT, {.i64 = -1}, -1, INT_MAX, VE},
  971. { "undershoot-pct", "Datarate undershoot (min) target (%)", OFFSET(rc_undershoot_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 100, VE},
  972. { "overshoot-pct", "Datarate overshoot (max) target (%)", OFFSET(rc_overshoot_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 1000, VE},
  973. { "minsection-pct", "GOP min bitrate (% of target)", OFFSET(minsection_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 100, VE},
  974. { "maxsection-pct", "GOP max bitrate (% of target)", OFFSET(maxsection_pct), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 5000, VE},
  975. { "frame-parallel", "Enable frame parallel decodability features", OFFSET(frame_parallel), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
  976. { "tiles", "Tile columns x rows", OFFSET(tile_cols), AV_OPT_TYPE_IMAGE_SIZE, { .str = NULL }, 0, 0, VE },
  977. { "tile-columns", "Log2 of number of tile columns to use", OFFSET(tile_cols_log2), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
  978. { "tile-rows", "Log2 of number of tile rows to use", OFFSET(tile_rows_log2), AV_OPT_TYPE_INT, {.i64 = -1}, -1, 6, VE},
  979. { "row-mt", "Enable row based multi-threading", OFFSET(row_mt), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
  980. { "enable-cdef", "Enable CDEF filtering", OFFSET(enable_cdef), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
  981. { "enable-global-motion", "Enable global motion", OFFSET(enable_global_motion), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
  982. { "enable-intrabc", "Enable intra block copy prediction mode", OFFSET(enable_intrabc), AV_OPT_TYPE_BOOL, {.i64 = -1}, -1, 1, VE},
  983. { NULL },
  984. };
  985. static const AVCodecDefault defaults[] = {
  986. { "b", "0" },
  987. { "qmin", "-1" },
  988. { "qmax", "-1" },
  989. { "g", "-1" },
  990. { "keyint_min", "-1" },
  991. { NULL },
  992. };
  993. static const AVClass class_aom = {
  994. .class_name = "libaom-av1 encoder",
  995. .item_name = av_default_item_name,
  996. .option = options,
  997. .version = LIBAVUTIL_VERSION_INT,
  998. };
  999. AVCodec ff_libaom_av1_encoder = {
  1000. .name = "libaom-av1",
  1001. .long_name = NULL_IF_CONFIG_SMALL("libaom AV1"),
  1002. .type = AVMEDIA_TYPE_VIDEO,
  1003. .id = AV_CODEC_ID_AV1,
  1004. .priv_data_size = sizeof(AOMContext),
  1005. .init = av1_init,
  1006. .encode2 = aom_encode,
  1007. .close = aom_free,
  1008. .capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_AUTO_THREADS | AV_CODEC_CAP_EXPERIMENTAL,
  1009. .profiles = NULL_IF_CONFIG_SMALL(ff_av1_profiles),
  1010. .priv_class = &class_aom,
  1011. .defaults = defaults,
  1012. .init_static_data = av1_init_static,
  1013. .wrapper_name = "libaom",
  1014. };