You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

294 lines
10.0KB

  1. /*
  2. * AVS2 encoding using the xavs2 library
  3. *
  4. * Copyright (C) 2018 Yiqun Xu, <yiqun.xu@vipl.ict.ac.cn>
  5. * Falei Luo, <falei.luo@gmail.com>
  6. * Huiwen Ren, <hwrenx@gmail.com>
  7. *
  8. * This file is part of FFmpeg.
  9. *
  10. * FFmpeg is free software; you can redistribute it and/or
  11. * modify it under the terms of the GNU Lesser General Public
  12. * License as published by the Free Software Foundation; either
  13. * version 2.1 of the License, or (at your option) any later version.
  14. *
  15. * FFmpeg is distributed in the hope that it will be useful,
  16. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  17. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  18. * Lesser General Public License for more details.
  19. *
  20. * You should have received a copy of the GNU Lesser General Public
  21. * License along with FFmpeg; if not, write to the Free Software
  22. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  23. */
  24. #include "xavs2.h"
  25. #include "mpeg12.h"
  26. #include "libavutil/avstring.h"
  27. #define xavs2_opt_set2(name, format, ...) do{ \
  28. char opt_str[16] = {0}; \
  29. int err; \
  30. av_strlcatf(opt_str, sizeof(opt_str), format, __VA_ARGS__); \
  31. err = cae->api->opt_set2(cae->param, name, opt_str); \
  32. if (err < 0) {\
  33. av_log(avctx, AV_LOG_WARNING, "Invalid value for %s: %s\n", name, opt_str);\
  34. }\
  35. } while(0);
  36. typedef struct XAVS2EContext {
  37. AVClass *class;
  38. int lcu_row_threads;
  39. int initial_qp;
  40. int qp;
  41. int max_qp;
  42. int min_qp;
  43. int preset_level;
  44. int log_level;
  45. void *encoder;
  46. char *xavs2_opts;
  47. xavs2_outpacket_t packet;
  48. xavs2_param_t *param;
  49. const xavs2_api_t *api;
  50. } XAVS2EContext;
  51. static av_cold int xavs2_init(AVCodecContext *avctx)
  52. {
  53. XAVS2EContext *cae= avctx->priv_data;
  54. int bit_depth, code;
  55. bit_depth = avctx->pix_fmt == AV_PIX_FMT_YUV420P ? 8 : 10;
  56. /* get API handler */
  57. cae->api = xavs2_api_get(bit_depth);
  58. if (!cae->api) {
  59. av_log(avctx, AV_LOG_ERROR, "api get failed\n");
  60. return AVERROR_EXTERNAL;
  61. }
  62. cae->param = cae->api->opt_alloc();
  63. if (!cae->param) {
  64. av_log(avctx, AV_LOG_ERROR, "param alloc failed\n");
  65. return AVERROR(ENOMEM);
  66. }
  67. xavs2_opt_set2("Width", "%d", avctx->width);
  68. xavs2_opt_set2("Height", "%d", avctx->height);
  69. xavs2_opt_set2("BFrames", "%d", avctx->max_b_frames);
  70. xavs2_opt_set2("BitDepth", "%d", bit_depth);
  71. xavs2_opt_set2("Log", "%d", cae->log_level);
  72. xavs2_opt_set2("Preset", "%d", cae->preset_level);
  73. xavs2_opt_set2("IntraPeriodMax", "%d", avctx->gop_size);
  74. xavs2_opt_set2("IntraPeriodMin", "%d", avctx->gop_size);
  75. xavs2_opt_set2("ThreadFrames", "%d", avctx->thread_count);
  76. xavs2_opt_set2("ThreadRows", "%d", cae->lcu_row_threads);
  77. xavs2_opt_set2("OpenGOP", "%d", !(avctx->flags & AV_CODEC_FLAG_CLOSED_GOP));
  78. if (cae->xavs2_opts) {
  79. AVDictionary *dict = NULL;
  80. AVDictionaryEntry *en = NULL;
  81. if (!av_dict_parse_string(&dict, cae->xavs2_opts, "=", ":", 0)) {
  82. while ((en = av_dict_get(dict, "", en, AV_DICT_IGNORE_SUFFIX))) {
  83. xavs2_opt_set2(en->key, "%s", en->value);
  84. }
  85. av_dict_free(&dict);
  86. }
  87. }
  88. /* Rate control */
  89. if (avctx->bit_rate > 0) {
  90. xavs2_opt_set2("RateControl", "%d", 1);
  91. xavs2_opt_set2("TargetBitRate", "%"PRId64"", avctx->bit_rate);
  92. xavs2_opt_set2("InitialQP", "%d", cae->initial_qp);
  93. xavs2_opt_set2("MaxQP", "%d", avctx->qmax >= 0 ? avctx->qmax : cae->max_qp);
  94. xavs2_opt_set2("MinQP", "%d", avctx->qmin >= 0 ? avctx->qmin : cae->min_qp);
  95. } else {
  96. xavs2_opt_set2("InitialQP", "%d", cae->qp);
  97. }
  98. ff_mpeg12_find_best_frame_rate(avctx->framerate, &code, NULL, NULL, 0);
  99. xavs2_opt_set2("FrameRate", "%d", code);
  100. cae->encoder = cae->api->encoder_create(cae->param);
  101. if (!cae->encoder) {
  102. av_log(avctx,AV_LOG_ERROR, "Can not create encoder. Null pointer returned\n");
  103. return AVERROR(EINVAL);
  104. }
  105. return 0;
  106. }
  107. static void xavs2_copy_frame_with_shift(xavs2_picture_t *pic, const AVFrame *frame, const int shift_in)
  108. {
  109. int j, k;
  110. for (k = 0; k < 3; k++) {
  111. int i_stride = pic->img.i_stride[k];
  112. for (j = 0; j < pic->img.i_lines[k]; j++) {
  113. uint16_t *p_plane = (uint16_t *)&pic->img.img_planes[k][j * i_stride];
  114. int i;
  115. uint8_t *p_buffer = frame->data[k] + frame->linesize[k] * j;
  116. memset(p_plane, 0, i_stride);
  117. for (i = 0; i < pic->img.i_width[k]; i++) {
  118. p_plane[i] = p_buffer[i] << shift_in;
  119. }
  120. }
  121. }
  122. }
  123. static void xavs2_copy_frame(xavs2_picture_t *pic, const AVFrame *frame)
  124. {
  125. int j, k;
  126. for (k = 0; k < 3; k++) {
  127. for (j = 0; j < pic->img.i_lines[k]; j++) {
  128. memcpy( pic->img.img_planes[k] + pic->img.i_stride[k] * j,
  129. frame->data[k]+frame->linesize[k] * j,
  130. pic->img.i_width[k] * pic->img.in_sample_size);
  131. }
  132. }
  133. }
  134. static int xavs2_encode_frame(AVCodecContext *avctx, AVPacket *pkt,
  135. const AVFrame *frame, int *got_packet)
  136. {
  137. XAVS2EContext *cae = avctx->priv_data;
  138. xavs2_picture_t pic;
  139. int ret;
  140. /* create the XAVS2 video encoder */
  141. /* read frame data and send to the XAVS2 video encoder */
  142. if (cae->api->encoder_get_buffer(cae->encoder, &pic) < 0) {
  143. av_log(avctx,AV_LOG_ERROR, "failed to get frame buffer\n");
  144. return AVERROR_EXTERNAL;
  145. }
  146. if (frame) {
  147. switch (frame->format) {
  148. case AV_PIX_FMT_YUV420P:
  149. if (pic.img.in_sample_size == pic.img.enc_sample_size) {
  150. xavs2_copy_frame(&pic, frame);
  151. } else {
  152. const int shift_in = atoi(cae->api->opt_get(cae->param, "SampleShift"));
  153. xavs2_copy_frame_with_shift(&pic, frame, shift_in);
  154. }
  155. break;
  156. case AV_PIX_FMT_YUV420P10:
  157. if (pic.img.in_sample_size == pic.img.enc_sample_size) {
  158. xavs2_copy_frame(&pic, frame);
  159. break;
  160. }
  161. default:
  162. av_log(avctx, AV_LOG_ERROR, "Unsupported pixel format\n");
  163. return AVERROR(EINVAL);
  164. break;
  165. }
  166. pic.i_state = 0;
  167. pic.i_pts = frame->pts;
  168. pic.i_type = XAVS2_TYPE_AUTO;
  169. ret = cae->api->encoder_encode(cae->encoder, &pic, &cae->packet);
  170. if (ret) {
  171. av_log(avctx, AV_LOG_ERROR, "encode failed\n");
  172. return AVERROR_EXTERNAL;
  173. }
  174. } else {
  175. cae->api->encoder_encode(cae->encoder, NULL, &cae->packet);
  176. }
  177. if ((cae->packet.len) && (cae->packet.state != XAVS2_STATE_FLUSH_END)){
  178. if (av_new_packet(pkt, cae->packet.len) < 0){
  179. av_log(avctx, AV_LOG_ERROR, "packet alloc failed\n");
  180. cae->api->encoder_packet_unref(cae->encoder, &cae->packet);
  181. return AVERROR(ENOMEM);
  182. }
  183. pkt->pts = cae->packet.pts;
  184. pkt->dts = cae->packet.dts;
  185. memcpy(pkt->data, cae->packet.stream, cae->packet.len);
  186. pkt->size = cae->packet.len;
  187. cae->api->encoder_packet_unref(cae->encoder, &cae->packet);
  188. *got_packet = 1;
  189. } else {
  190. *got_packet = 0;
  191. }
  192. return 0;
  193. }
  194. static av_cold int xavs2_close(AVCodecContext *avctx)
  195. {
  196. XAVS2EContext *cae = avctx->priv_data;
  197. /* destroy the encoder */
  198. if (cae->api) {
  199. cae->api->encoder_destroy(cae->encoder);
  200. if (cae->param) {
  201. cae->api->opt_destroy(cae->param);
  202. }
  203. }
  204. return 0;
  205. }
  206. #define OFFSET(x) offsetof(XAVS2EContext, x)
  207. #define VE AV_OPT_FLAG_VIDEO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
  208. static const AVOption options[] = {
  209. { "lcu_row_threads" , "number of parallel threads for rows" , OFFSET(lcu_row_threads) , AV_OPT_TYPE_INT, {.i64 = 0 }, 0, INT_MAX, VE },
  210. { "initial_qp" , "Quantization initial parameter" , OFFSET(initial_qp) , AV_OPT_TYPE_INT, {.i64 = 34 }, 1, 63, VE },
  211. { "qp" , "Quantization parameter" , OFFSET(qp) , AV_OPT_TYPE_INT, {.i64 = 34 }, 1, 63, VE },
  212. { "max_qp" , "max qp for rate control" , OFFSET(max_qp) , AV_OPT_TYPE_INT, {.i64 = 55 }, 0, 63, VE },
  213. { "min_qp" , "min qp for rate control" , OFFSET(min_qp) , AV_OPT_TYPE_INT, {.i64 = 20 }, 0, 63, VE },
  214. { "speed_level" , "Speed level, higher is better but slower", OFFSET(preset_level) , AV_OPT_TYPE_INT, {.i64 = 0 }, 0, 9, VE },
  215. { "log_level" , "log level: -1: none, 0: error, 1: warning, 2: info, 3: debug", OFFSET(log_level) , AV_OPT_TYPE_INT, {.i64 = 0 }, -1, 3, VE },
  216. { "xavs2-params" , "set the xavs2 configuration using a :-separated list of key=value parameters", OFFSET(xavs2_opts), AV_OPT_TYPE_STRING, { 0 }, 0, 0, VE },
  217. { NULL },
  218. };
  219. static const AVClass libxavs2 = {
  220. .class_name = "XAVS2EContext",
  221. .item_name = av_default_item_name,
  222. .option = options,
  223. .version = LIBAVUTIL_VERSION_INT,
  224. };
  225. static const AVCodecDefault xavs2_defaults[] = {
  226. { "b", "0" },
  227. { "g", "48"},
  228. { "bf", "7" },
  229. { NULL },
  230. };
  231. AVCodec ff_libxavs2_encoder = {
  232. .name = "libxavs2",
  233. .long_name = NULL_IF_CONFIG_SMALL("libxavs2 AVS2-P2/IEEE1857.4"),
  234. .type = AVMEDIA_TYPE_VIDEO,
  235. .id = AV_CODEC_ID_AVS2,
  236. .priv_data_size = sizeof(XAVS2EContext),
  237. .init = xavs2_init,
  238. .encode2 = xavs2_encode_frame,
  239. .close = xavs2_close,
  240. .capabilities = AV_CODEC_CAP_DELAY | AV_CODEC_CAP_AUTO_THREADS,
  241. .pix_fmts = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUV420P,
  242. AV_PIX_FMT_NONE },
  243. .priv_class = &libxavs2,
  244. .defaults = xavs2_defaults,
  245. .wrapper_name = "libxavs2",
  246. } ;