You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

1116 lines
37KB

  1. /*
  2. * Intel MediaSDK QSV encoder utility functions
  3. *
  4. * copyright (c) 2013 Yukinori Yamazoe
  5. * copyright (c) 2015 Anton Khirnov
  6. *
  7. * This file is part of Libav.
  8. *
  9. * Libav is free software; you can redistribute it and/or
  10. * modify it under the terms of the GNU Lesser General Public
  11. * License as published by the Free Software Foundation; either
  12. * version 2.1 of the License, or (at your option) any later version.
  13. *
  14. * Libav is distributed in the hope that it will be useful,
  15. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  16. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  17. * Lesser General Public License for more details.
  18. *
  19. * You should have received a copy of the GNU Lesser General Public
  20. * License along with Libav; if not, write to the Free Software
  21. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  22. */
  23. #include <string.h>
  24. #include <sys/types.h>
  25. #include <mfx/mfxvideo.h>
  26. #include "libavutil/common.h"
  27. #include "libavutil/hwcontext.h"
  28. #include "libavutil/hwcontext_qsv.h"
  29. #include "libavutil/mem.h"
  30. #include "libavutil/log.h"
  31. #include "libavutil/time.h"
  32. #include "libavutil/imgutils.h"
  33. #include "avcodec.h"
  34. #include "internal.h"
  35. #include "qsv.h"
  36. #include "qsv_internal.h"
  37. #include "qsvenc.h"
  38. static const struct {
  39. mfxU16 profile;
  40. const char *name;
  41. } profile_names[] = {
  42. { MFX_PROFILE_AVC_BASELINE, "baseline" },
  43. { MFX_PROFILE_AVC_MAIN, "main" },
  44. { MFX_PROFILE_AVC_EXTENDED, "extended" },
  45. { MFX_PROFILE_AVC_HIGH, "high" },
  46. #if QSV_VERSION_ATLEAST(1, 15)
  47. { MFX_PROFILE_AVC_HIGH_422, "high 422" },
  48. #endif
  49. #if QSV_VERSION_ATLEAST(1, 4)
  50. { MFX_PROFILE_AVC_CONSTRAINED_BASELINE, "constrained baseline" },
  51. { MFX_PROFILE_AVC_CONSTRAINED_HIGH, "constrained high" },
  52. { MFX_PROFILE_AVC_PROGRESSIVE_HIGH, "progressive high" },
  53. #endif
  54. { MFX_PROFILE_MPEG2_SIMPLE, "simple" },
  55. { MFX_PROFILE_MPEG2_MAIN, "main" },
  56. { MFX_PROFILE_MPEG2_HIGH, "high" },
  57. { MFX_PROFILE_VC1_SIMPLE, "simple" },
  58. { MFX_PROFILE_VC1_MAIN, "main" },
  59. { MFX_PROFILE_VC1_ADVANCED, "advanced" },
  60. #if QSV_VERSION_ATLEAST(1, 8)
  61. { MFX_PROFILE_HEVC_MAIN, "main" },
  62. { MFX_PROFILE_HEVC_MAIN10, "main10" },
  63. { MFX_PROFILE_HEVC_MAINSP, "mainsp" },
  64. #endif
  65. };
  66. static const char *print_profile(mfxU16 profile)
  67. {
  68. int i;
  69. for (i = 0; i < FF_ARRAY_ELEMS(profile_names); i++)
  70. if (profile == profile_names[i].profile)
  71. return profile_names[i].name;
  72. return "unknown";
  73. }
  74. static const struct {
  75. mfxU16 rc_mode;
  76. const char *name;
  77. } rc_names[] = {
  78. { MFX_RATECONTROL_CBR, "CBR" },
  79. { MFX_RATECONTROL_VBR, "VBR" },
  80. { MFX_RATECONTROL_CQP, "CQP" },
  81. { MFX_RATECONTROL_AVBR, "AVBR" },
  82. #if QSV_HAVE_LA
  83. { MFX_RATECONTROL_LA, "LA" },
  84. #endif
  85. #if QSV_HAVE_ICQ
  86. { MFX_RATECONTROL_ICQ, "ICQ" },
  87. { MFX_RATECONTROL_LA_ICQ, "LA_ICQ" },
  88. #endif
  89. #if QSV_HAVE_VCM
  90. { MFX_RATECONTROL_VCM, "VCM" },
  91. #endif
  92. #if QSV_VERSION_ATLEAST(1, 10)
  93. { MFX_RATECONTROL_LA_EXT, "LA_EXT" },
  94. #endif
  95. #if QSV_HAVE_LA_HRD
  96. { MFX_RATECONTROL_LA_HRD, "LA_HRD" },
  97. #endif
  98. #if QSV_HAVE_QVBR
  99. { MFX_RATECONTROL_QVBR, "QVBR" },
  100. #endif
  101. };
  102. static const char *print_ratecontrol(mfxU16 rc_mode)
  103. {
  104. int i;
  105. for (i = 0; i < FF_ARRAY_ELEMS(rc_names); i++)
  106. if (rc_mode == rc_names[i].rc_mode)
  107. return rc_names[i].name;
  108. return "unknown";
  109. }
  110. static const char *print_threestate(mfxU16 val)
  111. {
  112. if (val == MFX_CODINGOPTION_ON)
  113. return "ON";
  114. else if (val == MFX_CODINGOPTION_OFF)
  115. return "OFF";
  116. return "unknown";
  117. }
  118. static void dump_video_param(AVCodecContext *avctx, QSVEncContext *q,
  119. mfxExtBuffer **coding_opts)
  120. {
  121. mfxInfoMFX *info = &q->param.mfx;
  122. mfxExtCodingOption *co = (mfxExtCodingOption*)coding_opts[0];
  123. #if QSV_HAVE_CO2
  124. mfxExtCodingOption2 *co2 = (mfxExtCodingOption2*)coding_opts[1];
  125. #endif
  126. #if QSV_HAVE_CO3
  127. mfxExtCodingOption3 *co3 = (mfxExtCodingOption3*)coding_opts[2];
  128. #endif
  129. av_log(avctx, AV_LOG_VERBOSE, "profile: %s; level: %"PRIu16"\n",
  130. print_profile(info->CodecProfile), info->CodecLevel);
  131. av_log(avctx, AV_LOG_VERBOSE, "GopPicSize: %"PRIu16"; GopRefDist: %"PRIu16"; GopOptFlag: ",
  132. info->GopPicSize, info->GopRefDist);
  133. if (info->GopOptFlag & MFX_GOP_CLOSED)
  134. av_log(avctx, AV_LOG_VERBOSE, "closed ");
  135. if (info->GopOptFlag & MFX_GOP_STRICT)
  136. av_log(avctx, AV_LOG_VERBOSE, "strict ");
  137. av_log(avctx, AV_LOG_VERBOSE, "; IdrInterval: %"PRIu16"\n", info->IdrInterval);
  138. av_log(avctx, AV_LOG_VERBOSE, "TargetUsage: %"PRIu16"; RateControlMethod: %s\n",
  139. info->TargetUsage, print_ratecontrol(info->RateControlMethod));
  140. if (info->RateControlMethod == MFX_RATECONTROL_CBR ||
  141. info->RateControlMethod == MFX_RATECONTROL_VBR
  142. #if QSV_HAVE_VCM
  143. || info->RateControlMethod == MFX_RATECONTROL_VCM
  144. #endif
  145. ) {
  146. av_log(avctx, AV_LOG_VERBOSE,
  147. "InitialDelayInKB: %"PRIu16"; TargetKbps: %"PRIu16"; MaxKbps: %"PRIu16"\n",
  148. info->InitialDelayInKB, info->TargetKbps, info->MaxKbps);
  149. } else if (info->RateControlMethod == MFX_RATECONTROL_CQP) {
  150. av_log(avctx, AV_LOG_VERBOSE, "QPI: %"PRIu16"; QPP: %"PRIu16"; QPB: %"PRIu16"\n",
  151. info->QPI, info->QPP, info->QPB);
  152. } else if (info->RateControlMethod == MFX_RATECONTROL_AVBR) {
  153. av_log(avctx, AV_LOG_VERBOSE,
  154. "TargetKbps: %"PRIu16"; Accuracy: %"PRIu16"; Convergence: %"PRIu16"\n",
  155. info->TargetKbps, info->Accuracy, info->Convergence);
  156. }
  157. #if QSV_HAVE_LA
  158. else if (info->RateControlMethod == MFX_RATECONTROL_LA
  159. #if QSV_HAVE_LA_HRD
  160. || info->RateControlMethod == MFX_RATECONTROL_LA_HRD
  161. #endif
  162. ) {
  163. av_log(avctx, AV_LOG_VERBOSE,
  164. "TargetKbps: %"PRIu16"; LookAheadDepth: %"PRIu16"\n",
  165. info->TargetKbps, co2->LookAheadDepth);
  166. }
  167. #endif
  168. #if QSV_HAVE_ICQ
  169. else if (info->RateControlMethod == MFX_RATECONTROL_ICQ) {
  170. av_log(avctx, AV_LOG_VERBOSE, "ICQQuality: %"PRIu16"\n", info->ICQQuality);
  171. } else if (info->RateControlMethod == MFX_RATECONTROL_LA_ICQ) {
  172. av_log(avctx, AV_LOG_VERBOSE, "ICQQuality: %"PRIu16"; LookAheadDepth: %"PRIu16"\n",
  173. info->ICQQuality, co2->LookAheadDepth);
  174. }
  175. #endif
  176. #if QSV_HAVE_QVBR
  177. else if (info->RateControlMethod == MFX_RATECONTROL_QVBR) {
  178. av_log(avctx, AV_LOG_VERBOSE, "QVBRQuality: %"PRIu16"\n",
  179. co3->QVBRQuality);
  180. }
  181. #endif
  182. av_log(avctx, AV_LOG_VERBOSE, "NumSlice: %"PRIu16"; NumRefFrame: %"PRIu16"\n",
  183. info->NumSlice, info->NumRefFrame);
  184. av_log(avctx, AV_LOG_VERBOSE, "RateDistortionOpt: %s\n",
  185. print_threestate(co->RateDistortionOpt));
  186. #if QSV_HAVE_CO2
  187. av_log(avctx, AV_LOG_VERBOSE,
  188. "RecoveryPointSEI: %s IntRefType: %"PRIu16"; IntRefCycleSize: %"PRIu16"; IntRefQPDelta: %"PRId16"\n",
  189. print_threestate(co->RecoveryPointSEI), co2->IntRefType, co2->IntRefCycleSize, co2->IntRefQPDelta);
  190. av_log(avctx, AV_LOG_VERBOSE, "MaxFrameSize: %"PRIu16"; ", co2->MaxFrameSize);
  191. #if QSV_HAVE_MAX_SLICE_SIZE
  192. av_log(avctx, AV_LOG_VERBOSE, "MaxSliceSize: %"PRIu16"; ", co2->MaxSliceSize);
  193. #endif
  194. av_log(avctx, AV_LOG_VERBOSE, "\n");
  195. av_log(avctx, AV_LOG_VERBOSE,
  196. "BitrateLimit: %s; MBBRC: %s; ExtBRC: %s\n",
  197. print_threestate(co2->BitrateLimit), print_threestate(co2->MBBRC),
  198. print_threestate(co2->ExtBRC));
  199. #if QSV_HAVE_TRELLIS
  200. av_log(avctx, AV_LOG_VERBOSE, "Trellis: ");
  201. if (co2->Trellis & MFX_TRELLIS_OFF) {
  202. av_log(avctx, AV_LOG_VERBOSE, "off");
  203. } else if (!co2->Trellis) {
  204. av_log(avctx, AV_LOG_VERBOSE, "auto");
  205. } else {
  206. if (co2->Trellis & MFX_TRELLIS_I) av_log(avctx, AV_LOG_VERBOSE, "I");
  207. if (co2->Trellis & MFX_TRELLIS_P) av_log(avctx, AV_LOG_VERBOSE, "P");
  208. if (co2->Trellis & MFX_TRELLIS_B) av_log(avctx, AV_LOG_VERBOSE, "B");
  209. }
  210. av_log(avctx, AV_LOG_VERBOSE, "\n");
  211. #endif
  212. #if QSV_VERSION_ATLEAST(1, 8)
  213. av_log(avctx, AV_LOG_VERBOSE,
  214. "RepeatPPS: %s; NumMbPerSlice: %"PRIu16"; LookAheadDS: ",
  215. print_threestate(co2->RepeatPPS), co2->NumMbPerSlice);
  216. switch (co2->LookAheadDS) {
  217. case MFX_LOOKAHEAD_DS_OFF: av_log(avctx, AV_LOG_VERBOSE, "off"); break;
  218. case MFX_LOOKAHEAD_DS_2x: av_log(avctx, AV_LOG_VERBOSE, "2x"); break;
  219. case MFX_LOOKAHEAD_DS_4x: av_log(avctx, AV_LOG_VERBOSE, "4x"); break;
  220. default: av_log(avctx, AV_LOG_VERBOSE, "unknown"); break;
  221. }
  222. av_log(avctx, AV_LOG_VERBOSE, "\n");
  223. av_log(avctx, AV_LOG_VERBOSE, "AdaptiveI: %s; AdaptiveB: %s; BRefType: ",
  224. print_threestate(co2->AdaptiveI), print_threestate(co2->AdaptiveB));
  225. switch (co2->BRefType) {
  226. case MFX_B_REF_OFF: av_log(avctx, AV_LOG_VERBOSE, "off"); break;
  227. case MFX_B_REF_PYRAMID: av_log(avctx, AV_LOG_VERBOSE, "pyramid"); break;
  228. default: av_log(avctx, AV_LOG_VERBOSE, "auto"); break;
  229. }
  230. av_log(avctx, AV_LOG_VERBOSE, "\n");
  231. #endif
  232. #if QSV_VERSION_ATLEAST(1, 9)
  233. av_log(avctx, AV_LOG_VERBOSE,
  234. "MinQPI: %"PRIu8"; MaxQPI: %"PRIu8"; MinQPP: %"PRIu8"; MaxQPP: %"PRIu8"; MinQPB: %"PRIu8"; MaxQPB: %"PRIu8"\n",
  235. co2->MinQPI, co2->MaxQPI, co2->MinQPP, co2->MaxQPP, co2->MinQPB, co2->MaxQPB);
  236. #endif
  237. #endif
  238. if (avctx->codec_id == AV_CODEC_ID_H264) {
  239. av_log(avctx, AV_LOG_VERBOSE, "Entropy coding: %s; MaxDecFrameBuffering: %"PRIu16"\n",
  240. co->CAVLC == MFX_CODINGOPTION_ON ? "CAVLC" : "CABAC", co->MaxDecFrameBuffering);
  241. av_log(avctx, AV_LOG_VERBOSE,
  242. "NalHrdConformance: %s; SingleSeiNalUnit: %s; VuiVclHrdParameters: %s VuiNalHrdParameters: %s\n",
  243. print_threestate(co->NalHrdConformance), print_threestate(co->SingleSeiNalUnit),
  244. print_threestate(co->VuiVclHrdParameters), print_threestate(co->VuiNalHrdParameters));
  245. }
  246. }
  247. static int select_rc_mode(AVCodecContext *avctx, QSVEncContext *q)
  248. {
  249. const char *rc_desc;
  250. mfxU16 rc_mode;
  251. int want_la = q->la_depth >= 0;
  252. int want_qscale = !!(avctx->flags & AV_CODEC_FLAG_QSCALE);
  253. int want_vcm = q->vcm;
  254. if (want_la && !QSV_HAVE_LA) {
  255. av_log(avctx, AV_LOG_ERROR,
  256. "Lookahead ratecontrol mode requested, but is not supported by this SDK version\n");
  257. return AVERROR(ENOSYS);
  258. }
  259. if (want_vcm && !QSV_HAVE_VCM) {
  260. av_log(avctx, AV_LOG_ERROR,
  261. "VCM ratecontrol mode requested, but is not supported by this SDK version\n");
  262. return AVERROR(ENOSYS);
  263. }
  264. if (want_la + want_qscale + want_vcm > 1) {
  265. av_log(avctx, AV_LOG_ERROR,
  266. "More than one of: { constant qscale, lookahead, VCM } requested, "
  267. "only one of them can be used at a time.\n");
  268. return AVERROR(EINVAL);
  269. }
  270. if (want_qscale) {
  271. rc_mode = MFX_RATECONTROL_CQP;
  272. rc_desc = "constant quantization parameter (CQP)";
  273. }
  274. #if QSV_HAVE_VCM
  275. else if (want_vcm) {
  276. rc_mode = MFX_RATECONTROL_VCM;
  277. rc_desc = "video conferencing mode (VCM)";
  278. }
  279. #endif
  280. #if QSV_HAVE_LA
  281. else if (want_la) {
  282. rc_mode = MFX_RATECONTROL_LA;
  283. rc_desc = "VBR with lookahead (LA)";
  284. #if QSV_HAVE_ICQ
  285. if (avctx->global_quality > 0) {
  286. rc_mode = MFX_RATECONTROL_LA_ICQ;
  287. rc_desc = "intelligent constant quality with lookahead (LA_ICQ)";
  288. }
  289. #endif
  290. }
  291. #endif
  292. #if QSV_HAVE_ICQ
  293. else if (avctx->global_quality > 0) {
  294. rc_mode = MFX_RATECONTROL_ICQ;
  295. rc_desc = "intelligent constant quality (ICQ)";
  296. }
  297. #endif
  298. else if (avctx->rc_max_rate == avctx->bit_rate) {
  299. rc_mode = MFX_RATECONTROL_CBR;
  300. rc_desc = "constant bitrate (CBR)";
  301. } else if (!avctx->rc_max_rate) {
  302. rc_mode = MFX_RATECONTROL_AVBR;
  303. rc_desc = "average variable bitrate (AVBR)";
  304. } else {
  305. rc_mode = MFX_RATECONTROL_VBR;
  306. rc_desc = "variable bitrate (VBR)";
  307. }
  308. q->param.mfx.RateControlMethod = rc_mode;
  309. av_log(avctx, AV_LOG_VERBOSE, "Using the %s ratecontrol method\n", rc_desc);
  310. return 0;
  311. }
  312. static int rc_supported(QSVEncContext *q)
  313. {
  314. mfxVideoParam param_out = { .mfx.CodecId = q->param.mfx.CodecId };
  315. mfxStatus ret;
  316. ret = MFXVideoENCODE_Query(q->session, &q->param, &param_out);
  317. if (ret < 0 ||
  318. param_out.mfx.RateControlMethod != q->param.mfx.RateControlMethod)
  319. return 0;
  320. return 1;
  321. }
  322. static int init_video_param(AVCodecContext *avctx, QSVEncContext *q)
  323. {
  324. float quant;
  325. int ret;
  326. ret = ff_qsv_codec_id_to_mfx(avctx->codec_id);
  327. if (ret < 0)
  328. return AVERROR_BUG;
  329. q->param.mfx.CodecId = ret;
  330. q->width_align = avctx->codec_id == AV_CODEC_ID_HEVC ? 32 : 16;
  331. if (avctx->level > 0)
  332. q->param.mfx.CodecLevel = avctx->level;
  333. q->param.mfx.CodecProfile = q->profile;
  334. q->param.mfx.TargetUsage = q->preset;
  335. q->param.mfx.GopPicSize = FFMAX(0, avctx->gop_size);
  336. q->param.mfx.GopRefDist = FFMAX(-1, avctx->max_b_frames) + 1;
  337. q->param.mfx.GopOptFlag = avctx->flags & AV_CODEC_FLAG_CLOSED_GOP ?
  338. MFX_GOP_CLOSED : 0;
  339. q->param.mfx.IdrInterval = q->idr_interval;
  340. q->param.mfx.NumSlice = avctx->slices;
  341. q->param.mfx.NumRefFrame = FFMAX(0, avctx->refs);
  342. q->param.mfx.EncodedOrder = 0;
  343. q->param.mfx.BufferSizeInKB = 0;
  344. if (avctx->hw_frames_ctx) {
  345. AVHWFramesContext *frames_ctx = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
  346. AVQSVFramesContext *frames_hwctx = frames_ctx->hwctx;
  347. q->param.mfx.FrameInfo = frames_hwctx->surfaces[0].Info;
  348. } else {
  349. enum AVPixelFormat sw_format = avctx->pix_fmt == AV_PIX_FMT_QSV ?
  350. avctx->sw_pix_fmt : avctx->pix_fmt;
  351. const AVPixFmtDescriptor *desc;
  352. desc = av_pix_fmt_desc_get(sw_format);
  353. if (!desc)
  354. return AVERROR_BUG;
  355. ff_qsv_map_pixfmt(sw_format, &q->param.mfx.FrameInfo.FourCC);
  356. q->param.mfx.FrameInfo.Width = FFALIGN(avctx->width, q->width_align);
  357. q->param.mfx.FrameInfo.Height = FFALIGN(avctx->height, 32);
  358. q->param.mfx.FrameInfo.CropX = 0;
  359. q->param.mfx.FrameInfo.CropY = 0;
  360. q->param.mfx.FrameInfo.CropW = avctx->width;
  361. q->param.mfx.FrameInfo.CropH = avctx->height;
  362. q->param.mfx.FrameInfo.AspectRatioW = avctx->sample_aspect_ratio.num;
  363. q->param.mfx.FrameInfo.AspectRatioH = avctx->sample_aspect_ratio.den;
  364. q->param.mfx.FrameInfo.PicStruct = MFX_PICSTRUCT_PROGRESSIVE;
  365. q->param.mfx.FrameInfo.ChromaFormat = MFX_CHROMAFORMAT_YUV420;
  366. q->param.mfx.FrameInfo.BitDepthLuma = desc->comp[0].depth;
  367. q->param.mfx.FrameInfo.BitDepthChroma = desc->comp[0].depth;
  368. q->param.mfx.FrameInfo.Shift = desc->comp[0].depth > 8;
  369. }
  370. if (avctx->framerate.den > 0 && avctx->framerate.num > 0) {
  371. q->param.mfx.FrameInfo.FrameRateExtN = avctx->framerate.num;
  372. q->param.mfx.FrameInfo.FrameRateExtD = avctx->framerate.den;
  373. } else {
  374. q->param.mfx.FrameInfo.FrameRateExtN = avctx->time_base.den;
  375. q->param.mfx.FrameInfo.FrameRateExtD = avctx->time_base.num;
  376. }
  377. ret = select_rc_mode(avctx, q);
  378. if (ret < 0)
  379. return ret;
  380. switch (q->param.mfx.RateControlMethod) {
  381. case MFX_RATECONTROL_CBR:
  382. case MFX_RATECONTROL_VBR:
  383. #if QSV_HAVE_VCM
  384. case MFX_RATECONTROL_VCM:
  385. #endif
  386. q->param.mfx.InitialDelayInKB = avctx->rc_initial_buffer_occupancy / 1000;
  387. q->param.mfx.TargetKbps = avctx->bit_rate / 1000;
  388. q->param.mfx.MaxKbps = avctx->rc_max_rate / 1000;
  389. break;
  390. case MFX_RATECONTROL_CQP:
  391. quant = avctx->global_quality / FF_QP2LAMBDA;
  392. q->param.mfx.QPI = av_clip(quant * fabs(avctx->i_quant_factor) + avctx->i_quant_offset, 0, 51);
  393. q->param.mfx.QPP = av_clip(quant, 0, 51);
  394. q->param.mfx.QPB = av_clip(quant * fabs(avctx->b_quant_factor) + avctx->b_quant_offset, 0, 51);
  395. break;
  396. case MFX_RATECONTROL_AVBR:
  397. q->param.mfx.TargetKbps = avctx->bit_rate / 1000;
  398. q->param.mfx.Convergence = q->avbr_convergence;
  399. q->param.mfx.Accuracy = q->avbr_accuracy;
  400. break;
  401. #if QSV_HAVE_LA
  402. case MFX_RATECONTROL_LA:
  403. q->param.mfx.TargetKbps = avctx->bit_rate / 1000;
  404. q->extco2.LookAheadDepth = q->la_depth;
  405. break;
  406. #if QSV_HAVE_ICQ
  407. case MFX_RATECONTROL_LA_ICQ:
  408. q->extco2.LookAheadDepth = q->la_depth;
  409. case MFX_RATECONTROL_ICQ:
  410. q->param.mfx.ICQQuality = avctx->global_quality;
  411. break;
  412. #endif
  413. #endif
  414. }
  415. // the HEVC encoder plugin currently fails if coding options
  416. // are provided
  417. if (avctx->codec_id != AV_CODEC_ID_HEVC) {
  418. q->extco.Header.BufferId = MFX_EXTBUFF_CODING_OPTION;
  419. q->extco.Header.BufferSz = sizeof(q->extco);
  420. #if FF_API_CODER_TYPE
  421. FF_DISABLE_DEPRECATION_WARNINGS
  422. if (avctx->coder_type != 0)
  423. q->cavlc = avctx->coder_type == FF_CODER_TYPE_VLC;
  424. FF_ENABLE_DEPRECATION_WARNINGS
  425. #endif
  426. q->extco.CAVLC = q->cavlc ? MFX_CODINGOPTION_ON
  427. : MFX_CODINGOPTION_UNKNOWN;
  428. if (q->rdo >= 0)
  429. q->extco.RateDistortionOpt = q->rdo > 0 ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  430. if (avctx->codec_id == AV_CODEC_ID_H264) {
  431. if (avctx->strict_std_compliance != FF_COMPLIANCE_NORMAL)
  432. q->extco.NalHrdConformance = avctx->strict_std_compliance > FF_COMPLIANCE_NORMAL ?
  433. MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  434. if (q->single_sei_nal_unit >= 0)
  435. q->extco.SingleSeiNalUnit = q->single_sei_nal_unit ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  436. if (q->recovery_point_sei >= 0)
  437. q->extco.RecoveryPointSEI = q->recovery_point_sei ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  438. q->extco.MaxDecFrameBuffering = q->max_dec_frame_buffering;
  439. }
  440. q->extparam_internal[q->nb_extparam_internal++] = (mfxExtBuffer *)&q->extco;
  441. #if QSV_HAVE_CO2
  442. if (avctx->codec_id == AV_CODEC_ID_H264) {
  443. q->extco2.Header.BufferId = MFX_EXTBUFF_CODING_OPTION2;
  444. q->extco2.Header.BufferSz = sizeof(q->extco2);
  445. if (q->int_ref_type >= 0)
  446. q->extco2.IntRefType = q->int_ref_type;
  447. if (q->int_ref_cycle_size >= 0)
  448. q->extco2.IntRefCycleSize = q->int_ref_cycle_size;
  449. if (q->int_ref_qp_delta != INT16_MIN)
  450. q->extco2.IntRefQPDelta = q->int_ref_qp_delta;
  451. if (q->bitrate_limit >= 0)
  452. q->extco2.BitrateLimit = q->bitrate_limit ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  453. if (q->mbbrc >= 0)
  454. q->extco2.MBBRC = q->mbbrc ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  455. if (q->extbrc >= 0)
  456. q->extco2.ExtBRC = q->extbrc ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  457. if (q->max_frame_size >= 0)
  458. q->extco2.MaxFrameSize = q->max_frame_size;
  459. #if QSV_HAVE_MAX_SLICE_SIZE
  460. if (q->max_slice_size >= 0)
  461. q->extco2.MaxSliceSize = q->max_slice_size;
  462. #endif
  463. #if QSV_HAVE_TRELLIS
  464. q->extco2.Trellis = q->trellis;
  465. #endif
  466. #if QSV_HAVE_BREF_TYPE
  467. #if FF_API_PRIVATE_OPT
  468. FF_DISABLE_DEPRECATION_WARNINGS
  469. if (avctx->b_frame_strategy >= 0)
  470. q->b_strategy = avctx->b_frame_strategy;
  471. FF_ENABLE_DEPRECATION_WARNINGS
  472. #endif
  473. if (q->b_strategy >= 0)
  474. q->extco2.BRefType = q->b_strategy ? MFX_B_REF_PYRAMID : MFX_B_REF_OFF;
  475. if (q->adaptive_i >= 0)
  476. q->extco2.AdaptiveI = q->adaptive_i ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  477. if (q->adaptive_b >= 0)
  478. q->extco2.AdaptiveB = q->adaptive_b ? MFX_CODINGOPTION_ON : MFX_CODINGOPTION_OFF;
  479. #endif
  480. q->extparam_internal[q->nb_extparam_internal++] = (mfxExtBuffer *)&q->extco2;
  481. }
  482. #endif
  483. }
  484. if (!rc_supported(q)) {
  485. av_log(avctx, AV_LOG_ERROR,
  486. "Selected ratecontrol mode is not supported by the QSV "
  487. "runtime. Choose a different mode.\n");
  488. return AVERROR(ENOSYS);
  489. }
  490. return 0;
  491. }
  492. static int qsv_retrieve_enc_params(AVCodecContext *avctx, QSVEncContext *q)
  493. {
  494. AVCPBProperties *cpb_props;
  495. uint8_t sps_buf[128];
  496. uint8_t pps_buf[128];
  497. mfxExtCodingOptionSPSPPS extradata = {
  498. .Header.BufferId = MFX_EXTBUFF_CODING_OPTION_SPSPPS,
  499. .Header.BufferSz = sizeof(extradata),
  500. .SPSBuffer = sps_buf, .SPSBufSize = sizeof(sps_buf),
  501. .PPSBuffer = pps_buf, .PPSBufSize = sizeof(pps_buf)
  502. };
  503. mfxExtCodingOption co = {
  504. .Header.BufferId = MFX_EXTBUFF_CODING_OPTION,
  505. .Header.BufferSz = sizeof(co),
  506. };
  507. #if QSV_HAVE_CO2
  508. mfxExtCodingOption2 co2 = {
  509. .Header.BufferId = MFX_EXTBUFF_CODING_OPTION2,
  510. .Header.BufferSz = sizeof(co2),
  511. };
  512. #endif
  513. #if QSV_HAVE_CO3
  514. mfxExtCodingOption3 co3 = {
  515. .Header.BufferId = MFX_EXTBUFF_CODING_OPTION3,
  516. .Header.BufferSz = sizeof(co3),
  517. };
  518. #endif
  519. mfxExtBuffer *ext_buffers[] = {
  520. (mfxExtBuffer*)&extradata,
  521. (mfxExtBuffer*)&co,
  522. #if QSV_HAVE_CO2
  523. (mfxExtBuffer*)&co2,
  524. #endif
  525. #if QSV_HAVE_CO3
  526. (mfxExtBuffer*)&co3,
  527. #endif
  528. };
  529. int need_pps = avctx->codec_id != AV_CODEC_ID_MPEG2VIDEO;
  530. int ret;
  531. q->param.ExtParam = ext_buffers;
  532. q->param.NumExtParam = FF_ARRAY_ELEMS(ext_buffers);
  533. ret = MFXVideoENCODE_GetVideoParam(q->session, &q->param);
  534. if (ret < 0)
  535. return ff_qsv_error(ret);
  536. q->packet_size = q->param.mfx.BufferSizeInKB * 1000;
  537. if (!extradata.SPSBufSize || (need_pps && !extradata.PPSBufSize)) {
  538. av_log(avctx, AV_LOG_ERROR, "No extradata returned from libmfx.\n");
  539. return AVERROR_UNKNOWN;
  540. }
  541. avctx->extradata = av_malloc(extradata.SPSBufSize + need_pps * extradata.PPSBufSize +
  542. AV_INPUT_BUFFER_PADDING_SIZE);
  543. if (!avctx->extradata)
  544. return AVERROR(ENOMEM);
  545. memcpy(avctx->extradata, sps_buf, extradata.SPSBufSize);
  546. if (need_pps)
  547. memcpy(avctx->extradata + extradata.SPSBufSize, pps_buf, extradata.PPSBufSize);
  548. avctx->extradata_size = extradata.SPSBufSize + need_pps * extradata.PPSBufSize;
  549. memset(avctx->extradata + avctx->extradata_size, 0, AV_INPUT_BUFFER_PADDING_SIZE);
  550. cpb_props = ff_add_cpb_side_data(avctx);
  551. if (!cpb_props)
  552. return AVERROR(ENOMEM);
  553. cpb_props->max_bitrate = avctx->rc_max_rate;
  554. cpb_props->min_bitrate = avctx->rc_min_rate;
  555. cpb_props->avg_bitrate = avctx->bit_rate;
  556. cpb_props->buffer_size = avctx->rc_buffer_size;
  557. dump_video_param(avctx, q, ext_buffers + 1);
  558. return 0;
  559. }
  560. static int qsv_init_opaque_alloc(AVCodecContext *avctx, QSVEncContext *q)
  561. {
  562. AVQSVContext *qsv = avctx->hwaccel_context;
  563. mfxFrameSurface1 *surfaces;
  564. int nb_surfaces, i;
  565. nb_surfaces = qsv->nb_opaque_surfaces + q->req.NumFrameSuggested + q->async_depth;
  566. q->opaque_alloc_buf = av_buffer_allocz(sizeof(*surfaces) * nb_surfaces);
  567. if (!q->opaque_alloc_buf)
  568. return AVERROR(ENOMEM);
  569. q->opaque_surfaces = av_malloc_array(nb_surfaces, sizeof(*q->opaque_surfaces));
  570. if (!q->opaque_surfaces)
  571. return AVERROR(ENOMEM);
  572. surfaces = (mfxFrameSurface1*)q->opaque_alloc_buf->data;
  573. for (i = 0; i < nb_surfaces; i++) {
  574. surfaces[i].Info = q->req.Info;
  575. q->opaque_surfaces[i] = surfaces + i;
  576. }
  577. q->opaque_alloc.Header.BufferId = MFX_EXTBUFF_OPAQUE_SURFACE_ALLOCATION;
  578. q->opaque_alloc.Header.BufferSz = sizeof(q->opaque_alloc);
  579. q->opaque_alloc.In.Surfaces = q->opaque_surfaces;
  580. q->opaque_alloc.In.NumSurface = nb_surfaces;
  581. q->opaque_alloc.In.Type = q->req.Type;
  582. q->extparam_internal[q->nb_extparam_internal++] = (mfxExtBuffer *)&q->opaque_alloc;
  583. qsv->nb_opaque_surfaces = nb_surfaces;
  584. qsv->opaque_surfaces = q->opaque_alloc_buf;
  585. qsv->opaque_alloc_type = q->req.Type;
  586. return 0;
  587. }
  588. static int qsvenc_init_session(AVCodecContext *avctx, QSVEncContext *q)
  589. {
  590. int ret;
  591. if (avctx->hwaccel_context) {
  592. AVQSVContext *qsv = avctx->hwaccel_context;
  593. q->session = qsv->session;
  594. } else if (avctx->hw_frames_ctx) {
  595. q->frames_ctx.hw_frames_ctx = av_buffer_ref(avctx->hw_frames_ctx);
  596. if (!q->frames_ctx.hw_frames_ctx)
  597. return AVERROR(ENOMEM);
  598. ret = ff_qsv_init_session_hwcontext(avctx, &q->internal_session,
  599. &q->frames_ctx, q->load_plugins,
  600. q->param.IOPattern == MFX_IOPATTERN_IN_OPAQUE_MEMORY);
  601. if (ret < 0) {
  602. av_buffer_unref(&q->frames_ctx.hw_frames_ctx);
  603. return ret;
  604. }
  605. q->session = q->internal_session;
  606. } else {
  607. ret = ff_qsv_init_internal_session(avctx, &q->internal_session,
  608. q->load_plugins);
  609. if (ret < 0)
  610. return ret;
  611. q->session = q->internal_session;
  612. }
  613. return 0;
  614. }
  615. int ff_qsv_enc_init(AVCodecContext *avctx, QSVEncContext *q)
  616. {
  617. int iopattern = 0;
  618. int opaque_alloc = 0;
  619. int ret;
  620. q->param.AsyncDepth = q->async_depth;
  621. q->async_fifo = av_fifo_alloc((1 + q->async_depth) *
  622. (sizeof(AVPacket) + sizeof(mfxSyncPoint*) + sizeof(mfxBitstream*)));
  623. if (!q->async_fifo)
  624. return AVERROR(ENOMEM);
  625. if (avctx->hwaccel_context) {
  626. AVQSVContext *qsv = avctx->hwaccel_context;
  627. iopattern = qsv->iopattern;
  628. opaque_alloc = qsv->opaque_alloc;
  629. }
  630. if (avctx->hw_frames_ctx) {
  631. AVHWFramesContext *frames_ctx = (AVHWFramesContext*)avctx->hw_frames_ctx->data;
  632. AVQSVFramesContext *frames_hwctx = frames_ctx->hwctx;
  633. if (!iopattern) {
  634. if (frames_hwctx->frame_type & MFX_MEMTYPE_OPAQUE_FRAME)
  635. iopattern = MFX_IOPATTERN_IN_OPAQUE_MEMORY;
  636. else if (frames_hwctx->frame_type &
  637. (MFX_MEMTYPE_VIDEO_MEMORY_DECODER_TARGET | MFX_MEMTYPE_VIDEO_MEMORY_PROCESSOR_TARGET))
  638. iopattern = MFX_IOPATTERN_IN_VIDEO_MEMORY;
  639. }
  640. }
  641. if (!iopattern)
  642. iopattern = MFX_IOPATTERN_IN_SYSTEM_MEMORY;
  643. q->param.IOPattern = iopattern;
  644. ret = qsvenc_init_session(avctx, q);
  645. if (ret < 0)
  646. return ret;
  647. ret = init_video_param(avctx, q);
  648. if (ret < 0)
  649. return ret;
  650. ret = MFXVideoENCODE_QueryIOSurf(q->session, &q->param, &q->req);
  651. if (ret < 0) {
  652. av_log(avctx, AV_LOG_ERROR, "Error querying the encoding parameters\n");
  653. return ff_qsv_error(ret);
  654. }
  655. if (opaque_alloc) {
  656. ret = qsv_init_opaque_alloc(avctx, q);
  657. if (ret < 0)
  658. return ret;
  659. }
  660. if (avctx->hwaccel_context) {
  661. AVQSVContext *qsv = avctx->hwaccel_context;
  662. int i, j;
  663. q->extparam = av_mallocz_array(qsv->nb_ext_buffers + q->nb_extparam_internal,
  664. sizeof(*q->extparam));
  665. if (!q->extparam)
  666. return AVERROR(ENOMEM);
  667. q->param.ExtParam = q->extparam;
  668. for (i = 0; i < qsv->nb_ext_buffers; i++)
  669. q->param.ExtParam[i] = qsv->ext_buffers[i];
  670. q->param.NumExtParam = qsv->nb_ext_buffers;
  671. for (i = 0; i < q->nb_extparam_internal; i++) {
  672. for (j = 0; j < qsv->nb_ext_buffers; j++) {
  673. if (qsv->ext_buffers[j]->BufferId == q->extparam_internal[i]->BufferId)
  674. break;
  675. }
  676. if (j < qsv->nb_ext_buffers)
  677. continue;
  678. q->param.ExtParam[q->param.NumExtParam++] = q->extparam_internal[i];
  679. }
  680. } else {
  681. q->param.ExtParam = q->extparam_internal;
  682. q->param.NumExtParam = q->nb_extparam_internal;
  683. }
  684. ret = MFXVideoENCODE_Init(q->session, &q->param);
  685. if (ret < 0) {
  686. av_log(avctx, AV_LOG_ERROR, "Error initializing the encoder\n");
  687. return ff_qsv_error(ret);
  688. }
  689. ret = qsv_retrieve_enc_params(avctx, q);
  690. if (ret < 0) {
  691. av_log(avctx, AV_LOG_ERROR, "Error retrieving encoding parameters.\n");
  692. return ret;
  693. }
  694. q->avctx = avctx;
  695. return 0;
  696. }
  697. static void clear_unused_frames(QSVEncContext *q)
  698. {
  699. QSVFrame *cur = q->work_frames;
  700. while (cur) {
  701. if (cur->surface && !cur->surface->Data.Locked) {
  702. cur->surface = NULL;
  703. av_frame_unref(cur->frame);
  704. }
  705. cur = cur->next;
  706. }
  707. }
  708. static int get_free_frame(QSVEncContext *q, QSVFrame **f)
  709. {
  710. QSVFrame *frame, **last;
  711. clear_unused_frames(q);
  712. frame = q->work_frames;
  713. last = &q->work_frames;
  714. while (frame) {
  715. if (!frame->surface) {
  716. *f = frame;
  717. return 0;
  718. }
  719. last = &frame->next;
  720. frame = frame->next;
  721. }
  722. frame = av_mallocz(sizeof(*frame));
  723. if (!frame)
  724. return AVERROR(ENOMEM);
  725. frame->frame = av_frame_alloc();
  726. if (!frame->frame) {
  727. av_freep(&frame);
  728. return AVERROR(ENOMEM);
  729. }
  730. *last = frame;
  731. *f = frame;
  732. return 0;
  733. }
  734. static int submit_frame(QSVEncContext *q, const AVFrame *frame,
  735. mfxFrameSurface1 **surface)
  736. {
  737. QSVFrame *qf;
  738. int ret;
  739. ret = get_free_frame(q, &qf);
  740. if (ret < 0)
  741. return ret;
  742. if (frame->format == AV_PIX_FMT_QSV) {
  743. ret = av_frame_ref(qf->frame, frame);
  744. if (ret < 0)
  745. return ret;
  746. qf->surface = (mfxFrameSurface1*)qf->frame->data[3];
  747. } else {
  748. /* make a copy if the input is not padded as libmfx requires */
  749. if (frame->height & 31 || frame->linesize[0] & (q->width_align - 1)) {
  750. qf->frame->height = FFALIGN(frame->height, 32);
  751. qf->frame->width = FFALIGN(frame->width, q->width_align);
  752. ret = ff_get_buffer(q->avctx, qf->frame, AV_GET_BUFFER_FLAG_REF);
  753. if (ret < 0)
  754. return ret;
  755. qf->frame->height = frame->height;
  756. qf->frame->width = frame->width;
  757. ret = av_frame_copy(qf->frame, frame);
  758. if (ret < 0) {
  759. av_frame_unref(qf->frame);
  760. return ret;
  761. }
  762. } else {
  763. ret = av_frame_ref(qf->frame, frame);
  764. if (ret < 0)
  765. return ret;
  766. }
  767. qf->surface_internal.Info = q->param.mfx.FrameInfo;
  768. qf->surface_internal.Info.PicStruct =
  769. !frame->interlaced_frame ? MFX_PICSTRUCT_PROGRESSIVE :
  770. frame->top_field_first ? MFX_PICSTRUCT_FIELD_TFF :
  771. MFX_PICSTRUCT_FIELD_BFF;
  772. if (frame->repeat_pict == 1)
  773. qf->surface_internal.Info.PicStruct |= MFX_PICSTRUCT_FIELD_REPEATED;
  774. else if (frame->repeat_pict == 2)
  775. qf->surface_internal.Info.PicStruct |= MFX_PICSTRUCT_FRAME_DOUBLING;
  776. else if (frame->repeat_pict == 4)
  777. qf->surface_internal.Info.PicStruct |= MFX_PICSTRUCT_FRAME_TRIPLING;
  778. qf->surface_internal.Data.PitchLow = qf->frame->linesize[0];
  779. qf->surface_internal.Data.Y = qf->frame->data[0];
  780. qf->surface_internal.Data.UV = qf->frame->data[1];
  781. qf->surface = &qf->surface_internal;
  782. }
  783. qf->surface->Data.TimeStamp = av_rescale_q(frame->pts, q->avctx->time_base, (AVRational){1, 90000});
  784. *surface = qf->surface;
  785. return 0;
  786. }
  787. static void print_interlace_msg(AVCodecContext *avctx, QSVEncContext *q)
  788. {
  789. if (q->param.mfx.CodecId == MFX_CODEC_AVC) {
  790. if (q->param.mfx.CodecProfile == MFX_PROFILE_AVC_BASELINE ||
  791. q->param.mfx.CodecLevel < MFX_LEVEL_AVC_21 ||
  792. q->param.mfx.CodecLevel > MFX_LEVEL_AVC_41)
  793. av_log(avctx, AV_LOG_WARNING,
  794. "Interlaced coding is supported"
  795. " at Main/High Profile Level 2.1-4.1\n");
  796. }
  797. }
  798. static int encode_frame(AVCodecContext *avctx, QSVEncContext *q,
  799. const AVFrame *frame)
  800. {
  801. AVPacket new_pkt = { 0 };
  802. mfxBitstream *bs;
  803. mfxFrameSurface1 *surf = NULL;
  804. mfxSyncPoint *sync = NULL;
  805. int ret;
  806. if (frame) {
  807. ret = submit_frame(q, frame, &surf);
  808. if (ret < 0) {
  809. av_log(avctx, AV_LOG_ERROR, "Error submitting the frame for encoding.\n");
  810. return ret;
  811. }
  812. }
  813. ret = av_new_packet(&new_pkt, q->packet_size);
  814. if (ret < 0) {
  815. av_log(avctx, AV_LOG_ERROR, "Error allocating the output packet\n");
  816. return ret;
  817. }
  818. bs = av_mallocz(sizeof(*bs));
  819. if (!bs) {
  820. av_packet_unref(&new_pkt);
  821. return AVERROR(ENOMEM);
  822. }
  823. bs->Data = new_pkt.data;
  824. bs->MaxLength = new_pkt.size;
  825. sync = av_mallocz(sizeof(*sync));
  826. if (!sync) {
  827. av_freep(&bs);
  828. av_packet_unref(&new_pkt);
  829. return AVERROR(ENOMEM);
  830. }
  831. do {
  832. ret = MFXVideoENCODE_EncodeFrameAsync(q->session, NULL, surf, bs, sync);
  833. if (ret == MFX_WRN_DEVICE_BUSY)
  834. av_usleep(1);
  835. } while (ret > 0);
  836. if (ret < 0) {
  837. av_packet_unref(&new_pkt);
  838. av_freep(&bs);
  839. av_freep(&sync);
  840. return (ret == MFX_ERR_MORE_DATA) ? 0 : ff_qsv_error(ret);
  841. }
  842. if (ret == MFX_WRN_INCOMPATIBLE_VIDEO_PARAM && frame->interlaced_frame)
  843. print_interlace_msg(avctx, q);
  844. if (*sync) {
  845. av_fifo_generic_write(q->async_fifo, &new_pkt, sizeof(new_pkt), NULL);
  846. av_fifo_generic_write(q->async_fifo, &sync, sizeof(sync), NULL);
  847. av_fifo_generic_write(q->async_fifo, &bs, sizeof(bs), NULL);
  848. } else {
  849. av_freep(&sync);
  850. av_packet_unref(&new_pkt);
  851. av_freep(&bs);
  852. }
  853. return 0;
  854. }
  855. int ff_qsv_encode(AVCodecContext *avctx, QSVEncContext *q,
  856. AVPacket *pkt, const AVFrame *frame, int *got_packet)
  857. {
  858. int ret;
  859. ret = encode_frame(avctx, q, frame);
  860. if (ret < 0)
  861. return ret;
  862. if (!av_fifo_space(q->async_fifo) ||
  863. (!frame && av_fifo_size(q->async_fifo))) {
  864. AVPacket new_pkt;
  865. mfxBitstream *bs;
  866. mfxSyncPoint *sync;
  867. av_fifo_generic_read(q->async_fifo, &new_pkt, sizeof(new_pkt), NULL);
  868. av_fifo_generic_read(q->async_fifo, &sync, sizeof(sync), NULL);
  869. av_fifo_generic_read(q->async_fifo, &bs, sizeof(bs), NULL);
  870. do {
  871. ret = MFXVideoCORE_SyncOperation(q->session, *sync, 1000);
  872. } while (ret == MFX_WRN_IN_EXECUTION);
  873. new_pkt.dts = av_rescale_q(bs->DecodeTimeStamp, (AVRational){1, 90000}, avctx->time_base);
  874. new_pkt.pts = av_rescale_q(bs->TimeStamp, (AVRational){1, 90000}, avctx->time_base);
  875. new_pkt.size = bs->DataLength;
  876. if (bs->FrameType & MFX_FRAMETYPE_IDR ||
  877. bs->FrameType & MFX_FRAMETYPE_xIDR)
  878. new_pkt.flags |= AV_PKT_FLAG_KEY;
  879. #if FF_API_CODED_FRAME
  880. FF_DISABLE_DEPRECATION_WARNINGS
  881. if (bs->FrameType & MFX_FRAMETYPE_I || bs->FrameType & MFX_FRAMETYPE_xI)
  882. avctx->coded_frame->pict_type = AV_PICTURE_TYPE_I;
  883. else if (bs->FrameType & MFX_FRAMETYPE_P || bs->FrameType & MFX_FRAMETYPE_xP)
  884. avctx->coded_frame->pict_type = AV_PICTURE_TYPE_P;
  885. else if (bs->FrameType & MFX_FRAMETYPE_B || bs->FrameType & MFX_FRAMETYPE_xB)
  886. avctx->coded_frame->pict_type = AV_PICTURE_TYPE_B;
  887. FF_ENABLE_DEPRECATION_WARNINGS
  888. #endif
  889. av_freep(&bs);
  890. av_freep(&sync);
  891. if (pkt->data) {
  892. if (pkt->size < new_pkt.size) {
  893. av_log(avctx, AV_LOG_ERROR, "Submitted buffer not large enough: %d < %d\n",
  894. pkt->size, new_pkt.size);
  895. av_packet_unref(&new_pkt);
  896. return AVERROR(EINVAL);
  897. }
  898. memcpy(pkt->data, new_pkt.data, new_pkt.size);
  899. pkt->size = new_pkt.size;
  900. ret = av_packet_copy_props(pkt, &new_pkt);
  901. av_packet_unref(&new_pkt);
  902. if (ret < 0)
  903. return ret;
  904. } else
  905. *pkt = new_pkt;
  906. *got_packet = 1;
  907. }
  908. return 0;
  909. }
  910. int ff_qsv_enc_close(AVCodecContext *avctx, QSVEncContext *q)
  911. {
  912. QSVFrame *cur;
  913. if (q->session)
  914. MFXVideoENCODE_Close(q->session);
  915. if (q->internal_session)
  916. MFXClose(q->internal_session);
  917. q->session = NULL;
  918. q->internal_session = NULL;
  919. av_buffer_unref(&q->frames_ctx.hw_frames_ctx);
  920. av_freep(&q->frames_ctx.mids);
  921. q->frames_ctx.nb_mids = 0;
  922. cur = q->work_frames;
  923. while (cur) {
  924. q->work_frames = cur->next;
  925. av_frame_free(&cur->frame);
  926. av_freep(&cur);
  927. cur = q->work_frames;
  928. }
  929. while (q->async_fifo && av_fifo_size(q->async_fifo)) {
  930. AVPacket pkt;
  931. mfxSyncPoint *sync;
  932. mfxBitstream *bs;
  933. av_fifo_generic_read(q->async_fifo, &pkt, sizeof(pkt), NULL);
  934. av_fifo_generic_read(q->async_fifo, &sync, sizeof(sync), NULL);
  935. av_fifo_generic_read(q->async_fifo, &bs, sizeof(bs), NULL);
  936. av_freep(&sync);
  937. av_freep(&bs);
  938. av_packet_unref(&pkt);
  939. }
  940. av_fifo_free(q->async_fifo);
  941. q->async_fifo = NULL;
  942. av_freep(&q->opaque_surfaces);
  943. av_buffer_unref(&q->opaque_alloc_buf);
  944. av_freep(&q->extparam);
  945. return 0;
  946. }