You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

810 lines
32KB

  1. /*
  2. * ffmpeg filter configuration
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. #include "ffmpeg.h"
  21. #include "libavfilter/avfilter.h"
  22. #include "libavfilter/buffersink.h"
  23. #include "libavresample/avresample.h"
  24. #include "libavutil/avassert.h"
  25. #include "libavutil/avstring.h"
  26. #include "libavutil/bprint.h"
  27. #include "libavutil/channel_layout.h"
  28. #include "libavutil/opt.h"
  29. #include "libavutil/pixdesc.h"
  30. #include "libavutil/pixfmt.h"
  31. #include "libavutil/imgutils.h"
  32. #include "libavutil/samplefmt.h"
  33. enum AVPixelFormat choose_pixel_fmt(AVStream *st, AVCodec *codec, enum AVPixelFormat target)
  34. {
  35. if (codec && codec->pix_fmts) {
  36. const enum AVPixelFormat *p = codec->pix_fmts;
  37. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(target);
  38. int has_alpha = desc ? desc->nb_components % 2 == 0 : 0;
  39. enum AVPixelFormat best= AV_PIX_FMT_NONE;
  40. if (st->codec->strict_std_compliance <= FF_COMPLIANCE_UNOFFICIAL) {
  41. if (st->codec->codec_id == AV_CODEC_ID_MJPEG) {
  42. p = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_NONE };
  43. } else if (st->codec->codec_id == AV_CODEC_ID_LJPEG) {
  44. p = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUV420P,
  45. AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_BGRA, AV_PIX_FMT_NONE };
  46. }
  47. }
  48. for (; *p != AV_PIX_FMT_NONE; p++) {
  49. best= avcodec_find_best_pix_fmt_of_2(best, *p, target, has_alpha, NULL);
  50. if (*p == target)
  51. break;
  52. }
  53. if (*p == AV_PIX_FMT_NONE) {
  54. if (target != AV_PIX_FMT_NONE)
  55. av_log(NULL, AV_LOG_WARNING,
  56. "Incompatible pixel format '%s' for codec '%s', auto-selecting format '%s'\n",
  57. av_get_pix_fmt_name(target),
  58. codec->name,
  59. av_get_pix_fmt_name(best));
  60. return best;
  61. }
  62. }
  63. return target;
  64. }
  65. void choose_sample_fmt(AVStream *st, AVCodec *codec)
  66. {
  67. if (codec && codec->sample_fmts) {
  68. const enum AVSampleFormat *p = codec->sample_fmts;
  69. for (; *p != -1; p++) {
  70. if (*p == st->codec->sample_fmt)
  71. break;
  72. }
  73. if (*p == -1) {
  74. if((codec->capabilities & CODEC_CAP_LOSSLESS) && av_get_sample_fmt_name(st->codec->sample_fmt) > av_get_sample_fmt_name(codec->sample_fmts[0]))
  75. av_log(NULL, AV_LOG_ERROR, "Conversion will not be lossless.\n");
  76. if(av_get_sample_fmt_name(st->codec->sample_fmt))
  77. av_log(NULL, AV_LOG_WARNING,
  78. "Incompatible sample format '%s' for codec '%s', auto-selecting format '%s'\n",
  79. av_get_sample_fmt_name(st->codec->sample_fmt),
  80. codec->name,
  81. av_get_sample_fmt_name(codec->sample_fmts[0]));
  82. st->codec->sample_fmt = codec->sample_fmts[0];
  83. }
  84. }
  85. }
  86. static char *choose_pix_fmts(OutputStream *ost)
  87. {
  88. if (ost->keep_pix_fmt) {
  89. if (ost->filter)
  90. avfilter_graph_set_auto_convert(ost->filter->graph->graph,
  91. AVFILTER_AUTO_CONVERT_NONE);
  92. if (ost->st->codec->pix_fmt == AV_PIX_FMT_NONE)
  93. return NULL;
  94. return av_strdup(av_get_pix_fmt_name(ost->st->codec->pix_fmt));
  95. }
  96. if (ost->st->codec->pix_fmt != AV_PIX_FMT_NONE) {
  97. return av_strdup(av_get_pix_fmt_name(choose_pixel_fmt(ost->st, ost->enc, ost->st->codec->pix_fmt)));
  98. } else if (ost->enc && ost->enc->pix_fmts) {
  99. const enum AVPixelFormat *p;
  100. AVIOContext *s = NULL;
  101. uint8_t *ret;
  102. int len;
  103. if (avio_open_dyn_buf(&s) < 0)
  104. exit(1);
  105. p = ost->enc->pix_fmts;
  106. if (ost->st->codec->strict_std_compliance <= FF_COMPLIANCE_UNOFFICIAL) {
  107. if (ost->st->codec->codec_id == AV_CODEC_ID_MJPEG) {
  108. p = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_NONE };
  109. } else if (ost->st->codec->codec_id == AV_CODEC_ID_LJPEG) {
  110. p = (const enum AVPixelFormat[]) { AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ422P, AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUV420P,
  111. AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV444P, AV_PIX_FMT_BGRA, AV_PIX_FMT_NONE };
  112. }
  113. }
  114. for (; *p != AV_PIX_FMT_NONE; p++) {
  115. const char *name = av_get_pix_fmt_name(*p);
  116. avio_printf(s, "%s|", name);
  117. }
  118. len = avio_close_dyn_buf(s, &ret);
  119. ret[len - 1] = 0;
  120. return ret;
  121. } else
  122. return NULL;
  123. }
  124. /* Define a function for building a string containing a list of
  125. * allowed formats. */
  126. #define DEF_CHOOSE_FORMAT(type, var, supported_list, none, get_name) \
  127. static char *choose_ ## var ## s(OutputStream *ost) \
  128. { \
  129. if (ost->st->codec->var != none) { \
  130. get_name(ost->st->codec->var); \
  131. return av_strdup(name); \
  132. } else if (ost->enc && ost->enc->supported_list) { \
  133. const type *p; \
  134. AVIOContext *s = NULL; \
  135. uint8_t *ret; \
  136. int len; \
  137. \
  138. if (avio_open_dyn_buf(&s) < 0) \
  139. exit(1); \
  140. \
  141. for (p = ost->enc->supported_list; *p != none; p++) { \
  142. get_name(*p); \
  143. avio_printf(s, "%s|", name); \
  144. } \
  145. len = avio_close_dyn_buf(s, &ret); \
  146. ret[len - 1] = 0; \
  147. return ret; \
  148. } else \
  149. return NULL; \
  150. }
  151. // DEF_CHOOSE_FORMAT(enum AVPixelFormat, pix_fmt, pix_fmts, AV_PIX_FMT_NONE,
  152. // GET_PIX_FMT_NAME)
  153. DEF_CHOOSE_FORMAT(enum AVSampleFormat, sample_fmt, sample_fmts,
  154. AV_SAMPLE_FMT_NONE, GET_SAMPLE_FMT_NAME)
  155. DEF_CHOOSE_FORMAT(int, sample_rate, supported_samplerates, 0,
  156. GET_SAMPLE_RATE_NAME)
  157. DEF_CHOOSE_FORMAT(uint64_t, channel_layout, channel_layouts, 0,
  158. GET_CH_LAYOUT_NAME)
  159. FilterGraph *init_simple_filtergraph(InputStream *ist, OutputStream *ost)
  160. {
  161. FilterGraph *fg = av_mallocz(sizeof(*fg));
  162. if (!fg)
  163. exit(1);
  164. fg->index = nb_filtergraphs;
  165. GROW_ARRAY(fg->outputs, fg->nb_outputs);
  166. if (!(fg->outputs[0] = av_mallocz(sizeof(*fg->outputs[0]))))
  167. exit(1);
  168. fg->outputs[0]->ost = ost;
  169. fg->outputs[0]->graph = fg;
  170. ost->filter = fg->outputs[0];
  171. GROW_ARRAY(fg->inputs, fg->nb_inputs);
  172. if (!(fg->inputs[0] = av_mallocz(sizeof(*fg->inputs[0]))))
  173. exit(1);
  174. fg->inputs[0]->ist = ist;
  175. fg->inputs[0]->graph = fg;
  176. GROW_ARRAY(ist->filters, ist->nb_filters);
  177. ist->filters[ist->nb_filters - 1] = fg->inputs[0];
  178. GROW_ARRAY(filtergraphs, nb_filtergraphs);
  179. filtergraphs[nb_filtergraphs - 1] = fg;
  180. return fg;
  181. }
  182. static void init_input_filter(FilterGraph *fg, AVFilterInOut *in)
  183. {
  184. InputStream *ist = NULL;
  185. enum AVMediaType type = avfilter_pad_get_type(in->filter_ctx->input_pads, in->pad_idx);
  186. int i;
  187. // TODO: support other filter types
  188. if (type != AVMEDIA_TYPE_VIDEO && type != AVMEDIA_TYPE_AUDIO) {
  189. av_log(NULL, AV_LOG_FATAL, "Only video and audio filters supported "
  190. "currently.\n");
  191. exit(1);
  192. }
  193. if (in->name) {
  194. AVFormatContext *s;
  195. AVStream *st = NULL;
  196. char *p;
  197. int file_idx = strtol(in->name, &p, 0);
  198. if (file_idx < 0 || file_idx >= nb_input_files) {
  199. av_log(NULL, AV_LOG_FATAL, "Invalid file index %d in filtergraph description %s.\n",
  200. file_idx, fg->graph_desc);
  201. exit(1);
  202. }
  203. s = input_files[file_idx]->ctx;
  204. for (i = 0; i < s->nb_streams; i++) {
  205. enum AVMediaType stream_type = s->streams[i]->codec->codec_type;
  206. if (stream_type != type &&
  207. !(stream_type == AVMEDIA_TYPE_SUBTITLE &&
  208. type == AVMEDIA_TYPE_VIDEO /* sub2video hack */))
  209. continue;
  210. if (check_stream_specifier(s, s->streams[i], *p == ':' ? p + 1 : p) == 1) {
  211. st = s->streams[i];
  212. break;
  213. }
  214. }
  215. if (!st) {
  216. av_log(NULL, AV_LOG_FATAL, "Stream specifier '%s' in filtergraph description %s "
  217. "matches no streams.\n", p, fg->graph_desc);
  218. exit(1);
  219. }
  220. ist = input_streams[input_files[file_idx]->ist_index + st->index];
  221. } else {
  222. /* find the first unused stream of corresponding type */
  223. for (i = 0; i < nb_input_streams; i++) {
  224. ist = input_streams[i];
  225. if (ist->st->codec->codec_type == type && ist->discard)
  226. break;
  227. }
  228. if (i == nb_input_streams) {
  229. av_log(NULL, AV_LOG_FATAL, "Cannot find a matching stream for "
  230. "unlabeled input pad %d on filter %s\n", in->pad_idx,
  231. in->filter_ctx->name);
  232. exit(1);
  233. }
  234. }
  235. av_assert0(ist);
  236. ist->discard = 0;
  237. ist->decoding_needed++;
  238. ist->st->discard = AVDISCARD_NONE;
  239. GROW_ARRAY(fg->inputs, fg->nb_inputs);
  240. if (!(fg->inputs[fg->nb_inputs - 1] = av_mallocz(sizeof(*fg->inputs[0]))))
  241. exit(1);
  242. fg->inputs[fg->nb_inputs - 1]->ist = ist;
  243. fg->inputs[fg->nb_inputs - 1]->graph = fg;
  244. GROW_ARRAY(ist->filters, ist->nb_filters);
  245. ist->filters[ist->nb_filters - 1] = fg->inputs[fg->nb_inputs - 1];
  246. }
  247. static int configure_output_video_filter(FilterGraph *fg, OutputFilter *ofilter, AVFilterInOut *out)
  248. {
  249. char *pix_fmts;
  250. OutputStream *ost = ofilter->ost;
  251. AVCodecContext *codec = ost->st->codec;
  252. AVFilterContext *last_filter = out->filter_ctx;
  253. int pad_idx = out->pad_idx;
  254. int ret;
  255. char name[255];
  256. snprintf(name, sizeof(name), "output stream %d:%d", ost->file_index, ost->index);
  257. ret = avfilter_graph_create_filter(&ofilter->filter,
  258. avfilter_get_by_name("buffersink"),
  259. name, NULL, NULL, fg->graph);
  260. if (ret < 0)
  261. return ret;
  262. if (codec->width || codec->height) {
  263. char args[255];
  264. AVFilterContext *filter;
  265. snprintf(args, sizeof(args), "%d:%d:flags=0x%X",
  266. codec->width,
  267. codec->height,
  268. (unsigned)ost->sws_flags);
  269. snprintf(name, sizeof(name), "scaler for output stream %d:%d",
  270. ost->file_index, ost->index);
  271. if ((ret = avfilter_graph_create_filter(&filter, avfilter_get_by_name("scale"),
  272. name, args, NULL, fg->graph)) < 0)
  273. return ret;
  274. if ((ret = avfilter_link(last_filter, pad_idx, filter, 0)) < 0)
  275. return ret;
  276. last_filter = filter;
  277. pad_idx = 0;
  278. }
  279. if ((pix_fmts = choose_pix_fmts(ost))) {
  280. AVFilterContext *filter;
  281. snprintf(name, sizeof(name), "pixel format for output stream %d:%d",
  282. ost->file_index, ost->index);
  283. ret = avfilter_graph_create_filter(&filter,
  284. avfilter_get_by_name("format"),
  285. "format", pix_fmts, NULL,
  286. fg->graph);
  287. av_freep(&pix_fmts);
  288. if (ret < 0)
  289. return ret;
  290. if ((ret = avfilter_link(last_filter, pad_idx, filter, 0)) < 0)
  291. return ret;
  292. last_filter = filter;
  293. pad_idx = 0;
  294. }
  295. if (ost->frame_rate.num && 0) {
  296. AVFilterContext *fps;
  297. char args[255];
  298. snprintf(args, sizeof(args), "fps=%d/%d", ost->frame_rate.num,
  299. ost->frame_rate.den);
  300. snprintf(name, sizeof(name), "fps for output stream %d:%d",
  301. ost->file_index, ost->index);
  302. ret = avfilter_graph_create_filter(&fps, avfilter_get_by_name("fps"),
  303. name, args, NULL, fg->graph);
  304. if (ret < 0)
  305. return ret;
  306. ret = avfilter_link(last_filter, pad_idx, fps, 0);
  307. if (ret < 0)
  308. return ret;
  309. last_filter = fps;
  310. pad_idx = 0;
  311. }
  312. if ((ret = avfilter_link(last_filter, pad_idx, ofilter->filter, 0)) < 0)
  313. return ret;
  314. return 0;
  315. }
  316. static int configure_output_audio_filter(FilterGraph *fg, OutputFilter *ofilter, AVFilterInOut *out)
  317. {
  318. OutputStream *ost = ofilter->ost;
  319. AVCodecContext *codec = ost->st->codec;
  320. AVFilterContext *last_filter = out->filter_ctx;
  321. int pad_idx = out->pad_idx;
  322. char *sample_fmts, *sample_rates, *channel_layouts;
  323. char name[255];
  324. int ret;
  325. snprintf(name, sizeof(name), "output stream %d:%d", ost->file_index, ost->index);
  326. ret = avfilter_graph_create_filter(&ofilter->filter,
  327. avfilter_get_by_name("abuffersink"),
  328. name, NULL, NULL, fg->graph);
  329. if (ret < 0)
  330. return ret;
  331. if ((ret = av_opt_set_int(ofilter->filter, "all_channel_counts", 1, AV_OPT_SEARCH_CHILDREN)) < 0)
  332. return ret;
  333. #define AUTO_INSERT_FILTER(opt_name, filter_name, arg) do { \
  334. AVFilterContext *filt_ctx; \
  335. \
  336. av_log(NULL, AV_LOG_INFO, opt_name " is forwarded to lavfi " \
  337. "similarly to -af " filter_name "=%s.\n", arg); \
  338. \
  339. ret = avfilter_graph_create_filter(&filt_ctx, \
  340. avfilter_get_by_name(filter_name), \
  341. filter_name, arg, NULL, fg->graph); \
  342. if (ret < 0) \
  343. return ret; \
  344. \
  345. ret = avfilter_link(last_filter, pad_idx, filt_ctx, 0); \
  346. if (ret < 0) \
  347. return ret; \
  348. \
  349. last_filter = filt_ctx; \
  350. pad_idx = 0; \
  351. } while (0)
  352. if (ost->audio_channels_mapped) {
  353. int i;
  354. AVBPrint pan_buf;
  355. av_bprint_init(&pan_buf, 256, 8192);
  356. av_bprintf(&pan_buf, "0x%"PRIx64,
  357. av_get_default_channel_layout(ost->audio_channels_mapped));
  358. for (i = 0; i < ost->audio_channels_mapped; i++)
  359. if (ost->audio_channels_map[i] != -1)
  360. av_bprintf(&pan_buf, ":c%d=c%d", i, ost->audio_channels_map[i]);
  361. AUTO_INSERT_FILTER("-map_channel", "pan", pan_buf.str);
  362. av_bprint_finalize(&pan_buf, NULL);
  363. }
  364. if (codec->channels && !codec->channel_layout)
  365. codec->channel_layout = av_get_default_channel_layout(codec->channels);
  366. sample_fmts = choose_sample_fmts(ost);
  367. sample_rates = choose_sample_rates(ost);
  368. channel_layouts = choose_channel_layouts(ost);
  369. if (sample_fmts || sample_rates || channel_layouts) {
  370. AVFilterContext *format;
  371. char args[256];
  372. args[0] = 0;
  373. if (sample_fmts)
  374. av_strlcatf(args, sizeof(args), "sample_fmts=%s:",
  375. sample_fmts);
  376. if (sample_rates)
  377. av_strlcatf(args, sizeof(args), "sample_rates=%s:",
  378. sample_rates);
  379. if (channel_layouts)
  380. av_strlcatf(args, sizeof(args), "channel_layouts=%s:",
  381. channel_layouts);
  382. av_freep(&sample_fmts);
  383. av_freep(&sample_rates);
  384. av_freep(&channel_layouts);
  385. snprintf(name, sizeof(name), "audio format for output stream %d:%d",
  386. ost->file_index, ost->index);
  387. ret = avfilter_graph_create_filter(&format,
  388. avfilter_get_by_name("aformat"),
  389. name, args, NULL, fg->graph);
  390. if (ret < 0)
  391. return ret;
  392. ret = avfilter_link(last_filter, pad_idx, format, 0);
  393. if (ret < 0)
  394. return ret;
  395. last_filter = format;
  396. pad_idx = 0;
  397. }
  398. if (audio_volume != 256 && 0) {
  399. char args[256];
  400. snprintf(args, sizeof(args), "%f", audio_volume / 256.);
  401. AUTO_INSERT_FILTER("-vol", "volume", args);
  402. }
  403. if ((ret = avfilter_link(last_filter, pad_idx, ofilter->filter, 0)) < 0)
  404. return ret;
  405. return 0;
  406. }
  407. #define DESCRIBE_FILTER_LINK(f, inout, in) \
  408. { \
  409. AVFilterContext *ctx = inout->filter_ctx; \
  410. AVFilterPad *pads = in ? ctx->input_pads : ctx->output_pads; \
  411. int nb_pads = in ? ctx->input_count : ctx->output_count; \
  412. AVIOContext *pb; \
  413. \
  414. if (avio_open_dyn_buf(&pb) < 0) \
  415. exit(1); \
  416. \
  417. avio_printf(pb, "%s", ctx->filter->name); \
  418. if (nb_pads > 1) \
  419. avio_printf(pb, ":%s", avfilter_pad_get_name(pads, inout->pad_idx));\
  420. avio_w8(pb, 0); \
  421. avio_close_dyn_buf(pb, &f->name); \
  422. }
  423. int configure_output_filter(FilterGraph *fg, OutputFilter *ofilter, AVFilterInOut *out)
  424. {
  425. av_freep(&ofilter->name);
  426. DESCRIBE_FILTER_LINK(ofilter, out, 0);
  427. switch (avfilter_pad_get_type(out->filter_ctx->output_pads, out->pad_idx)) {
  428. case AVMEDIA_TYPE_VIDEO: return configure_output_video_filter(fg, ofilter, out);
  429. case AVMEDIA_TYPE_AUDIO: return configure_output_audio_filter(fg, ofilter, out);
  430. default: av_assert0(0);
  431. }
  432. }
  433. static int sub2video_prepare(InputStream *ist)
  434. {
  435. AVFormatContext *avf = input_files[ist->file_index]->ctx;
  436. int i, w, h;
  437. /* Compute the size of the canvas for the subtitles stream.
  438. If the subtitles codec has set a size, use it. Otherwise use the
  439. maximum dimensions of the video streams in the same file. */
  440. w = ist->st->codec->width;
  441. h = ist->st->codec->height;
  442. if (!(w && h)) {
  443. for (i = 0; i < avf->nb_streams; i++) {
  444. if (avf->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO) {
  445. w = FFMAX(w, avf->streams[i]->codec->width);
  446. h = FFMAX(h, avf->streams[i]->codec->height);
  447. }
  448. }
  449. if (!(w && h)) {
  450. w = FFMAX(w, 720);
  451. h = FFMAX(h, 576);
  452. }
  453. av_log(avf, AV_LOG_INFO, "sub2video: using %dx%d canvas\n", w, h);
  454. }
  455. ist->sub2video.w = ist->st->codec->width = ist->resample_width = w;
  456. ist->sub2video.h = ist->st->codec->height = ist->resample_height = h;
  457. /* rectangles are AV_PIX_FMT_PAL8, but we have no guarantee that the
  458. palettes for all rectangles are identical or compatible */
  459. ist->resample_pix_fmt = ist->st->codec->pix_fmt = AV_PIX_FMT_RGB32;
  460. ist->sub2video.frame = av_frame_alloc();
  461. if (!ist->sub2video.frame)
  462. return AVERROR(ENOMEM);
  463. return 0;
  464. }
  465. static int configure_input_video_filter(FilterGraph *fg, InputFilter *ifilter,
  466. AVFilterInOut *in)
  467. {
  468. AVFilterContext *first_filter = in->filter_ctx;
  469. AVFilter *filter = avfilter_get_by_name("buffer");
  470. InputStream *ist = ifilter->ist;
  471. AVRational tb = ist->framerate.num ? av_inv_q(ist->framerate) :
  472. ist->st->time_base;
  473. AVRational fr = ist->framerate;
  474. AVRational sar;
  475. AVBPrint args;
  476. char name[255];
  477. int pad_idx = in->pad_idx;
  478. int ret;
  479. if (!fr.num)
  480. fr = av_guess_frame_rate(input_files[ist->file_index]->ctx, ist->st, NULL);
  481. if (ist->st->codec->codec_type == AVMEDIA_TYPE_SUBTITLE) {
  482. ret = sub2video_prepare(ist);
  483. if (ret < 0)
  484. return ret;
  485. }
  486. sar = ist->st->sample_aspect_ratio.num ?
  487. ist->st->sample_aspect_ratio :
  488. ist->st->codec->sample_aspect_ratio;
  489. if(!sar.den)
  490. sar = (AVRational){0,1};
  491. av_bprint_init(&args, 0, 1);
  492. av_bprintf(&args,
  493. "video_size=%dx%d:pix_fmt=%d:time_base=%d/%d:"
  494. "pixel_aspect=%d/%d:sws_param=flags=%d", ist->resample_width,
  495. ist->resample_height, ist->resample_pix_fmt,
  496. tb.num, tb.den, sar.num, sar.den,
  497. SWS_BILINEAR + ((ist->st->codec->flags&CODEC_FLAG_BITEXACT) ? SWS_BITEXACT:0));
  498. if (fr.num && fr.den)
  499. av_bprintf(&args, ":frame_rate=%d/%d", fr.num, fr.den);
  500. snprintf(name, sizeof(name), "graph %d input from stream %d:%d", fg->index,
  501. ist->file_index, ist->st->index);
  502. if ((ret = avfilter_graph_create_filter(&ifilter->filter, filter, name,
  503. args.str, NULL, fg->graph)) < 0)
  504. return ret;
  505. if (ist->framerate.num) {
  506. AVFilterContext *setpts;
  507. snprintf(name, sizeof(name), "force CFR for input from stream %d:%d",
  508. ist->file_index, ist->st->index);
  509. if ((ret = avfilter_graph_create_filter(&setpts,
  510. avfilter_get_by_name("setpts"),
  511. name, "N", NULL,
  512. fg->graph)) < 0)
  513. return ret;
  514. if ((ret = avfilter_link(setpts, 0, first_filter, pad_idx)) < 0)
  515. return ret;
  516. first_filter = setpts;
  517. pad_idx = 0;
  518. }
  519. if (do_deinterlace) {
  520. AVFilterContext *yadif;
  521. snprintf(name, sizeof(name), "deinterlace input from stream %d:%d",
  522. ist->file_index, ist->st->index);
  523. if ((ret = avfilter_graph_create_filter(&yadif,
  524. avfilter_get_by_name("yadif"),
  525. name, "", NULL,
  526. fg->graph)) < 0)
  527. return ret;
  528. if ((ret = avfilter_link(yadif, 0, first_filter, pad_idx)) < 0)
  529. return ret;
  530. first_filter = yadif;
  531. pad_idx = 0;
  532. }
  533. if ((ret = avfilter_link(ifilter->filter, 0, first_filter, pad_idx)) < 0)
  534. return ret;
  535. return 0;
  536. }
  537. static int configure_input_audio_filter(FilterGraph *fg, InputFilter *ifilter,
  538. AVFilterInOut *in)
  539. {
  540. AVFilterContext *first_filter = in->filter_ctx;
  541. AVFilter *filter = avfilter_get_by_name("abuffer");
  542. InputStream *ist = ifilter->ist;
  543. int pad_idx = in->pad_idx;
  544. AVBPrint args;
  545. char name[255];
  546. int ret;
  547. av_bprint_init(&args, 0, AV_BPRINT_SIZE_AUTOMATIC);
  548. av_bprintf(&args, "time_base=%d/%d:sample_rate=%d:sample_fmt=%s",
  549. 1, ist->st->codec->sample_rate,
  550. ist->st->codec->sample_rate,
  551. av_get_sample_fmt_name(ist->st->codec->sample_fmt));
  552. if (ist->st->codec->channel_layout)
  553. av_bprintf(&args, ":channel_layout=0x%"PRIx64,
  554. ist->st->codec->channel_layout);
  555. else
  556. av_bprintf(&args, ":channels=%d", ist->st->codec->channels);
  557. snprintf(name, sizeof(name), "graph %d input from stream %d:%d", fg->index,
  558. ist->file_index, ist->st->index);
  559. if ((ret = avfilter_graph_create_filter(&ifilter->filter, filter,
  560. name, args.str, NULL,
  561. fg->graph)) < 0)
  562. return ret;
  563. #define AUTO_INSERT_FILTER_INPUT(opt_name, filter_name, arg) do { \
  564. AVFilterContext *filt_ctx; \
  565. \
  566. av_log(NULL, AV_LOG_INFO, opt_name " is forwarded to lavfi " \
  567. "similarly to -af " filter_name "=%s.\n", arg); \
  568. \
  569. snprintf(name, sizeof(name), "graph %d %s for input stream %d:%d", \
  570. fg->index, filter_name, ist->file_index, ist->st->index); \
  571. ret = avfilter_graph_create_filter(&filt_ctx, \
  572. avfilter_get_by_name(filter_name), \
  573. name, arg, NULL, fg->graph); \
  574. if (ret < 0) \
  575. return ret; \
  576. \
  577. ret = avfilter_link(filt_ctx, 0, first_filter, pad_idx); \
  578. if (ret < 0) \
  579. return ret; \
  580. \
  581. first_filter = filt_ctx; \
  582. } while (0)
  583. if (audio_sync_method > 0) {
  584. char args[256] = {0};
  585. av_strlcatf(args, sizeof(args), "async=%d", audio_sync_method);
  586. if (audio_drift_threshold != 0.1)
  587. av_strlcatf(args, sizeof(args), ":min_hard_comp=%f", audio_drift_threshold);
  588. if (!fg->reconfiguration)
  589. av_strlcatf(args, sizeof(args), ":first_pts=0");
  590. AUTO_INSERT_FILTER_INPUT("-async", "aresample", args);
  591. }
  592. // if (ost->audio_channels_mapped) {
  593. // int i;
  594. // AVBPrint pan_buf;
  595. // av_bprint_init(&pan_buf, 256, 8192);
  596. // av_bprintf(&pan_buf, "0x%"PRIx64,
  597. // av_get_default_channel_layout(ost->audio_channels_mapped));
  598. // for (i = 0; i < ost->audio_channels_mapped; i++)
  599. // if (ost->audio_channels_map[i] != -1)
  600. // av_bprintf(&pan_buf, ":c%d=c%d", i, ost->audio_channels_map[i]);
  601. // AUTO_INSERT_FILTER_INPUT("-map_channel", "pan", pan_buf.str);
  602. // av_bprint_finalize(&pan_buf, NULL);
  603. // }
  604. if (audio_volume != 256) {
  605. char args[256];
  606. av_log(NULL, AV_LOG_WARNING, "-vol has been deprecated. Use the volume "
  607. "audio filter instead.\n");
  608. snprintf(args, sizeof(args), "%f", audio_volume / 256.);
  609. AUTO_INSERT_FILTER_INPUT("-vol", "volume", args);
  610. }
  611. if ((ret = avfilter_link(ifilter->filter, 0, first_filter, pad_idx)) < 0)
  612. return ret;
  613. return 0;
  614. }
  615. static int configure_input_filter(FilterGraph *fg, InputFilter *ifilter,
  616. AVFilterInOut *in)
  617. {
  618. av_freep(&ifilter->name);
  619. DESCRIBE_FILTER_LINK(ifilter, in, 1);
  620. switch (avfilter_pad_get_type(in->filter_ctx->input_pads, in->pad_idx)) {
  621. case AVMEDIA_TYPE_VIDEO: return configure_input_video_filter(fg, ifilter, in);
  622. case AVMEDIA_TYPE_AUDIO: return configure_input_audio_filter(fg, ifilter, in);
  623. default: av_assert0(0);
  624. }
  625. }
  626. int configure_filtergraph(FilterGraph *fg)
  627. {
  628. AVFilterInOut *inputs, *outputs, *cur;
  629. int ret, i, init = !fg->graph, simple = !fg->graph_desc;
  630. const char *graph_desc = simple ? fg->outputs[0]->ost->avfilter :
  631. fg->graph_desc;
  632. avfilter_graph_free(&fg->graph);
  633. if (!(fg->graph = avfilter_graph_alloc()))
  634. return AVERROR(ENOMEM);
  635. if (simple) {
  636. OutputStream *ost = fg->outputs[0]->ost;
  637. char args[512];
  638. AVDictionaryEntry *e = NULL;
  639. snprintf(args, sizeof(args), "flags=0x%X", (unsigned)ost->sws_flags);
  640. fg->graph->scale_sws_opts = av_strdup(args);
  641. args[0] = 0;
  642. while ((e = av_dict_get(ost->swr_opts, "", e,
  643. AV_DICT_IGNORE_SUFFIX))) {
  644. av_strlcatf(args, sizeof(args), "%s=%s:", e->key, e->value);
  645. }
  646. if (strlen(args))
  647. args[strlen(args)-1] = 0;
  648. av_opt_set(fg->graph, "aresample_swr_opts", args, 0);
  649. args[0] = '\0';
  650. while ((e = av_dict_get(fg->outputs[0]->ost->resample_opts, "", e,
  651. AV_DICT_IGNORE_SUFFIX))) {
  652. av_strlcatf(args, sizeof(args), "%s=%s:", e->key, e->value);
  653. }
  654. if (strlen(args))
  655. args[strlen(args) - 1] = '\0';
  656. fg->graph->resample_lavr_opts = av_strdup(args);
  657. }
  658. if ((ret = avfilter_graph_parse2(fg->graph, graph_desc, &inputs, &outputs)) < 0)
  659. return ret;
  660. if (simple && (!inputs || inputs->next || !outputs || outputs->next)) {
  661. av_log(NULL, AV_LOG_ERROR, "Simple filtergraph '%s' does not have "
  662. "exactly one input and output.\n", graph_desc);
  663. return AVERROR(EINVAL);
  664. }
  665. for (cur = inputs; !simple && init && cur; cur = cur->next)
  666. init_input_filter(fg, cur);
  667. for (cur = inputs, i = 0; cur; cur = cur->next, i++)
  668. if ((ret = configure_input_filter(fg, fg->inputs[i], cur)) < 0)
  669. return ret;
  670. avfilter_inout_free(&inputs);
  671. if (!init || simple) {
  672. /* we already know the mappings between lavfi outputs and output streams,
  673. * so we can finish the setup */
  674. for (cur = outputs, i = 0; cur; cur = cur->next, i++)
  675. configure_output_filter(fg, fg->outputs[i], cur);
  676. avfilter_inout_free(&outputs);
  677. if ((ret = avfilter_graph_config(fg->graph, NULL)) < 0)
  678. return ret;
  679. } else {
  680. /* wait until output mappings are processed */
  681. for (cur = outputs; cur;) {
  682. GROW_ARRAY(fg->outputs, fg->nb_outputs);
  683. if (!(fg->outputs[fg->nb_outputs - 1] = av_mallocz(sizeof(*fg->outputs[0]))))
  684. exit(1);
  685. fg->outputs[fg->nb_outputs - 1]->graph = fg;
  686. fg->outputs[fg->nb_outputs - 1]->out_tmp = cur;
  687. cur = cur->next;
  688. fg->outputs[fg->nb_outputs - 1]->out_tmp->next = NULL;
  689. }
  690. }
  691. fg->reconfiguration = 1;
  692. return 0;
  693. }
  694. int ist_in_filtergraph(FilterGraph *fg, InputStream *ist)
  695. {
  696. int i;
  697. for (i = 0; i < fg->nb_inputs; i++)
  698. if (fg->inputs[i]->ist == ist)
  699. return 1;
  700. return 0;
  701. }