You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

549 lines
18KB

  1. /*
  2. * Copyright (c) 2008 Vitor Sessak
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. /**
  21. * @file
  22. * memory buffer source filter
  23. */
  24. #include <float.h>
  25. #include "libavutil/channel_layout.h"
  26. #include "libavutil/common.h"
  27. #include "libavutil/fifo.h"
  28. #include "libavutil/frame.h"
  29. #include "libavutil/imgutils.h"
  30. #include "libavutil/internal.h"
  31. #include "libavutil/opt.h"
  32. #include "libavutil/samplefmt.h"
  33. #include "audio.h"
  34. #include "avfilter.h"
  35. #include "buffersrc.h"
  36. #include "formats.h"
  37. #include "internal.h"
  38. #include "video.h"
  39. #include "avcodec.h"
  40. typedef struct {
  41. const AVClass *class;
  42. AVFifoBuffer *fifo;
  43. AVRational time_base; ///< time_base to set in the output link
  44. AVRational frame_rate; ///< frame_rate to set in the output link
  45. unsigned nb_failed_requests;
  46. unsigned warning_limit;
  47. /* video only */
  48. int w, h;
  49. enum AVPixelFormat pix_fmt;
  50. AVRational pixel_aspect;
  51. char *sws_param;
  52. /* audio only */
  53. int sample_rate;
  54. enum AVSampleFormat sample_fmt;
  55. int channels;
  56. uint64_t channel_layout;
  57. char *channel_layout_str;
  58. int eof;
  59. } BufferSourceContext;
  60. #define CHECK_VIDEO_PARAM_CHANGE(s, c, width, height, format)\
  61. if (c->w != width || c->h != height || c->pix_fmt != format) {\
  62. av_log(s, AV_LOG_INFO, "Changing frame properties on the fly is not supported by all filters.\n");\
  63. }
  64. #define CHECK_AUDIO_PARAM_CHANGE(s, c, srate, ch_layout, ch_count, format)\
  65. if (c->sample_fmt != format || c->sample_rate != srate ||\
  66. c->channel_layout != ch_layout || c->channels != ch_count) {\
  67. av_log(s, AV_LOG_ERROR, "Changing frame properties on the fly is not supported.\n");\
  68. return AVERROR(EINVAL);\
  69. }
  70. int attribute_align_arg av_buffersrc_write_frame(AVFilterContext *ctx, const AVFrame *frame)
  71. {
  72. return av_buffersrc_add_frame_flags(ctx, (AVFrame *)frame,
  73. AV_BUFFERSRC_FLAG_KEEP_REF);
  74. }
  75. int attribute_align_arg av_buffersrc_add_frame(AVFilterContext *ctx, AVFrame *frame)
  76. {
  77. return av_buffersrc_add_frame_flags(ctx, frame, 0);
  78. }
  79. static int av_buffersrc_add_frame_internal(AVFilterContext *ctx,
  80. AVFrame *frame, int flags);
  81. int attribute_align_arg av_buffersrc_add_frame_flags(AVFilterContext *ctx, AVFrame *frame, int flags)
  82. {
  83. AVFrame *copy = NULL;
  84. int ret = 0;
  85. if (frame && frame->channel_layout &&
  86. av_get_channel_layout_nb_channels(frame->channel_layout) != av_frame_get_channels(frame)) {
  87. av_log(0, AV_LOG_ERROR, "Layout indicates a different number of channels than actually present\n");
  88. return AVERROR(EINVAL);
  89. }
  90. if (!(flags & AV_BUFFERSRC_FLAG_KEEP_REF) || !frame)
  91. return av_buffersrc_add_frame_internal(ctx, frame, flags);
  92. if (!(copy = av_frame_alloc()))
  93. return AVERROR(ENOMEM);
  94. ret = av_frame_ref(copy, frame);
  95. if (ret >= 0)
  96. ret = av_buffersrc_add_frame_internal(ctx, copy, flags);
  97. av_frame_free(&copy);
  98. return ret;
  99. }
  100. static int av_buffersrc_add_frame_internal(AVFilterContext *ctx,
  101. AVFrame *frame, int flags)
  102. {
  103. BufferSourceContext *s = ctx->priv;
  104. AVFrame *copy;
  105. int ret;
  106. s->nb_failed_requests = 0;
  107. if (!frame) {
  108. s->eof = 1;
  109. return 0;
  110. } else if (s->eof)
  111. return AVERROR(EINVAL);
  112. if (!(flags & AV_BUFFERSRC_FLAG_NO_CHECK_FORMAT)) {
  113. switch (ctx->outputs[0]->type) {
  114. case AVMEDIA_TYPE_VIDEO:
  115. CHECK_VIDEO_PARAM_CHANGE(ctx, s, frame->width, frame->height,
  116. frame->format);
  117. break;
  118. case AVMEDIA_TYPE_AUDIO:
  119. /* For layouts unknown on input but known on link after negotiation. */
  120. if (!frame->channel_layout)
  121. frame->channel_layout = s->channel_layout;
  122. CHECK_AUDIO_PARAM_CHANGE(ctx, s, frame->sample_rate, frame->channel_layout,
  123. av_frame_get_channels(frame), frame->format);
  124. break;
  125. default:
  126. return AVERROR(EINVAL);
  127. }
  128. }
  129. if (!av_fifo_space(s->fifo) &&
  130. (ret = av_fifo_realloc2(s->fifo, av_fifo_size(s->fifo) +
  131. sizeof(copy))) < 0)
  132. return ret;
  133. if (!(copy = av_frame_alloc()))
  134. return AVERROR(ENOMEM);
  135. av_frame_move_ref(copy, frame);
  136. if ((ret = av_fifo_generic_write(s->fifo, &copy, sizeof(copy), NULL)) < 0) {
  137. av_frame_move_ref(frame, copy);
  138. av_frame_free(&copy);
  139. return ret;
  140. }
  141. if ((flags & AV_BUFFERSRC_FLAG_PUSH))
  142. if ((ret = ctx->output_pads[0].request_frame(ctx->outputs[0])) < 0)
  143. return ret;
  144. return 0;
  145. }
  146. #if FF_API_AVFILTERBUFFER
  147. FF_DISABLE_DEPRECATION_WARNINGS
  148. static void compat_free_buffer(void *opaque, uint8_t *data)
  149. {
  150. AVFilterBufferRef *buf = opaque;
  151. AV_NOWARN_DEPRECATED(
  152. avfilter_unref_buffer(buf);
  153. )
  154. }
  155. static void compat_unref_buffer(void *opaque, uint8_t *data)
  156. {
  157. AVBufferRef *buf = opaque;
  158. AV_NOWARN_DEPRECATED(
  159. av_buffer_unref(&buf);
  160. )
  161. }
  162. int av_buffersrc_add_ref(AVFilterContext *ctx, AVFilterBufferRef *buf,
  163. int flags)
  164. {
  165. BufferSourceContext *s = ctx->priv;
  166. AVFrame *frame = NULL;
  167. AVBufferRef *dummy_buf = NULL;
  168. int ret = 0, planes, i;
  169. if (!buf) {
  170. s->eof = 1;
  171. return 0;
  172. } else if (s->eof)
  173. return AVERROR(EINVAL);
  174. frame = av_frame_alloc();
  175. if (!frame)
  176. return AVERROR(ENOMEM);
  177. dummy_buf = av_buffer_create(NULL, 0, compat_free_buffer, buf,
  178. (buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY);
  179. if (!dummy_buf) {
  180. ret = AVERROR(ENOMEM);
  181. goto fail;
  182. }
  183. AV_NOWARN_DEPRECATED(
  184. if ((ret = avfilter_copy_buf_props(frame, buf)) < 0)
  185. goto fail;
  186. )
  187. #define WRAP_PLANE(ref_out, data, data_size) \
  188. do { \
  189. AVBufferRef *dummy_ref = av_buffer_ref(dummy_buf); \
  190. if (!dummy_ref) { \
  191. ret = AVERROR(ENOMEM); \
  192. goto fail; \
  193. } \
  194. ref_out = av_buffer_create(data, data_size, compat_unref_buffer, \
  195. dummy_ref, (buf->perms & AV_PERM_WRITE) ? 0 : AV_BUFFER_FLAG_READONLY); \
  196. if (!ref_out) { \
  197. av_frame_unref(frame); \
  198. ret = AVERROR(ENOMEM); \
  199. goto fail; \
  200. } \
  201. } while (0)
  202. if (ctx->outputs[0]->type == AVMEDIA_TYPE_VIDEO) {
  203. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
  204. planes = av_pix_fmt_count_planes(frame->format);
  205. if (!desc || planes <= 0) {
  206. ret = AVERROR(EINVAL);
  207. goto fail;
  208. }
  209. for (i = 0; i < planes; i++) {
  210. int v_shift = (i == 1 || i == 2) ? desc->log2_chroma_h : 0;
  211. int plane_size = (frame->height >> v_shift) * frame->linesize[i];
  212. WRAP_PLANE(frame->buf[i], frame->data[i], plane_size);
  213. }
  214. } else {
  215. int planar = av_sample_fmt_is_planar(frame->format);
  216. int channels = av_get_channel_layout_nb_channels(frame->channel_layout);
  217. planes = planar ? channels : 1;
  218. if (planes > FF_ARRAY_ELEMS(frame->buf)) {
  219. frame->nb_extended_buf = planes - FF_ARRAY_ELEMS(frame->buf);
  220. frame->extended_buf = av_mallocz(sizeof(*frame->extended_buf) *
  221. frame->nb_extended_buf);
  222. if (!frame->extended_buf) {
  223. ret = AVERROR(ENOMEM);
  224. goto fail;
  225. }
  226. }
  227. for (i = 0; i < FFMIN(planes, FF_ARRAY_ELEMS(frame->buf)); i++)
  228. WRAP_PLANE(frame->buf[i], frame->extended_data[i], frame->linesize[0]);
  229. for (i = 0; i < planes - FF_ARRAY_ELEMS(frame->buf); i++)
  230. WRAP_PLANE(frame->extended_buf[i],
  231. frame->extended_data[i + FF_ARRAY_ELEMS(frame->buf)],
  232. frame->linesize[0]);
  233. }
  234. ret = av_buffersrc_add_frame_flags(ctx, frame, flags);
  235. fail:
  236. av_buffer_unref(&dummy_buf);
  237. av_frame_free(&frame);
  238. return ret;
  239. }
  240. FF_ENABLE_DEPRECATION_WARNINGS
  241. int av_buffersrc_buffer(AVFilterContext *ctx, AVFilterBufferRef *buf)
  242. {
  243. return av_buffersrc_add_ref(ctx, buf, 0);
  244. }
  245. #endif
  246. static av_cold int init_video(AVFilterContext *ctx)
  247. {
  248. BufferSourceContext *c = ctx->priv;
  249. if (c->pix_fmt == AV_PIX_FMT_NONE || !c->w || !c->h || av_q2d(c->time_base) <= 0) {
  250. av_log(ctx, AV_LOG_ERROR, "Invalid parameters provided.\n");
  251. return AVERROR(EINVAL);
  252. }
  253. if (!(c->fifo = av_fifo_alloc(sizeof(AVFrame*))))
  254. return AVERROR(ENOMEM);
  255. av_log(ctx, AV_LOG_VERBOSE, "w:%d h:%d pixfmt:%s tb:%d/%d fr:%d/%d sar:%d/%d sws_param:%s\n",
  256. c->w, c->h, av_get_pix_fmt_name(c->pix_fmt),
  257. c->time_base.num, c->time_base.den, c->frame_rate.num, c->frame_rate.den,
  258. c->pixel_aspect.num, c->pixel_aspect.den, (char *)av_x_if_null(c->sws_param, ""));
  259. c->warning_limit = 100;
  260. return 0;
  261. }
  262. unsigned av_buffersrc_get_nb_failed_requests(AVFilterContext *buffer_src)
  263. {
  264. return ((BufferSourceContext *)buffer_src->priv)->nb_failed_requests;
  265. }
  266. #define OFFSET(x) offsetof(BufferSourceContext, x)
  267. #define A AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_AUDIO_PARAM
  268. #define V AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
  269. static const AVOption buffer_options[] = {
  270. { "width", NULL, OFFSET(w), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  271. { "video_size", NULL, OFFSET(w), AV_OPT_TYPE_IMAGE_SIZE, .flags = V },
  272. { "height", NULL, OFFSET(h), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  273. { "pix_fmt", NULL, OFFSET(pix_fmt), AV_OPT_TYPE_PIXEL_FMT, { .i64 = AV_PIX_FMT_NONE }, .min = AV_PIX_FMT_NONE, .max = INT_MAX, .flags = V },
  274. #if FF_API_OLD_FILTER_OPTS
  275. /* those 4 are for compatibility with the old option passing system where each filter
  276. * did its own parsing */
  277. { "time_base_num", "deprecated, do not use", OFFSET(time_base.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  278. { "time_base_den", "deprecated, do not use", OFFSET(time_base.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  279. { "sar_num", "deprecated, do not use", OFFSET(pixel_aspect.num), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  280. { "sar_den", "deprecated, do not use", OFFSET(pixel_aspect.den), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, V },
  281. #endif
  282. { "sar", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V },
  283. { "pixel_aspect", "sample aspect ratio", OFFSET(pixel_aspect), AV_OPT_TYPE_RATIONAL, { .dbl = 1 }, 0, DBL_MAX, V },
  284. { "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V },
  285. { "frame_rate", NULL, OFFSET(frame_rate), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, DBL_MAX, V },
  286. { "sws_param", NULL, OFFSET(sws_param), AV_OPT_TYPE_STRING, .flags = V },
  287. { NULL },
  288. };
  289. AVFILTER_DEFINE_CLASS(buffer);
  290. static const AVOption abuffer_options[] = {
  291. { "time_base", NULL, OFFSET(time_base), AV_OPT_TYPE_RATIONAL, { .dbl = 0 }, 0, INT_MAX, A },
  292. { "sample_rate", NULL, OFFSET(sample_rate), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A },
  293. { "sample_fmt", NULL, OFFSET(sample_fmt), AV_OPT_TYPE_SAMPLE_FMT, { .i64 = AV_SAMPLE_FMT_NONE }, .min = AV_SAMPLE_FMT_NONE, .max = INT_MAX, .flags = A },
  294. { "channel_layout", NULL, OFFSET(channel_layout_str), AV_OPT_TYPE_STRING, .flags = A },
  295. { "channels", NULL, OFFSET(channels), AV_OPT_TYPE_INT, { .i64 = 0 }, 0, INT_MAX, A },
  296. { NULL },
  297. };
  298. AVFILTER_DEFINE_CLASS(abuffer);
  299. static av_cold int init_audio(AVFilterContext *ctx)
  300. {
  301. BufferSourceContext *s = ctx->priv;
  302. int ret = 0;
  303. if (s->sample_fmt == AV_SAMPLE_FMT_NONE) {
  304. av_log(ctx, AV_LOG_ERROR, "Sample format was not set or was invalid\n");
  305. return AVERROR(EINVAL);
  306. }
  307. if (s->channel_layout_str) {
  308. int n;
  309. /* TODO reindent */
  310. s->channel_layout = av_get_channel_layout(s->channel_layout_str);
  311. if (!s->channel_layout) {
  312. av_log(ctx, AV_LOG_ERROR, "Invalid channel layout %s.\n",
  313. s->channel_layout_str);
  314. return AVERROR(EINVAL);
  315. }
  316. n = av_get_channel_layout_nb_channels(s->channel_layout);
  317. if (s->channels) {
  318. if (n != s->channels) {
  319. av_log(ctx, AV_LOG_ERROR,
  320. "Mismatching channel count %d and layout '%s' "
  321. "(%d channels)\n",
  322. s->channels, s->channel_layout_str, n);
  323. return AVERROR(EINVAL);
  324. }
  325. }
  326. s->channels = n;
  327. } else if (!s->channels) {
  328. av_log(ctx, AV_LOG_ERROR, "Neither number of channels nor "
  329. "channel layout specified\n");
  330. return AVERROR(EINVAL);
  331. }
  332. if (!(s->fifo = av_fifo_alloc(sizeof(AVFrame*))))
  333. return AVERROR(ENOMEM);
  334. if (!s->time_base.num)
  335. s->time_base = (AVRational){1, s->sample_rate};
  336. av_log(ctx, AV_LOG_VERBOSE,
  337. "tb:%d/%d samplefmt:%s samplerate:%d chlayout:%s\n",
  338. s->time_base.num, s->time_base.den, av_get_sample_fmt_name(s->sample_fmt),
  339. s->sample_rate, s->channel_layout_str);
  340. s->warning_limit = 100;
  341. return ret;
  342. }
  343. static av_cold void uninit(AVFilterContext *ctx)
  344. {
  345. BufferSourceContext *s = ctx->priv;
  346. while (s->fifo && av_fifo_size(s->fifo)) {
  347. AVFrame *frame;
  348. av_fifo_generic_read(s->fifo, &frame, sizeof(frame), NULL);
  349. av_frame_free(&frame);
  350. }
  351. av_fifo_free(s->fifo);
  352. s->fifo = NULL;
  353. }
  354. static int query_formats(AVFilterContext *ctx)
  355. {
  356. BufferSourceContext *c = ctx->priv;
  357. AVFilterChannelLayouts *channel_layouts = NULL;
  358. AVFilterFormats *formats = NULL;
  359. AVFilterFormats *samplerates = NULL;
  360. switch (ctx->outputs[0]->type) {
  361. case AVMEDIA_TYPE_VIDEO:
  362. ff_add_format(&formats, c->pix_fmt);
  363. ff_set_common_formats(ctx, formats);
  364. break;
  365. case AVMEDIA_TYPE_AUDIO:
  366. ff_add_format(&formats, c->sample_fmt);
  367. ff_set_common_formats(ctx, formats);
  368. ff_add_format(&samplerates, c->sample_rate);
  369. ff_set_common_samplerates(ctx, samplerates);
  370. ff_add_channel_layout(&channel_layouts,
  371. c->channel_layout ? c->channel_layout :
  372. FF_COUNT2LAYOUT(c->channels));
  373. ff_set_common_channel_layouts(ctx, channel_layouts);
  374. break;
  375. default:
  376. return AVERROR(EINVAL);
  377. }
  378. return 0;
  379. }
  380. static int config_props(AVFilterLink *link)
  381. {
  382. BufferSourceContext *c = link->src->priv;
  383. switch (link->type) {
  384. case AVMEDIA_TYPE_VIDEO:
  385. link->w = c->w;
  386. link->h = c->h;
  387. link->sample_aspect_ratio = c->pixel_aspect;
  388. break;
  389. case AVMEDIA_TYPE_AUDIO:
  390. if (!c->channel_layout)
  391. c->channel_layout = link->channel_layout;
  392. break;
  393. default:
  394. return AVERROR(EINVAL);
  395. }
  396. link->time_base = c->time_base;
  397. link->frame_rate = c->frame_rate;
  398. return 0;
  399. }
  400. static int request_frame(AVFilterLink *link)
  401. {
  402. BufferSourceContext *c = link->src->priv;
  403. AVFrame *frame;
  404. if (!av_fifo_size(c->fifo)) {
  405. if (c->eof)
  406. return AVERROR_EOF;
  407. c->nb_failed_requests++;
  408. return AVERROR(EAGAIN);
  409. }
  410. av_fifo_generic_read(c->fifo, &frame, sizeof(frame), NULL);
  411. return ff_filter_frame(link, frame);
  412. }
  413. static int poll_frame(AVFilterLink *link)
  414. {
  415. BufferSourceContext *c = link->src->priv;
  416. int size = av_fifo_size(c->fifo);
  417. if (!size && c->eof)
  418. return AVERROR_EOF;
  419. return size/sizeof(AVFrame*);
  420. }
  421. static const AVFilterPad avfilter_vsrc_buffer_outputs[] = {
  422. {
  423. .name = "default",
  424. .type = AVMEDIA_TYPE_VIDEO,
  425. .request_frame = request_frame,
  426. .poll_frame = poll_frame,
  427. .config_props = config_props,
  428. },
  429. { NULL }
  430. };
  431. AVFilter ff_vsrc_buffer = {
  432. .name = "buffer",
  433. .description = NULL_IF_CONFIG_SMALL("Buffer video frames, and make them accessible to the filterchain."),
  434. .priv_size = sizeof(BufferSourceContext),
  435. .query_formats = query_formats,
  436. .init = init_video,
  437. .uninit = uninit,
  438. .inputs = NULL,
  439. .outputs = avfilter_vsrc_buffer_outputs,
  440. .priv_class = &buffer_class,
  441. };
  442. static const AVFilterPad avfilter_asrc_abuffer_outputs[] = {
  443. {
  444. .name = "default",
  445. .type = AVMEDIA_TYPE_AUDIO,
  446. .request_frame = request_frame,
  447. .poll_frame = poll_frame,
  448. .config_props = config_props,
  449. },
  450. { NULL }
  451. };
  452. AVFilter ff_asrc_abuffer = {
  453. .name = "abuffer",
  454. .description = NULL_IF_CONFIG_SMALL("Buffer audio frames, and make them accessible to the filterchain."),
  455. .priv_size = sizeof(BufferSourceContext),
  456. .query_formats = query_formats,
  457. .init = init_audio,
  458. .uninit = uninit,
  459. .inputs = NULL,
  460. .outputs = avfilter_asrc_abuffer_outputs,
  461. .priv_class = &abuffer_class,
  462. };