You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

468 lines
19KB

  1. /*
  2. * Copyright (c) 2013 Paul B Mahol
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. #include "libavutil/imgutils.h"
  21. #include "libavutil/eval.h"
  22. #include "libavutil/opt.h"
  23. #include "libavutil/pixfmt.h"
  24. #include "avfilter.h"
  25. #include "bufferqueue.h"
  26. #include "formats.h"
  27. #include "internal.h"
  28. #include "video.h"
  29. #define TOP 0
  30. #define BOTTOM 1
  31. enum BlendMode {
  32. BLEND_UNSET = -1,
  33. BLEND_NORMAL,
  34. BLEND_ADDITION,
  35. BLEND_AND,
  36. BLEND_AVERAGE,
  37. BLEND_BURN,
  38. BLEND_DARKEN,
  39. BLEND_DIFFERENCE,
  40. BLEND_DIVIDE,
  41. BLEND_DODGE,
  42. BLEND_EXCLUSION,
  43. BLEND_HARDLIGHT,
  44. BLEND_LIGHTEN,
  45. BLEND_MULTIPLY,
  46. BLEND_NEGATION,
  47. BLEND_OR,
  48. BLEND_OVERLAY,
  49. BLEND_PHOENIX,
  50. BLEND_PINLIGHT,
  51. BLEND_REFLECT,
  52. BLEND_SCREEN,
  53. BLEND_SOFTLIGHT,
  54. BLEND_SUBTRACT,
  55. BLEND_VIVIDLIGHT,
  56. BLEND_XOR,
  57. BLEND_NB
  58. };
  59. static const char *const var_names[] = { "X", "Y", "W", "H", "SW", "SH", "T", "N", "A", "B", "TOP", "BOTTOM", NULL };
  60. enum { VAR_X, VAR_Y, VAR_W, VAR_H, VAR_SW, VAR_SH, VAR_T, VAR_N, VAR_A, VAR_B, VAR_TOP, VAR_BOTTOM, VAR_VARS_NB };
  61. typedef struct FilterParams {
  62. enum BlendMode mode;
  63. double values[VAR_VARS_NB];
  64. double opacity;
  65. AVExpr *e;
  66. char *expr_str;
  67. void (*blend)(const uint8_t *top, int top_linesize,
  68. const uint8_t *bottom, int bottom_linesize,
  69. uint8_t *dst, int dst_linesize,
  70. int width, int height, struct FilterParams *param);
  71. } FilterParams;
  72. typedef struct {
  73. const AVClass *class;
  74. struct FFBufQueue queue_top;
  75. struct FFBufQueue queue_bottom;
  76. int hsub, vsub; ///< chroma subsampling values
  77. int frame_requested;
  78. int framenum;
  79. char *all_expr;
  80. enum BlendMode all_mode;
  81. double all_opacity;
  82. FilterParams params[4];
  83. } BlendContext;
  84. #define OFFSET(x) offsetof(BlendContext, x)
  85. #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
  86. static const AVOption blend_options[] = {
  87. { "c0_mode", "set component #0 blend mode", OFFSET(params[0].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
  88. { "c1_mode", "set component #1 blend mode", OFFSET(params[1].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
  89. { "c2_mode", "set component #2 blend mode", OFFSET(params[2].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
  90. { "c3_mode", "set component #3 blend mode", OFFSET(params[3].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},
  91. { "all_mode", "set blend mode for all components", OFFSET(all_mode), AV_OPT_TYPE_INT, {.i64=-1},-1, BLEND_NB-1, FLAGS, "mode"},
  92. { "addition", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_ADDITION}, 0, 0, FLAGS, "mode" },
  93. { "and", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AND}, 0, 0, FLAGS, "mode" },
  94. { "average", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AVERAGE}, 0, 0, FLAGS, "mode" },
  95. { "burn", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_BURN}, 0, 0, FLAGS, "mode" },
  96. { "darken", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DARKEN}, 0, 0, FLAGS, "mode" },
  97. { "difference", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIFFERENCE}, 0, 0, FLAGS, "mode" },
  98. { "divide", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIVIDE}, 0, 0, FLAGS, "mode" },
  99. { "dodge", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DODGE}, 0, 0, FLAGS, "mode" },
  100. { "exclusion", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_EXCLUSION}, 0, 0, FLAGS, "mode" },
  101. { "hardlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_HARDLIGHT}, 0, 0, FLAGS, "mode" },
  102. { "lighten", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_LIGHTEN}, 0, 0, FLAGS, "mode" },
  103. { "multiply", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_MULTIPLY}, 0, 0, FLAGS, "mode" },
  104. { "negation", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NEGATION}, 0, 0, FLAGS, "mode" },
  105. { "normal", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NORMAL}, 0, 0, FLAGS, "mode" },
  106. { "or", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OR}, 0, 0, FLAGS, "mode" },
  107. { "overlay", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OVERLAY}, 0, 0, FLAGS, "mode" },
  108. { "phoenix", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PHOENIX}, 0, 0, FLAGS, "mode" },
  109. { "pinlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PINLIGHT}, 0, 0, FLAGS, "mode" },
  110. { "reflect", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_REFLECT}, 0, 0, FLAGS, "mode" },
  111. { "screen", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SCREEN}, 0, 0, FLAGS, "mode" },
  112. { "softlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SOFTLIGHT}, 0, 0, FLAGS, "mode" },
  113. { "subtract", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SUBTRACT}, 0, 0, FLAGS, "mode" },
  114. { "vividlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_VIVIDLIGHT}, 0, 0, FLAGS, "mode" },
  115. { "xor", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_XOR}, 0, 0, FLAGS, "mode" },
  116. { "c0_expr", "set color component #0 expression", OFFSET(params[0].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
  117. { "c1_expr", "set color component #1 expression", OFFSET(params[1].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
  118. { "c2_expr", "set color component #2 expression", OFFSET(params[2].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
  119. { "c3_expr", "set color component #3 expression", OFFSET(params[3].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
  120. { "all_expr", "set expression for all color components", OFFSET(all_expr), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },
  121. { "c0_opacity", "set color component #0 opacity", OFFSET(params[0].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
  122. { "c1_opacity", "set color component #1 opacity", OFFSET(params[1].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
  123. { "c2_opacity", "set color component #2 opacity", OFFSET(params[2].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
  124. { "c3_opacity", "set color component #3 opacity", OFFSET(params[3].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },
  125. { "all_opacity", "set opacity for all color components", OFFSET(all_opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS},
  126. { NULL },
  127. };
  128. AVFILTER_DEFINE_CLASS(blend);
  129. static void blend_normal(const uint8_t *top, int top_linesize,
  130. const uint8_t *bottom, int bottom_linesize,
  131. uint8_t *dst, int dst_linesize,
  132. int width, int height, FilterParams *param)
  133. {
  134. av_image_copy_plane(dst, dst_linesize, top, top_linesize, width, height);
  135. }
  136. #define DEFINE_BLEND(name, expr) \
  137. static void blend_## name(const uint8_t *top, int top_linesize, \
  138. const uint8_t *bottom, int bottom_linesize, \
  139. uint8_t *dst, int dst_linesize, \
  140. int width, int height, FilterParams *param) \
  141. { \
  142. double opacity = param->opacity; \
  143. int i, j; \
  144. \
  145. for (i = 0; i < height; i++) { \
  146. for (j = 0; j < width; j++) { \
  147. dst[j] = top[j] + ((expr) - top[j]) * opacity; \
  148. } \
  149. dst += dst_linesize; \
  150. top += top_linesize; \
  151. bottom += bottom_linesize; \
  152. } \
  153. }
  154. #define A top[j]
  155. #define B bottom[j]
  156. #define MULTIPLY(x, a, b) (x * ((a * b) / 255))
  157. #define SCREEN(x, a, b) (255 - x * ((255 - a) * (255 - b) / 255))
  158. #define BURN(a, b) ((a == 0) ? a : FFMAX(0, 255 - ((255 - b) << 8) / a))
  159. #define DODGE(a, b) ((a == 255) ? a : FFMIN(255, ((b << 8) / (255 - a))))
  160. DEFINE_BLEND(addition, FFMIN(255, A + B))
  161. DEFINE_BLEND(average, (A + B) / 2)
  162. DEFINE_BLEND(subtract, FFMAX(0, A - B))
  163. DEFINE_BLEND(multiply, MULTIPLY(1, A, B))
  164. DEFINE_BLEND(negation, 255 - FFABS(255 - A - B))
  165. DEFINE_BLEND(difference, FFABS(A - B))
  166. DEFINE_BLEND(screen, SCREEN(1, A, B))
  167. DEFINE_BLEND(overlay, (A < 128) ? MULTIPLY(2, A, B) : SCREEN(2, A, B))
  168. DEFINE_BLEND(hardlight, (B < 128) ? MULTIPLY(2, B, A) : SCREEN(2, B, A))
  169. DEFINE_BLEND(darken, FFMIN(A, B))
  170. DEFINE_BLEND(lighten, FFMAX(A, B))
  171. DEFINE_BLEND(divide, ((float)A / ((float)B) * 255))
  172. DEFINE_BLEND(dodge, DODGE(A, B))
  173. DEFINE_BLEND(burn, BURN(A, B))
  174. DEFINE_BLEND(softlight, (A > 127) ? B + (255 - B) * (A - 127.5) / 127.5 * (0.5 - FFABS(B - 127.5) / 255): B - B * ((127.5 - A) / 127.5) * (0.5 - FFABS(B - 127.5)/255))
  175. DEFINE_BLEND(exclusion, A + B - 2 * A * B / 255)
  176. DEFINE_BLEND(pinlight, (B < 128) ? FFMIN(A, 2 * B) : FFMAX(A, 2 * (B - 128)))
  177. DEFINE_BLEND(phoenix, FFMIN(A, B) - FFMAX(A, B) + 255)
  178. DEFINE_BLEND(reflect, (B == 255) ? B : FFMIN(255, (A * A / (255 - B))))
  179. DEFINE_BLEND(and, A & B)
  180. DEFINE_BLEND(or, A | B)
  181. DEFINE_BLEND(xor, A ^ B)
  182. DEFINE_BLEND(vividlight, (B < 128) ? BURN(A, 2 * B) : DODGE(A, 2 * (B - 128)))
  183. static void blend_expr(const uint8_t *top, int top_linesize,
  184. const uint8_t *bottom, int bottom_linesize,
  185. uint8_t *dst, int dst_linesize,
  186. int width, int height,
  187. FilterParams *param)
  188. {
  189. AVExpr *e = param->e;
  190. double *values = param->values;
  191. int y, x;
  192. for (y = 0; y < height; y++) {
  193. values[VAR_Y] = y;
  194. for (x = 0; x < width; x++) {
  195. values[VAR_X] = x;
  196. values[VAR_TOP] = values[VAR_A] = top[x];
  197. values[VAR_BOTTOM] = values[VAR_B] = bottom[x];
  198. dst[x] = av_expr_eval(e, values, NULL);
  199. }
  200. dst += dst_linesize;
  201. top += top_linesize;
  202. bottom += bottom_linesize;
  203. }
  204. }
  205. static av_cold int init(AVFilterContext *ctx)
  206. {
  207. BlendContext *b = ctx->priv;
  208. int ret, plane;
  209. for (plane = 0; plane < FF_ARRAY_ELEMS(b->params); plane++) {
  210. FilterParams *param = &b->params[plane];
  211. if (b->all_mode >= 0)
  212. param->mode = b->all_mode;
  213. if (b->all_opacity < 1)
  214. param->opacity = b->all_opacity;
  215. switch (param->mode) {
  216. case BLEND_ADDITION: param->blend = blend_addition; break;
  217. case BLEND_AND: param->blend = blend_and; break;
  218. case BLEND_AVERAGE: param->blend = blend_average; break;
  219. case BLEND_BURN: param->blend = blend_burn; break;
  220. case BLEND_DARKEN: param->blend = blend_darken; break;
  221. case BLEND_DIFFERENCE: param->blend = blend_difference; break;
  222. case BLEND_DIVIDE: param->blend = blend_divide; break;
  223. case BLEND_DODGE: param->blend = blend_dodge; break;
  224. case BLEND_EXCLUSION: param->blend = blend_exclusion; break;
  225. case BLEND_HARDLIGHT: param->blend = blend_hardlight; break;
  226. case BLEND_LIGHTEN: param->blend = blend_lighten; break;
  227. case BLEND_MULTIPLY: param->blend = blend_multiply; break;
  228. case BLEND_NEGATION: param->blend = blend_negation; break;
  229. case BLEND_NORMAL: param->blend = blend_normal; break;
  230. case BLEND_OR: param->blend = blend_or; break;
  231. case BLEND_OVERLAY: param->blend = blend_overlay; break;
  232. case BLEND_PHOENIX: param->blend = blend_phoenix; break;
  233. case BLEND_PINLIGHT: param->blend = blend_pinlight; break;
  234. case BLEND_REFLECT: param->blend = blend_reflect; break;
  235. case BLEND_SCREEN: param->blend = blend_screen; break;
  236. case BLEND_SOFTLIGHT: param->blend = blend_softlight; break;
  237. case BLEND_SUBTRACT: param->blend = blend_subtract; break;
  238. case BLEND_VIVIDLIGHT: param->blend = blend_vividlight; break;
  239. case BLEND_XOR: param->blend = blend_xor; break;
  240. }
  241. if (b->all_expr && !param->expr_str) {
  242. param->expr_str = av_strdup(b->all_expr);
  243. if (!param->expr_str)
  244. return AVERROR(ENOMEM);
  245. }
  246. if (param->expr_str) {
  247. ret = av_expr_parse(&param->e, param->expr_str, var_names,
  248. NULL, NULL, NULL, NULL, 0, ctx);
  249. if (ret < 0)
  250. return ret;
  251. param->blend = blend_expr;
  252. }
  253. }
  254. return 0;
  255. }
  256. static int query_formats(AVFilterContext *ctx)
  257. {
  258. static const enum AVPixelFormat pix_fmts[] = {
  259. AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA420P,
  260. AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P,AV_PIX_FMT_YUVJ420P,
  261. AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV410P,
  262. AV_PIX_FMT_GBRP, AV_PIX_FMT_GRAY8, AV_PIX_FMT_NONE
  263. };
  264. ff_set_common_formats(ctx, ff_make_format_list(pix_fmts));
  265. return 0;
  266. }
  267. static int config_output(AVFilterLink *outlink)
  268. {
  269. AVFilterContext *ctx = outlink->src;
  270. AVFilterLink *toplink = ctx->inputs[TOP];
  271. AVFilterLink *bottomlink = ctx->inputs[BOTTOM];
  272. if (toplink->format != bottomlink->format) {
  273. av_log(ctx, AV_LOG_ERROR, "inputs must be of same pixel format\n");
  274. return AVERROR(EINVAL);
  275. }
  276. if (toplink->w != bottomlink->w ||
  277. toplink->h != bottomlink->h ||
  278. toplink->sample_aspect_ratio.num != bottomlink->sample_aspect_ratio.num ||
  279. toplink->sample_aspect_ratio.den != bottomlink->sample_aspect_ratio.den) {
  280. av_log(ctx, AV_LOG_ERROR, "First input link %s parameters "
  281. "(size %dx%d, SAR %d:%d) do not match the corresponding "
  282. "second input link %s parameters (%dx%d, SAR %d:%d)\n",
  283. ctx->input_pads[TOP].name, toplink->w, toplink->h,
  284. toplink->sample_aspect_ratio.num,
  285. toplink->sample_aspect_ratio.den,
  286. ctx->input_pads[BOTTOM].name, bottomlink->w, bottomlink->h,
  287. bottomlink->sample_aspect_ratio.num,
  288. bottomlink->sample_aspect_ratio.den);
  289. return AVERROR(EINVAL);
  290. }
  291. outlink->w = toplink->w;
  292. outlink->h = bottomlink->h;
  293. outlink->time_base = toplink->time_base;
  294. outlink->sample_aspect_ratio = toplink->sample_aspect_ratio;
  295. outlink->frame_rate = toplink->frame_rate;
  296. return 0;
  297. }
  298. static int config_input_top(AVFilterLink *inlink)
  299. {
  300. BlendContext *b = inlink->dst->priv;
  301. const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(inlink->format);
  302. b->hsub = pix_desc->log2_chroma_w;
  303. b->vsub = pix_desc->log2_chroma_h;
  304. return 0;
  305. }
  306. static av_cold void uninit(AVFilterContext *ctx)
  307. {
  308. BlendContext *b = ctx->priv;
  309. int i;
  310. ff_bufqueue_discard_all(&b->queue_top);
  311. ff_bufqueue_discard_all(&b->queue_bottom);
  312. for (i = 0; i < FF_ARRAY_ELEMS(b->params); i++)
  313. av_expr_free(b->params[i].e);
  314. }
  315. static int request_frame(AVFilterLink *outlink)
  316. {
  317. AVFilterContext *ctx = outlink->src;
  318. BlendContext *b = ctx->priv;
  319. int in, ret;
  320. b->frame_requested = 1;
  321. while (b->frame_requested) {
  322. in = ff_bufqueue_peek(&b->queue_top, 0) ? BOTTOM : TOP;
  323. ret = ff_request_frame(ctx->inputs[in]);
  324. if (ret < 0)
  325. return ret;
  326. }
  327. return 0;
  328. }
  329. static void blend_frame(AVFilterContext *ctx,
  330. AVFrame *top_buf,
  331. AVFrame *bottom_buf,
  332. AVFrame *dst_buf)
  333. {
  334. BlendContext *b = ctx->priv;
  335. AVFilterLink *inlink = ctx->inputs[0];
  336. FilterParams *param;
  337. int plane;
  338. for (plane = 0; dst_buf->data[plane]; plane++) {
  339. int hsub = plane == 1 || plane == 2 ? b->hsub : 0;
  340. int vsub = plane == 1 || plane == 2 ? b->vsub : 0;
  341. int outw = dst_buf->width >> hsub;
  342. int outh = dst_buf->height >> vsub;
  343. uint8_t *dst = dst_buf->data[plane];
  344. uint8_t *top = top_buf->data[plane];
  345. uint8_t *bottom = bottom_buf->data[plane];
  346. param = &b->params[plane];
  347. param->values[VAR_N] = b->framenum++;
  348. param->values[VAR_T] = dst_buf->pts == AV_NOPTS_VALUE ? NAN : dst_buf->pts * av_q2d(inlink->time_base);
  349. param->values[VAR_W] = outw;
  350. param->values[VAR_H] = outh;
  351. param->values[VAR_SW] = outw / dst_buf->width;
  352. param->values[VAR_SH] = outh / dst_buf->height;
  353. param->blend(top, top_buf->linesize[plane],
  354. bottom, bottom_buf->linesize[plane],
  355. dst, dst_buf->linesize[plane], outw, outh, param);
  356. }
  357. }
  358. static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
  359. {
  360. AVFilterContext *ctx = inlink->dst;
  361. AVFilterLink *outlink = ctx->outputs[0];
  362. BlendContext *b = ctx->priv;
  363. int ret = 0;
  364. int is_bottom = (inlink == ctx->inputs[BOTTOM]);
  365. struct FFBufQueue *queue =
  366. (is_bottom ? &b->queue_bottom : &b->queue_top);
  367. ff_bufqueue_add(ctx, queue, buf);
  368. while (1) {
  369. AVFrame *top_buf, *bottom_buf, *out_buf;
  370. if (!ff_bufqueue_peek(&b->queue_top, 0) ||
  371. !ff_bufqueue_peek(&b->queue_bottom, 0)) break;
  372. top_buf = ff_bufqueue_get(&b->queue_top);
  373. bottom_buf = ff_bufqueue_get(&b->queue_bottom);
  374. out_buf = ff_get_video_buffer(outlink, outlink->w, outlink->h);
  375. if (!out_buf) {
  376. return AVERROR(ENOMEM);
  377. }
  378. av_frame_copy_props(out_buf, top_buf);
  379. b->frame_requested = 0;
  380. blend_frame(ctx, top_buf, bottom_buf, out_buf);
  381. ret = ff_filter_frame(ctx->outputs[0], out_buf);
  382. av_frame_free(&top_buf);
  383. av_frame_free(&bottom_buf);
  384. }
  385. return ret;
  386. }
  387. static const AVFilterPad blend_inputs[] = {
  388. {
  389. .name = "top",
  390. .type = AVMEDIA_TYPE_VIDEO,
  391. .config_props = config_input_top,
  392. .filter_frame = filter_frame,
  393. },{
  394. .name = "bottom",
  395. .type = AVMEDIA_TYPE_VIDEO,
  396. .filter_frame = filter_frame,
  397. },
  398. { NULL }
  399. };
  400. static const AVFilterPad blend_outputs[] = {
  401. {
  402. .name = "default",
  403. .type = AVMEDIA_TYPE_VIDEO,
  404. .config_props = config_output,
  405. .request_frame = request_frame,
  406. },
  407. { NULL }
  408. };
  409. AVFilter avfilter_vf_blend = {
  410. .name = "blend",
  411. .description = NULL_IF_CONFIG_SMALL("Blend two video frames into each other."),
  412. .init = init,
  413. .uninit = uninit,
  414. .priv_size = sizeof(BlendContext),
  415. .query_formats = query_formats,
  416. .inputs = blend_inputs,
  417. .outputs = blend_outputs,
  418. .priv_class = &blend_class,
  419. };