You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

639 lines
30KB

  1. /*
  2. * Copyright (c) 2013 Paul B Mahol
  3. *
  4. * This file is part of FFmpeg.
  5. *
  6. * FFmpeg is free software; you can redistribute it and/or
  7. * modify it under the terms of the GNU Lesser General Public
  8. * License as published by the Free Software Foundation; either
  9. * version 2.1 of the License, or (at your option) any later version.
  10. *
  11. * FFmpeg is distributed in the hope that it will be useful,
  12. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  13. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  14. * Lesser General Public License for more details.
  15. *
  16. * You should have received a copy of the GNU Lesser General Public
  17. * License along with FFmpeg; if not, write to the Free Software
  18. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  19. */
  20. #include "libavutil/imgutils.h"
  21. #include "libavutil/eval.h"
  22. #include "libavutil/opt.h"
  23. #include "libavutil/pixfmt.h"
  24. #include "avfilter.h"
  25. #include "bufferqueue.h"
  26. #include "formats.h"
  27. #include "internal.h"
  28. #include "dualinput.h"
  29. #include "video.h"
  30. #define TOP 0
  31. #define BOTTOM 1
  32. enum BlendMode {
  33. BLEND_UNSET = -1,
  34. BLEND_NORMAL,
  35. BLEND_ADDITION,
  36. BLEND_AND,
  37. BLEND_AVERAGE,
  38. BLEND_BURN,
  39. BLEND_DARKEN,
  40. BLEND_DIFFERENCE,
  41. BLEND_DIFFERENCE128,
  42. BLEND_DIVIDE,
  43. BLEND_DODGE,
  44. BLEND_EXCLUSION,
  45. BLEND_HARDLIGHT,
  46. BLEND_LIGHTEN,
  47. BLEND_MULTIPLY,
  48. BLEND_NEGATION,
  49. BLEND_OR,
  50. BLEND_OVERLAY,
  51. BLEND_PHOENIX,
  52. BLEND_PINLIGHT,
  53. BLEND_REFLECT,
  54. BLEND_SCREEN,
  55. BLEND_SOFTLIGHT,
  56. BLEND_SUBTRACT,
  57. BLEND_VIVIDLIGHT,
  58. BLEND_XOR,
  59. BLEND_HARDMIX,
  60. BLEND_LINEARLIGHT,
  61. BLEND_GLOW,
  62. BLEND_NB
  63. };
  64. static const char *const var_names[] = { "X", "Y", "W", "H", "SW", "SH", "T", "N", "A", "B", "TOP", "BOTTOM", NULL };
  65. enum { VAR_X, VAR_Y, VAR_W, VAR_H, VAR_SW, VAR_SH, VAR_T, VAR_N, VAR_A, VAR_B, VAR_TOP, VAR_BOTTOM, VAR_VARS_NB };
  66. typedef struct FilterParams {
  67. enum BlendMode mode;
  68. double opacity;
  69. AVExpr *e;
  70. char *expr_str;
  71. void (*blend)(const uint8_t *top, int top_linesize,
  72. const uint8_t *bottom, int bottom_linesize,
  73. uint8_t *dst, int dst_linesize,
  74. int width, int start, int end,
  75. struct FilterParams *param, double *values);
  76. } FilterParams;
  77. typedef struct ThreadData {
  78. const AVFrame *top, *bottom;
  79. AVFrame *dst;
  80. AVFilterLink *inlink;
  81. int plane;
  82. int w, h;
  83. FilterParams *param;
  84. } ThreadData;
  85. typedef struct {
  86. const AVClass *class;
  87. FFDualInputContext dinput;
  88. int hsub, vsub; ///< chroma subsampling values
  89. int nb_planes;
  90. char *all_expr;
  91. enum BlendMode all_mode;
  92. double all_opacity;
  93. FilterParams params[4];
  94. int tblend;
  95. AVFrame *prev_frame; /* only used with tblend */
  96. } BlendContext;
  97. #define COMMON_OPTIONS \
  98. { "c0_mode", "set component #0 blend mode", OFFSET(params[0].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},\
  99. { "c1_mode", "set component #1 blend mode", OFFSET(params[1].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},\
  100. { "c2_mode", "set component #2 blend mode", OFFSET(params[2].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},\
  101. { "c3_mode", "set component #3 blend mode", OFFSET(params[3].mode), AV_OPT_TYPE_INT, {.i64=0}, 0, BLEND_NB-1, FLAGS, "mode"},\
  102. { "all_mode", "set blend mode for all components", OFFSET(all_mode), AV_OPT_TYPE_INT, {.i64=-1},-1, BLEND_NB-1, FLAGS, "mode"},\
  103. { "addition", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_ADDITION}, 0, 0, FLAGS, "mode" },\
  104. { "and", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AND}, 0, 0, FLAGS, "mode" },\
  105. { "average", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_AVERAGE}, 0, 0, FLAGS, "mode" },\
  106. { "burn", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_BURN}, 0, 0, FLAGS, "mode" },\
  107. { "darken", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DARKEN}, 0, 0, FLAGS, "mode" },\
  108. { "difference", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIFFERENCE}, 0, 0, FLAGS, "mode" },\
  109. { "difference128", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIFFERENCE128}, 0, 0, FLAGS, "mode" },\
  110. { "divide", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DIVIDE}, 0, 0, FLAGS, "mode" },\
  111. { "dodge", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_DODGE}, 0, 0, FLAGS, "mode" },\
  112. { "exclusion", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_EXCLUSION}, 0, 0, FLAGS, "mode" },\
  113. { "glow", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_GLOW}, 0, 0, FLAGS, "mode" },\
  114. { "hardlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_HARDLIGHT}, 0, 0, FLAGS, "mode" },\
  115. { "hardmix", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_HARDMIX}, 0, 0, FLAGS, "mode" },\
  116. { "lighten", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_LIGHTEN}, 0, 0, FLAGS, "mode" },\
  117. { "linearlight","", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_LINEARLIGHT},0, 0, FLAGS, "mode" },\
  118. { "multiply", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_MULTIPLY}, 0, 0, FLAGS, "mode" },\
  119. { "negation", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NEGATION}, 0, 0, FLAGS, "mode" },\
  120. { "normal", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_NORMAL}, 0, 0, FLAGS, "mode" },\
  121. { "or", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OR}, 0, 0, FLAGS, "mode" },\
  122. { "overlay", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_OVERLAY}, 0, 0, FLAGS, "mode" },\
  123. { "phoenix", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PHOENIX}, 0, 0, FLAGS, "mode" },\
  124. { "pinlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_PINLIGHT}, 0, 0, FLAGS, "mode" },\
  125. { "reflect", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_REFLECT}, 0, 0, FLAGS, "mode" },\
  126. { "screen", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SCREEN}, 0, 0, FLAGS, "mode" },\
  127. { "softlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SOFTLIGHT}, 0, 0, FLAGS, "mode" },\
  128. { "subtract", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_SUBTRACT}, 0, 0, FLAGS, "mode" },\
  129. { "vividlight", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_VIVIDLIGHT}, 0, 0, FLAGS, "mode" },\
  130. { "xor", "", 0, AV_OPT_TYPE_CONST, {.i64=BLEND_XOR}, 0, 0, FLAGS, "mode" },\
  131. { "c0_expr", "set color component #0 expression", OFFSET(params[0].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },\
  132. { "c1_expr", "set color component #1 expression", OFFSET(params[1].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },\
  133. { "c2_expr", "set color component #2 expression", OFFSET(params[2].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },\
  134. { "c3_expr", "set color component #3 expression", OFFSET(params[3].expr_str), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },\
  135. { "all_expr", "set expression for all color components", OFFSET(all_expr), AV_OPT_TYPE_STRING, {.str=NULL}, CHAR_MIN, CHAR_MAX, FLAGS },\
  136. { "c0_opacity", "set color component #0 opacity", OFFSET(params[0].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },\
  137. { "c1_opacity", "set color component #1 opacity", OFFSET(params[1].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },\
  138. { "c2_opacity", "set color component #2 opacity", OFFSET(params[2].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },\
  139. { "c3_opacity", "set color component #3 opacity", OFFSET(params[3].opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS },\
  140. { "all_opacity", "set opacity for all color components", OFFSET(all_opacity), AV_OPT_TYPE_DOUBLE, {.dbl=1}, 0, 1, FLAGS}
  141. #define OFFSET(x) offsetof(BlendContext, x)
  142. #define FLAGS AV_OPT_FLAG_FILTERING_PARAM|AV_OPT_FLAG_VIDEO_PARAM
  143. static const AVOption blend_options[] = {
  144. COMMON_OPTIONS,
  145. { "shortest", "force termination when the shortest input terminates", OFFSET(dinput.shortest), AV_OPT_TYPE_INT, {.i64=0}, 0, 1, FLAGS },
  146. { "repeatlast", "repeat last bottom frame", OFFSET(dinput.repeatlast), AV_OPT_TYPE_INT, {.i64=1}, 0, 1, FLAGS },
  147. { NULL }
  148. };
  149. AVFILTER_DEFINE_CLASS(blend);
  150. static void blend_normal(const uint8_t *top, int top_linesize,
  151. const uint8_t *bottom, int bottom_linesize,
  152. uint8_t *dst, int dst_linesize,
  153. int width, int start, int end,
  154. FilterParams *param, double *values)
  155. {
  156. av_image_copy_plane(dst, dst_linesize, top, top_linesize, width, end - start);
  157. }
  158. #define DEFINE_BLEND8(name, expr) \
  159. static void blend_## name##_8bit(const uint8_t *top, int top_linesize, \
  160. const uint8_t *bottom, int bottom_linesize, \
  161. uint8_t *dst, int dst_linesize, \
  162. int width, int start, int end, \
  163. FilterParams *param, double *values) \
  164. { \
  165. double opacity = param->opacity; \
  166. int i, j; \
  167. \
  168. for (i = start; i < end; i++) { \
  169. for (j = 0; j < width; j++) { \
  170. dst[j] = top[j] + ((expr) - top[j]) * opacity; \
  171. } \
  172. dst += dst_linesize; \
  173. top += top_linesize; \
  174. bottom += bottom_linesize; \
  175. } \
  176. }
  177. #define DEFINE_BLEND16(name, expr) \
  178. static void blend_## name##_16bit(const uint8_t *_top, int top_linesize, \
  179. const uint8_t *_bottom, int bottom_linesize, \
  180. uint8_t *_dst, int dst_linesize, \
  181. int width, int start, int end, \
  182. FilterParams *param, double *values) \
  183. { \
  184. const uint16_t *top = (uint16_t*)_top; \
  185. const uint16_t *bottom = (uint16_t*)_bottom; \
  186. uint16_t *dst = (uint16_t*)_dst; \
  187. double opacity = param->opacity; \
  188. int i, j; \
  189. dst_linesize /= 2; \
  190. top_linesize /= 2; \
  191. bottom_linesize /= 2; \
  192. \
  193. for (i = start; i < end; i++) { \
  194. for (j = 0; j < width; j++) { \
  195. dst[j] = top[j] + ((expr) - top[j]) * opacity; \
  196. } \
  197. dst += dst_linesize; \
  198. top += top_linesize; \
  199. bottom += bottom_linesize; \
  200. } \
  201. }
  202. #define A top[j]
  203. #define B bottom[j]
  204. #define MULTIPLY(x, a, b) ((x) * (((a) * (b)) / 255))
  205. #define SCREEN(x, a, b) (255 - (x) * ((255 - (a)) * (255 - (b)) / 255))
  206. #define BURN(a, b) (((a) == 0) ? (a) : FFMAX(0, 255 - ((255 - (b)) << 8) / (a)))
  207. #define DODGE(a, b) (((a) == 255) ? (a) : FFMIN(255, (((b) << 8) / (255 - (a)))))
  208. DEFINE_BLEND8(addition, FFMIN(255, A + B))
  209. DEFINE_BLEND8(average, (A + B) / 2)
  210. DEFINE_BLEND8(subtract, FFMAX(0, A - B))
  211. DEFINE_BLEND8(multiply, MULTIPLY(1, A, B))
  212. DEFINE_BLEND8(negation, 255 - FFABS(255 - A - B))
  213. DEFINE_BLEND8(difference, FFABS(A - B))
  214. DEFINE_BLEND8(difference128, av_clip_uint8(128 + A - B))
  215. DEFINE_BLEND8(screen, SCREEN(1, A, B))
  216. DEFINE_BLEND8(overlay, (A < 128) ? MULTIPLY(2, A, B) : SCREEN(2, A, B))
  217. DEFINE_BLEND8(hardlight, (B < 128) ? MULTIPLY(2, B, A) : SCREEN(2, B, A))
  218. DEFINE_BLEND8(hardmix, (A < (255 - B)) ? 0: 255)
  219. DEFINE_BLEND8(darken, FFMIN(A, B))
  220. DEFINE_BLEND8(lighten, FFMAX(A, B))
  221. DEFINE_BLEND8(divide, av_clip_uint8(((float)A / ((float)B) * 255)))
  222. DEFINE_BLEND8(dodge, DODGE(A, B))
  223. DEFINE_BLEND8(burn, BURN(A, B))
  224. DEFINE_BLEND8(softlight, (A > 127) ? B + (255 - B) * (A - 127.5) / 127.5 * (0.5 - FFABS(B - 127.5) / 255): B - B * ((127.5 - A) / 127.5) * (0.5 - FFABS(B - 127.5)/255))
  225. DEFINE_BLEND8(exclusion, A + B - 2 * A * B / 255)
  226. DEFINE_BLEND8(pinlight, (B < 128) ? FFMIN(A, 2 * B) : FFMAX(A, 2 * (B - 128)))
  227. DEFINE_BLEND8(phoenix, FFMIN(A, B) - FFMAX(A, B) + 255)
  228. DEFINE_BLEND8(reflect, (B == 255) ? B : FFMIN(255, (A * A / (255 - B))))
  229. DEFINE_BLEND8(glow, (A == 255) ? A : FFMIN(255, (B * B / (255 - A))))
  230. DEFINE_BLEND8(and, A & B)
  231. DEFINE_BLEND8(or, A | B)
  232. DEFINE_BLEND8(xor, A ^ B)
  233. DEFINE_BLEND8(vividlight, (A < 128) ? BURN(2 * A, B) : DODGE(2 * (A - 128), B))
  234. DEFINE_BLEND8(linearlight,av_clip_uint8((B < 128) ? B + 2 * A - 255 : B + 2 * (A - 128)))
  235. #undef MULTIPLY
  236. #undef SCREEN
  237. #undef BURN
  238. #undef DODGE
  239. #define MULTIPLY(x, a, b) ((x) * (((a) * (b)) / 65535))
  240. #define SCREEN(x, a, b) (65535 - (x) * ((65535 - (a)) * (65535 - (b)) / 65535))
  241. #define BURN(a, b) (((a) == 0) ? (a) : FFMAX(0, 65535 - ((65535 - (b)) << 16) / (a)))
  242. #define DODGE(a, b) (((a) == 65535) ? (a) : FFMIN(65535, (((b) << 16) / (65535 - (a)))))
  243. DEFINE_BLEND16(addition, FFMIN(65535, A + B))
  244. DEFINE_BLEND16(average, (A + B) / 2)
  245. DEFINE_BLEND16(subtract, FFMAX(0, A - B))
  246. DEFINE_BLEND16(multiply, MULTIPLY(1, A, B))
  247. DEFINE_BLEND16(negation, 65535 - FFABS(65535 - A - B))
  248. DEFINE_BLEND16(difference, FFABS(A - B))
  249. DEFINE_BLEND16(difference128, av_clip_uint16(32768 + A - B))
  250. DEFINE_BLEND16(screen, SCREEN(1, A, B))
  251. DEFINE_BLEND16(overlay, (A < 32768) ? MULTIPLY(2, A, B) : SCREEN(2, A, B))
  252. DEFINE_BLEND16(hardlight, (B < 32768) ? MULTIPLY(2, B, A) : SCREEN(2, B, A))
  253. DEFINE_BLEND16(hardmix, (A < (65535 - B)) ? 0: 65535)
  254. DEFINE_BLEND16(darken, FFMIN(A, B))
  255. DEFINE_BLEND16(lighten, FFMAX(A, B))
  256. DEFINE_BLEND16(divide, av_clip_uint16(((float)A / ((float)B) * 65535)))
  257. DEFINE_BLEND16(dodge, DODGE(A, B))
  258. DEFINE_BLEND16(burn, BURN(A, B))
  259. DEFINE_BLEND16(softlight, (A > 32767) ? B + (65535 - B) * (A - 32767.5) / 32767.5 * (0.5 - FFABS(B - 32767.5) / 65535): B - B * ((32767.5 - A) / 32767.5) * (0.5 - FFABS(B - 32767.5)/65535))
  260. DEFINE_BLEND16(exclusion, A + B - 2 * A * B / 65535)
  261. DEFINE_BLEND16(pinlight, (B < 32768) ? FFMIN(A, 2 * B) : FFMAX(A, 2 * (B - 32768)))
  262. DEFINE_BLEND16(phoenix, FFMIN(A, B) - FFMAX(A, B) + 65535)
  263. DEFINE_BLEND16(reflect, (B == 65535) ? B : FFMIN(65535, (A * A / (65535 - B))))
  264. DEFINE_BLEND16(glow, (A == 65535) ? A : FFMIN(65535, (B * B / (65535 - A))))
  265. DEFINE_BLEND16(and, A & B)
  266. DEFINE_BLEND16(or, A | B)
  267. DEFINE_BLEND16(xor, A ^ B)
  268. DEFINE_BLEND16(vividlight, (A < 32768) ? BURN(2 * A, B) : DODGE(2 * (A - 32768), B))
  269. DEFINE_BLEND16(linearlight,av_clip_uint16((B < 32768) ? B + 2 * A - 65535 : B + 2 * (A - 32768)))
  270. #define DEFINE_BLEND_EXPR(type, name, div) \
  271. static void blend_expr_## name(const uint8_t *_top, int top_linesize, \
  272. const uint8_t *_bottom, int bottom_linesize, \
  273. uint8_t *_dst, int dst_linesize, \
  274. int width, int start, int end, \
  275. FilterParams *param, double *values) \
  276. { \
  277. const type *top = (type*)_top; \
  278. const type *bottom = (type*)_bottom; \
  279. type *dst = (type*)_dst; \
  280. AVExpr *e = param->e; \
  281. int y, x; \
  282. dst_linesize /= div; \
  283. top_linesize /= div; \
  284. bottom_linesize /= div; \
  285. \
  286. for (y = start; y < end; y++) { \
  287. values[VAR_Y] = y; \
  288. for (x = 0; x < width; x++) { \
  289. values[VAR_X] = x; \
  290. values[VAR_TOP] = values[VAR_A] = top[x]; \
  291. values[VAR_BOTTOM] = values[VAR_B] = bottom[x]; \
  292. dst[x] = av_expr_eval(e, values, NULL); \
  293. } \
  294. dst += dst_linesize; \
  295. top += top_linesize; \
  296. bottom += bottom_linesize; \
  297. } \
  298. }
  299. DEFINE_BLEND_EXPR(uint8_t, 8bit, 1)
  300. DEFINE_BLEND_EXPR(uint16_t, 16bit, 2)
  301. static int filter_slice(AVFilterContext *ctx, void *arg, int jobnr, int nb_jobs)
  302. {
  303. ThreadData *td = arg;
  304. int slice_start = (td->h * jobnr ) / nb_jobs;
  305. int slice_end = (td->h * (jobnr+1)) / nb_jobs;
  306. const uint8_t *top = td->top->data[td->plane];
  307. const uint8_t *bottom = td->bottom->data[td->plane];
  308. uint8_t *dst = td->dst->data[td->plane];
  309. double values[VAR_VARS_NB];
  310. values[VAR_N] = td->inlink->frame_count;
  311. values[VAR_T] = td->dst->pts == AV_NOPTS_VALUE ? NAN : td->dst->pts * av_q2d(td->inlink->time_base);
  312. values[VAR_W] = td->w;
  313. values[VAR_H] = td->h;
  314. values[VAR_SW] = td->w / (double)td->dst->width;
  315. values[VAR_SH] = td->h / (double)td->dst->height;
  316. td->param->blend(top + slice_start * td->top->linesize[td->plane],
  317. td->top->linesize[td->plane],
  318. bottom + slice_start * td->bottom->linesize[td->plane],
  319. td->bottom->linesize[td->plane],
  320. dst + slice_start * td->dst->linesize[td->plane],
  321. td->dst->linesize[td->plane],
  322. td->w, slice_start, slice_end, td->param, &values[0]);
  323. return 0;
  324. }
  325. static AVFrame *blend_frame(AVFilterContext *ctx, AVFrame *top_buf,
  326. const AVFrame *bottom_buf)
  327. {
  328. BlendContext *s = ctx->priv;
  329. AVFilterLink *inlink = ctx->inputs[0];
  330. AVFilterLink *outlink = ctx->outputs[0];
  331. AVFrame *dst_buf;
  332. int plane;
  333. dst_buf = ff_get_video_buffer(outlink, outlink->w, outlink->h);
  334. if (!dst_buf)
  335. return top_buf;
  336. av_frame_copy_props(dst_buf, top_buf);
  337. for (plane = 0; plane < s->nb_planes; plane++) {
  338. int hsub = plane == 1 || plane == 2 ? s->hsub : 0;
  339. int vsub = plane == 1 || plane == 2 ? s->vsub : 0;
  340. int outw = FF_CEIL_RSHIFT(dst_buf->width, hsub);
  341. int outh = FF_CEIL_RSHIFT(dst_buf->height, vsub);
  342. FilterParams *param = &s->params[plane];
  343. ThreadData td = { .top = top_buf, .bottom = bottom_buf, .dst = dst_buf,
  344. .w = outw, .h = outh, .param = param, .plane = plane,
  345. .inlink = inlink };
  346. ctx->internal->execute(ctx, filter_slice, &td, NULL, FFMIN(outh, ctx->graph->nb_threads));
  347. }
  348. if (!s->tblend)
  349. av_frame_free(&top_buf);
  350. return dst_buf;
  351. }
  352. static av_cold int init(AVFilterContext *ctx)
  353. {
  354. BlendContext *s = ctx->priv;
  355. s->tblend = !strcmp(ctx->filter->name, "tblend");
  356. s->dinput.process = blend_frame;
  357. return 0;
  358. }
  359. static int query_formats(AVFilterContext *ctx)
  360. {
  361. static const enum AVPixelFormat pix_fmts[] = {
  362. AV_PIX_FMT_YUVA444P, AV_PIX_FMT_YUVA422P, AV_PIX_FMT_YUVA420P,
  363. AV_PIX_FMT_YUVJ444P, AV_PIX_FMT_YUVJ440P, AV_PIX_FMT_YUVJ422P,AV_PIX_FMT_YUVJ420P, AV_PIX_FMT_YUVJ411P,
  364. AV_PIX_FMT_YUV444P, AV_PIX_FMT_YUV440P, AV_PIX_FMT_YUV422P, AV_PIX_FMT_YUV420P, AV_PIX_FMT_YUV411P, AV_PIX_FMT_YUV410P,
  365. AV_PIX_FMT_GBRP, AV_PIX_FMT_GBRAP, AV_PIX_FMT_GRAY8,
  366. AV_PIX_FMT_YUV420P16, AV_PIX_FMT_YUV422P16, AV_PIX_FMT_YUV444P16,
  367. AV_PIX_FMT_YUVA420P16, AV_PIX_FMT_YUVA422P16, AV_PIX_FMT_YUVA444P16,
  368. AV_PIX_FMT_GBRP16, AV_PIX_FMT_GRAY16,
  369. AV_PIX_FMT_NONE
  370. };
  371. AVFilterFormats *fmts_list = ff_make_format_list(pix_fmts);
  372. if (!fmts_list)
  373. return AVERROR(ENOMEM);
  374. return ff_set_common_formats(ctx, fmts_list);
  375. }
  376. static av_cold void uninit(AVFilterContext *ctx)
  377. {
  378. BlendContext *s = ctx->priv;
  379. int i;
  380. ff_dualinput_uninit(&s->dinput);
  381. av_frame_free(&s->prev_frame);
  382. for (i = 0; i < FF_ARRAY_ELEMS(s->params); i++)
  383. av_expr_free(s->params[i].e);
  384. }
  385. static int config_output(AVFilterLink *outlink)
  386. {
  387. AVFilterContext *ctx = outlink->src;
  388. AVFilterLink *toplink = ctx->inputs[TOP];
  389. BlendContext *s = ctx->priv;
  390. const AVPixFmtDescriptor *pix_desc = av_pix_fmt_desc_get(toplink->format);
  391. int ret, plane, is_16bit;
  392. if (!s->tblend) {
  393. AVFilterLink *bottomlink = ctx->inputs[BOTTOM];
  394. if (toplink->format != bottomlink->format) {
  395. av_log(ctx, AV_LOG_ERROR, "inputs must be of same pixel format\n");
  396. return AVERROR(EINVAL);
  397. }
  398. if (toplink->w != bottomlink->w ||
  399. toplink->h != bottomlink->h ||
  400. toplink->sample_aspect_ratio.num != bottomlink->sample_aspect_ratio.num ||
  401. toplink->sample_aspect_ratio.den != bottomlink->sample_aspect_ratio.den) {
  402. av_log(ctx, AV_LOG_ERROR, "First input link %s parameters "
  403. "(size %dx%d, SAR %d:%d) do not match the corresponding "
  404. "second input link %s parameters (%dx%d, SAR %d:%d)\n",
  405. ctx->input_pads[TOP].name, toplink->w, toplink->h,
  406. toplink->sample_aspect_ratio.num,
  407. toplink->sample_aspect_ratio.den,
  408. ctx->input_pads[BOTTOM].name, bottomlink->w, bottomlink->h,
  409. bottomlink->sample_aspect_ratio.num,
  410. bottomlink->sample_aspect_ratio.den);
  411. return AVERROR(EINVAL);
  412. }
  413. }
  414. outlink->w = toplink->w;
  415. outlink->h = toplink->h;
  416. outlink->time_base = toplink->time_base;
  417. outlink->sample_aspect_ratio = toplink->sample_aspect_ratio;
  418. outlink->frame_rate = toplink->frame_rate;
  419. s->hsub = pix_desc->log2_chroma_w;
  420. s->vsub = pix_desc->log2_chroma_h;
  421. is_16bit = pix_desc->comp[0].depth_minus1 == 15;
  422. s->nb_planes = av_pix_fmt_count_planes(toplink->format);
  423. if (s->tblend)
  424. outlink->flags |= FF_LINK_FLAG_REQUEST_LOOP;
  425. else if ((ret = ff_dualinput_init(ctx, &s->dinput)) < 0)
  426. return ret;
  427. for (plane = 0; plane < FF_ARRAY_ELEMS(s->params); plane++) {
  428. FilterParams *param = &s->params[plane];
  429. if (s->all_mode >= 0)
  430. param->mode = s->all_mode;
  431. if (s->all_opacity < 1)
  432. param->opacity = s->all_opacity;
  433. switch (param->mode) {
  434. case BLEND_ADDITION: param->blend = is_16bit ? blend_addition_16bit : blend_addition_8bit; break;
  435. case BLEND_AND: param->blend = is_16bit ? blend_and_16bit : blend_and_8bit; break;
  436. case BLEND_AVERAGE: param->blend = is_16bit ? blend_average_16bit : blend_average_8bit; break;
  437. case BLEND_BURN: param->blend = is_16bit ? blend_burn_16bit : blend_burn_8bit; break;
  438. case BLEND_DARKEN: param->blend = is_16bit ? blend_darken_16bit : blend_darken_8bit; break;
  439. case BLEND_DIFFERENCE: param->blend = is_16bit ? blend_difference_16bit : blend_difference_8bit; break;
  440. case BLEND_DIFFERENCE128: param->blend = is_16bit ? blend_difference128_16bit: blend_difference128_8bit; break;
  441. case BLEND_DIVIDE: param->blend = is_16bit ? blend_divide_16bit : blend_divide_8bit; break;
  442. case BLEND_DODGE: param->blend = is_16bit ? blend_dodge_16bit : blend_dodge_8bit; break;
  443. case BLEND_EXCLUSION: param->blend = is_16bit ? blend_exclusion_16bit : blend_exclusion_8bit; break;
  444. case BLEND_GLOW: param->blend = is_16bit ? blend_glow_16bit : blend_glow_8bit; break;
  445. case BLEND_HARDLIGHT: param->blend = is_16bit ? blend_hardlight_16bit : blend_hardlight_8bit; break;
  446. case BLEND_HARDMIX: param->blend = is_16bit ? blend_hardmix_16bit : blend_hardmix_8bit; break;
  447. case BLEND_LIGHTEN: param->blend = is_16bit ? blend_lighten_16bit : blend_lighten_8bit; break;
  448. case BLEND_LINEARLIGHT:param->blend = is_16bit ? blend_linearlight_16bit: blend_linearlight_8bit;break;
  449. case BLEND_MULTIPLY: param->blend = is_16bit ? blend_multiply_16bit : blend_multiply_8bit; break;
  450. case BLEND_NEGATION: param->blend = is_16bit ? blend_negation_16bit : blend_negation_8bit; break;
  451. case BLEND_NORMAL: param->blend = blend_normal; break;
  452. case BLEND_OR: param->blend = is_16bit ? blend_or_16bit : blend_or_8bit; break;
  453. case BLEND_OVERLAY: param->blend = is_16bit ? blend_overlay_16bit : blend_overlay_8bit; break;
  454. case BLEND_PHOENIX: param->blend = is_16bit ? blend_phoenix_16bit : blend_phoenix_8bit; break;
  455. case BLEND_PINLIGHT: param->blend = is_16bit ? blend_pinlight_16bit : blend_pinlight_8bit; break;
  456. case BLEND_REFLECT: param->blend = is_16bit ? blend_reflect_16bit : blend_reflect_8bit; break;
  457. case BLEND_SCREEN: param->blend = is_16bit ? blend_screen_16bit : blend_screen_8bit; break;
  458. case BLEND_SOFTLIGHT: param->blend = is_16bit ? blend_softlight_16bit : blend_softlight_8bit; break;
  459. case BLEND_SUBTRACT: param->blend = is_16bit ? blend_subtract_16bit : blend_subtract_8bit; break;
  460. case BLEND_VIVIDLIGHT: param->blend = is_16bit ? blend_vividlight_16bit : blend_vividlight_8bit; break;
  461. case BLEND_XOR: param->blend = is_16bit ? blend_xor_16bit : blend_xor_8bit; break;
  462. }
  463. if (s->all_expr && !param->expr_str) {
  464. param->expr_str = av_strdup(s->all_expr);
  465. if (!param->expr_str)
  466. return AVERROR(ENOMEM);
  467. }
  468. if (param->expr_str) {
  469. ret = av_expr_parse(&param->e, param->expr_str, var_names,
  470. NULL, NULL, NULL, NULL, 0, ctx);
  471. if (ret < 0)
  472. return ret;
  473. param->blend = is_16bit? blend_expr_16bit : blend_expr_8bit;
  474. }
  475. }
  476. return 0;
  477. }
  478. #if CONFIG_BLEND_FILTER
  479. static int request_frame(AVFilterLink *outlink)
  480. {
  481. BlendContext *s = outlink->src->priv;
  482. return ff_dualinput_request_frame(&s->dinput, outlink);
  483. }
  484. static int filter_frame(AVFilterLink *inlink, AVFrame *buf)
  485. {
  486. BlendContext *s = inlink->dst->priv;
  487. return ff_dualinput_filter_frame(&s->dinput, inlink, buf);
  488. }
  489. static const AVFilterPad blend_inputs[] = {
  490. {
  491. .name = "top",
  492. .type = AVMEDIA_TYPE_VIDEO,
  493. .filter_frame = filter_frame,
  494. },{
  495. .name = "bottom",
  496. .type = AVMEDIA_TYPE_VIDEO,
  497. .filter_frame = filter_frame,
  498. },
  499. { NULL }
  500. };
  501. static const AVFilterPad blend_outputs[] = {
  502. {
  503. .name = "default",
  504. .type = AVMEDIA_TYPE_VIDEO,
  505. .config_props = config_output,
  506. .request_frame = request_frame,
  507. },
  508. { NULL }
  509. };
  510. AVFilter ff_vf_blend = {
  511. .name = "blend",
  512. .description = NULL_IF_CONFIG_SMALL("Blend two video frames into each other."),
  513. .init = init,
  514. .uninit = uninit,
  515. .priv_size = sizeof(BlendContext),
  516. .query_formats = query_formats,
  517. .inputs = blend_inputs,
  518. .outputs = blend_outputs,
  519. .priv_class = &blend_class,
  520. .flags = AVFILTER_FLAG_SUPPORT_TIMELINE_INTERNAL | AVFILTER_FLAG_SLICE_THREADS,
  521. };
  522. #endif
  523. #if CONFIG_TBLEND_FILTER
  524. static int tblend_filter_frame(AVFilterLink *inlink, AVFrame *frame)
  525. {
  526. BlendContext *s = inlink->dst->priv;
  527. AVFilterLink *outlink = inlink->dst->outputs[0];
  528. if (s->prev_frame) {
  529. AVFrame *out = blend_frame(inlink->dst, frame, s->prev_frame);
  530. av_frame_free(&s->prev_frame);
  531. s->prev_frame = frame;
  532. return ff_filter_frame(outlink, out);
  533. }
  534. s->prev_frame = frame;
  535. return 0;
  536. }
  537. static const AVOption tblend_options[] = {
  538. COMMON_OPTIONS,
  539. { NULL }
  540. };
  541. AVFILTER_DEFINE_CLASS(tblend);
  542. static const AVFilterPad tblend_inputs[] = {
  543. {
  544. .name = "default",
  545. .type = AVMEDIA_TYPE_VIDEO,
  546. .filter_frame = tblend_filter_frame,
  547. },
  548. { NULL }
  549. };
  550. static const AVFilterPad tblend_outputs[] = {
  551. {
  552. .name = "default",
  553. .type = AVMEDIA_TYPE_VIDEO,
  554. .config_props = config_output,
  555. },
  556. { NULL }
  557. };
  558. AVFilter ff_vf_tblend = {
  559. .name = "tblend",
  560. .description = NULL_IF_CONFIG_SMALL("Blend successive frames."),
  561. .priv_size = sizeof(BlendContext),
  562. .priv_class = &tblend_class,
  563. .query_formats = query_formats,
  564. .init = init,
  565. .uninit = uninit,
  566. .inputs = tblend_inputs,
  567. .outputs = tblend_outputs,
  568. .flags = AVFILTER_FLAG_SLICE_THREADS,
  569. };
  570. #endif