You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

549 lines
16KB

  1. /*
  2. *
  3. * This file is part of FFmpeg.
  4. *
  5. * FFmpeg is free software; you can redistribute it and/or
  6. * modify it under the terms of the GNU Lesser General Public
  7. * License as published by the Free Software Foundation; either
  8. * version 2.1 of the License, or (at your option) any later version.
  9. *
  10. * FFmpeg is distributed in the hope that it will be useful,
  11. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  12. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  13. * Lesser General Public License for more details.
  14. *
  15. * You should have received a copy of the GNU Lesser General Public
  16. * License along with FFmpeg; if not, write to the Free Software
  17. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  18. */
  19. #include "channel_layout.h"
  20. #include "buffer.h"
  21. #include "common.h"
  22. #include "dict.h"
  23. #include "frame.h"
  24. #include "imgutils.h"
  25. #include "mem.h"
  26. #include "samplefmt.h"
  27. #define MAKE_ACCESSORS(str, name, type, field) \
  28. type av_##name##_get_##field(const str *s) { return s->field; } \
  29. void av_##name##_set_##field(str *s, type v) { s->field = v; }
  30. MAKE_ACCESSORS(AVFrame, frame, int64_t, best_effort_timestamp)
  31. MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_duration)
  32. MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_pos)
  33. MAKE_ACCESSORS(AVFrame, frame, int64_t, channel_layout)
  34. MAKE_ACCESSORS(AVFrame, frame, int, channels)
  35. MAKE_ACCESSORS(AVFrame, frame, int, sample_rate)
  36. MAKE_ACCESSORS(AVFrame, frame, AVDictionary *, metadata)
  37. MAKE_ACCESSORS(AVFrame, frame, int, decode_error_flags)
  38. MAKE_ACCESSORS(AVFrame, frame, int, pkt_size)
  39. AVDictionary **avpriv_frame_get_metadatap(AVFrame *frame) {return &frame->metadata;};
  40. int av_frame_set_qp_table(AVFrame *f, AVBufferRef *buf, int stride, int qp_type)
  41. {
  42. av_buffer_unref(&f->qp_table_buf);
  43. f->qp_table_buf = buf;
  44. f->qscale_table = buf->data;
  45. f->qstride = stride;
  46. f->qscale_type = qp_type;
  47. return 0;
  48. }
  49. int8_t *av_frame_get_qp_table(AVFrame *f, int *stride, int *type)
  50. {
  51. *stride = f->qstride;
  52. *type = f->qscale_type;
  53. if (!f->qp_table_buf)
  54. return NULL;
  55. return f->qp_table_buf->data;
  56. }
  57. static void get_frame_defaults(AVFrame *frame)
  58. {
  59. if (frame->extended_data != frame->data)
  60. av_freep(&frame->extended_data);
  61. memset(frame, 0, sizeof(*frame));
  62. frame->pts =
  63. frame->pkt_dts =
  64. frame->pkt_pts = AV_NOPTS_VALUE;
  65. av_frame_set_best_effort_timestamp(frame, AV_NOPTS_VALUE);
  66. av_frame_set_pkt_duration (frame, 0);
  67. av_frame_set_pkt_pos (frame, -1);
  68. av_frame_set_pkt_size (frame, -1);
  69. frame->key_frame = 1;
  70. frame->sample_aspect_ratio = (AVRational){ 0, 1 };
  71. frame->format = -1; /* unknown */
  72. frame->extended_data = frame->data;
  73. }
  74. AVFrame *av_frame_alloc(void)
  75. {
  76. AVFrame *frame = av_mallocz(sizeof(*frame));
  77. if (!frame)
  78. return NULL;
  79. frame->extended_data = NULL;
  80. get_frame_defaults(frame);
  81. return frame;
  82. }
  83. void av_frame_free(AVFrame **frame)
  84. {
  85. if (!frame || !*frame)
  86. return;
  87. av_frame_unref(*frame);
  88. av_freep(frame);
  89. }
  90. static int get_video_buffer(AVFrame *frame, int align)
  91. {
  92. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
  93. int ret, i;
  94. if (!desc)
  95. return AVERROR(EINVAL);
  96. if ((ret = av_image_check_size(frame->width, frame->height, 0, NULL)) < 0)
  97. return ret;
  98. if (!frame->linesize[0]) {
  99. ret = av_image_fill_linesizes(frame->linesize, frame->format,
  100. frame->width);
  101. if (ret < 0)
  102. return ret;
  103. for (i = 0; i < 4 && frame->linesize[i]; i++)
  104. frame->linesize[i] = FFALIGN(frame->linesize[i], align);
  105. }
  106. for (i = 0; i < 4 && frame->linesize[i]; i++) {
  107. int h = FFALIGN(frame->height, 32);
  108. if (i == 1 || i == 2)
  109. h = -((-h) >> desc->log2_chroma_h);
  110. frame->buf[i] = av_buffer_alloc(frame->linesize[i] * h + 16);
  111. if (!frame->buf[i])
  112. goto fail;
  113. frame->data[i] = frame->buf[i]->data;
  114. }
  115. if (desc->flags & PIX_FMT_PAL || desc->flags & PIX_FMT_PSEUDOPAL) {
  116. av_buffer_unref(&frame->buf[1]);
  117. frame->buf[1] = av_buffer_alloc(1024);
  118. if (!frame->buf[1])
  119. goto fail;
  120. frame->data[1] = frame->buf[1]->data;
  121. }
  122. frame->extended_data = frame->data;
  123. return 0;
  124. fail:
  125. av_frame_unref(frame);
  126. return AVERROR(ENOMEM);
  127. }
  128. static int get_audio_buffer(AVFrame *frame, int align)
  129. {
  130. int channels = av_get_channel_layout_nb_channels(frame->channel_layout);
  131. int planar = av_sample_fmt_is_planar(frame->format);
  132. int planes = planar ? channels : 1;
  133. int ret, i;
  134. if (!frame->linesize[0]) {
  135. ret = av_samples_get_buffer_size(&frame->linesize[0], channels,
  136. frame->nb_samples, frame->format,
  137. align);
  138. if (ret < 0)
  139. return ret;
  140. }
  141. if (planes > AV_NUM_DATA_POINTERS) {
  142. frame->extended_data = av_mallocz(planes *
  143. sizeof(*frame->extended_data));
  144. frame->extended_buf = av_mallocz((planes - AV_NUM_DATA_POINTERS) *
  145. sizeof(*frame->extended_buf));
  146. if (!frame->extended_data || !frame->extended_buf) {
  147. av_freep(&frame->extended_data);
  148. av_freep(&frame->extended_buf);
  149. return AVERROR(ENOMEM);
  150. }
  151. frame->nb_extended_buf = planes - AV_NUM_DATA_POINTERS;
  152. } else
  153. frame->extended_data = frame->data;
  154. for (i = 0; i < FFMIN(planes, AV_NUM_DATA_POINTERS); i++) {
  155. frame->buf[i] = av_buffer_alloc(frame->linesize[0]);
  156. if (!frame->buf[i]) {
  157. av_frame_unref(frame);
  158. return AVERROR(ENOMEM);
  159. }
  160. frame->extended_data[i] = frame->data[i] = frame->buf[i]->data;
  161. }
  162. for (i = 0; i < planes - AV_NUM_DATA_POINTERS; i++) {
  163. frame->extended_buf[i] = av_buffer_alloc(frame->linesize[0]);
  164. if (!frame->extended_buf[i]) {
  165. av_frame_unref(frame);
  166. return AVERROR(ENOMEM);
  167. }
  168. frame->extended_data[i + AV_NUM_DATA_POINTERS] = frame->extended_buf[i]->data;
  169. }
  170. return 0;
  171. }
  172. int av_frame_get_buffer(AVFrame *frame, int align)
  173. {
  174. if (frame->format < 0)
  175. return AVERROR(EINVAL);
  176. if (frame->width > 0 && frame->height > 0)
  177. return get_video_buffer(frame, align);
  178. else if (frame->nb_samples > 0 && frame->channel_layout)
  179. return get_audio_buffer(frame, align);
  180. return AVERROR(EINVAL);
  181. }
  182. int av_frame_ref(AVFrame *dst, AVFrame *src)
  183. {
  184. int i, ret = 0;
  185. dst->format = src->format;
  186. dst->width = src->width;
  187. dst->height = src->height;
  188. dst->channels = src->channels;
  189. dst->channel_layout = src->channel_layout;
  190. dst->nb_samples = src->nb_samples;
  191. ret = av_frame_copy_props(dst, src);
  192. if (ret < 0)
  193. return ret;
  194. /* duplicate the frame data if it's not refcounted */
  195. if (!src->buf[0]) {
  196. ret = av_frame_get_buffer(dst, 32);
  197. if (ret < 0)
  198. return ret;
  199. if (src->nb_samples) {
  200. int ch = av_get_channel_layout_nb_channels(src->channel_layout);
  201. av_samples_copy(dst->extended_data, src->extended_data, 0, 0,
  202. dst->nb_samples, ch, dst->format);
  203. } else {
  204. av_image_copy(dst->data, dst->linesize, src->data, src->linesize,
  205. dst->format, dst->width, dst->height);
  206. }
  207. return 0;
  208. }
  209. /* ref the buffers */
  210. for (i = 0; i < FF_ARRAY_ELEMS(src->buf) && src->buf[i]; i++) {
  211. dst->buf[i] = av_buffer_ref(src->buf[i]);
  212. if (!dst->buf[i]) {
  213. ret = AVERROR(ENOMEM);
  214. goto fail;
  215. }
  216. }
  217. if (src->extended_buf) {
  218. dst->extended_buf = av_mallocz(sizeof(*dst->extended_buf) *
  219. src->nb_extended_buf);
  220. if (!dst->extended_buf) {
  221. ret = AVERROR(ENOMEM);
  222. goto fail;
  223. }
  224. dst->nb_extended_buf = src->nb_extended_buf;
  225. for (i = 0; i < src->nb_extended_buf; i++) {
  226. dst->extended_buf[i] = av_buffer_ref(src->extended_buf[i]);
  227. if (!dst->extended_buf[i]) {
  228. ret = AVERROR(ENOMEM);
  229. goto fail;
  230. }
  231. }
  232. }
  233. /* duplicate extended data */
  234. if (src->extended_data != src->data) {
  235. int ch = av_get_channel_layout_nb_channels(src->channel_layout);
  236. if (!ch) {
  237. ret = AVERROR(EINVAL);
  238. goto fail;
  239. }
  240. dst->extended_data = av_malloc(sizeof(*dst->extended_data) * ch);
  241. if (!dst->extended_data) {
  242. ret = AVERROR(ENOMEM);
  243. goto fail;
  244. }
  245. memcpy(dst->extended_data, src->extended_data, sizeof(*src->extended_data) * ch);
  246. } else
  247. dst->extended_data = dst->data;
  248. memcpy(dst->data, src->data, sizeof(src->data));
  249. memcpy(dst->linesize, src->linesize, sizeof(src->linesize));
  250. return 0;
  251. fail:
  252. av_frame_unref(dst);
  253. return ret;
  254. }
  255. AVFrame *av_frame_clone(AVFrame *src)
  256. {
  257. AVFrame *ret = av_frame_alloc();
  258. if (!ret)
  259. return NULL;
  260. if (av_frame_ref(ret, src) < 0)
  261. av_frame_free(&ret);
  262. return ret;
  263. }
  264. void av_frame_unref(AVFrame *frame)
  265. {
  266. int i;
  267. for (i = 0; i < frame->nb_side_data; i++) {
  268. av_freep(&frame->side_data[i]->data);
  269. av_dict_free(&frame->side_data[i]->metadata);
  270. av_freep(&frame->side_data[i]);
  271. }
  272. av_freep(&frame->side_data);
  273. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
  274. av_buffer_unref(&frame->buf[i]);
  275. for (i = 0; i < frame->nb_extended_buf; i++)
  276. av_buffer_unref(&frame->extended_buf[i]);
  277. av_freep(&frame->extended_buf);
  278. av_dict_free(&frame->metadata);
  279. av_buffer_unref(&frame->qp_table_buf);
  280. get_frame_defaults(frame);
  281. }
  282. void av_frame_move_ref(AVFrame *dst, AVFrame *src)
  283. {
  284. *dst = *src;
  285. if (src->extended_data == src->data)
  286. dst->extended_data = dst->data;
  287. memset(src, 0, sizeof(*src));
  288. get_frame_defaults(src);
  289. }
  290. int av_frame_is_writable(AVFrame *frame)
  291. {
  292. int i, ret = 1;
  293. /* assume non-refcounted frames are not writable */
  294. if (!frame->buf[0])
  295. return 0;
  296. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++)
  297. ret &= !!av_buffer_is_writable(frame->buf[i]);
  298. for (i = 0; i < frame->nb_extended_buf; i++)
  299. ret &= !!av_buffer_is_writable(frame->extended_buf[i]);
  300. return ret;
  301. }
  302. int av_frame_make_writable(AVFrame *frame)
  303. {
  304. AVFrame tmp;
  305. int ret;
  306. if (!frame->buf[0])
  307. return AVERROR(EINVAL);
  308. if (av_frame_is_writable(frame))
  309. return 0;
  310. memset(&tmp, 0, sizeof(tmp));
  311. tmp.format = frame->format;
  312. tmp.width = frame->width;
  313. tmp.height = frame->height;
  314. tmp.channels = frame->channels;
  315. tmp.channel_layout = frame->channel_layout;
  316. tmp.nb_samples = frame->nb_samples;
  317. ret = av_frame_get_buffer(&tmp, 32);
  318. if (ret < 0)
  319. return ret;
  320. if (tmp.nb_samples) {
  321. int ch = av_get_channel_layout_nb_channels(tmp.channel_layout);
  322. av_samples_copy(tmp.extended_data, frame->extended_data, 0, 0,
  323. frame->nb_samples, ch, frame->format);
  324. } else {
  325. av_image_copy(tmp.data, tmp.linesize, frame->data, frame->linesize,
  326. frame->format, frame->width, frame->height);
  327. }
  328. ret = av_frame_copy_props(&tmp, frame);
  329. if (ret < 0) {
  330. av_frame_unref(&tmp);
  331. return ret;
  332. }
  333. av_frame_unref(frame);
  334. *frame = tmp;
  335. if (tmp.data == tmp.extended_data)
  336. frame->extended_data = frame->data;
  337. return 0;
  338. }
  339. int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
  340. {
  341. int i;
  342. dst->key_frame = src->key_frame;
  343. dst->pict_type = src->pict_type;
  344. dst->sample_aspect_ratio = src->sample_aspect_ratio;
  345. dst->pts = src->pts;
  346. dst->interlaced_frame = src->interlaced_frame;
  347. dst->top_field_first = src->top_field_first;
  348. dst->sample_rate = src->sample_rate;
  349. dst->opaque = src->opaque;
  350. #if FF_API_AVFRAME_LAVC
  351. dst->type = src->type;
  352. #endif
  353. dst->pkt_pts = src->pkt_pts;
  354. dst->pkt_dts = src->pkt_dts;
  355. dst->pkt_pos = src->pkt_pos;
  356. dst->pkt_size = src->pkt_size;
  357. dst->pkt_duration = src->pkt_duration;
  358. dst->reordered_opaque = src->reordered_opaque;
  359. dst->quality = src->quality;
  360. dst->best_effort_timestamp = src->best_effort_timestamp;
  361. dst->coded_picture_number = src->coded_picture_number;
  362. dst->display_picture_number = src->display_picture_number;
  363. dst->decode_error_flags = src->decode_error_flags;
  364. av_dict_copy(&dst->metadata, src->metadata, 0);
  365. for (i = 0; i < src->nb_side_data; i++) {
  366. const AVFrameSideData *sd_src = src->side_data[i];
  367. AVFrameSideData *sd_dst = av_frame_new_side_data(dst, sd_src->type,
  368. sd_src->size);
  369. if (!sd_dst) {
  370. for (i = 0; i < dst->nb_side_data; i++) {
  371. av_freep(&dst->side_data[i]->data);
  372. av_freep(&dst->side_data[i]);
  373. av_dict_free(&dst->side_data[i]->metadata);
  374. }
  375. av_freep(&dst->side_data);
  376. return AVERROR(ENOMEM);
  377. }
  378. memcpy(sd_dst->data, sd_src->data, sd_src->size);
  379. av_dict_copy(&sd_dst->metadata, sd_src->metadata, 0);
  380. }
  381. dst->qscale_table = NULL;
  382. dst->qstride = 0;
  383. dst->qscale_type = 0;
  384. if (src->qp_table_buf) {
  385. dst->qp_table_buf = av_buffer_ref(src->qp_table_buf);
  386. if (dst->qp_table_buf) {
  387. dst->qscale_table = dst->qp_table_buf->data;
  388. dst->qstride = src->qstride;
  389. dst->qscale_type = src->qscale_type;
  390. }
  391. }
  392. return 0;
  393. }
  394. AVBufferRef *av_frame_get_plane_buffer(AVFrame *frame, int plane)
  395. {
  396. uint8_t *data;
  397. int planes, i;
  398. if (frame->nb_samples) {
  399. int channels = av_get_channel_layout_nb_channels(frame->channel_layout);
  400. if (!channels)
  401. return NULL;
  402. planes = av_sample_fmt_is_planar(frame->format) ? channels : 1;
  403. } else
  404. planes = 4;
  405. if (plane < 0 || plane >= planes || !frame->extended_data[plane])
  406. return NULL;
  407. data = frame->extended_data[plane];
  408. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++) {
  409. AVBufferRef *buf = frame->buf[i];
  410. if (data >= buf->data && data < buf->data + buf->size)
  411. return buf;
  412. }
  413. for (i = 0; i < frame->nb_extended_buf; i++) {
  414. AVBufferRef *buf = frame->extended_buf[i];
  415. if (data >= buf->data && data < buf->data + buf->size)
  416. return buf;
  417. }
  418. return NULL;
  419. }
  420. AVFrameSideData *av_frame_new_side_data(AVFrame *frame,
  421. enum AVFrameSideDataType type,
  422. int size)
  423. {
  424. AVFrameSideData *ret, **tmp;
  425. if (frame->nb_side_data > INT_MAX / sizeof(*frame->side_data) - 1)
  426. return NULL;
  427. tmp = av_realloc(frame->side_data,
  428. (frame->nb_side_data + 1) * sizeof(*frame->side_data));
  429. if (!tmp)
  430. return NULL;
  431. frame->side_data = tmp;
  432. ret = av_mallocz(sizeof(*ret));
  433. if (!ret)
  434. return NULL;
  435. ret->data = av_malloc(size);
  436. if (!ret->data) {
  437. av_freep(&ret);
  438. return NULL;
  439. }
  440. ret->size = size;
  441. ret->type = type;
  442. frame->side_data[frame->nb_side_data++] = ret;
  443. return ret;
  444. }
  445. AVFrameSideData *av_frame_get_side_data(AVFrame *frame,
  446. enum AVFrameSideDataType type)
  447. {
  448. int i;
  449. for (i = 0; i < frame->nb_side_data; i++) {
  450. if (frame->side_data[i]->type == type)
  451. return frame->side_data[i];
  452. }
  453. return NULL;
  454. }