You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

571 lines
17KB

  1. /*
  2. *
  3. * This file is part of FFmpeg.
  4. *
  5. * FFmpeg is free software; you can redistribute it and/or
  6. * modify it under the terms of the GNU Lesser General Public
  7. * License as published by the Free Software Foundation; either
  8. * version 2.1 of the License, or (at your option) any later version.
  9. *
  10. * FFmpeg is distributed in the hope that it will be useful,
  11. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  12. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  13. * Lesser General Public License for more details.
  14. *
  15. * You should have received a copy of the GNU Lesser General Public
  16. * License along with FFmpeg; if not, write to the Free Software
  17. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  18. */
  19. #include "channel_layout.h"
  20. #include "avassert.h"
  21. #include "buffer.h"
  22. #include "common.h"
  23. #include "dict.h"
  24. #include "frame.h"
  25. #include "imgutils.h"
  26. #include "mem.h"
  27. #include "samplefmt.h"
  28. #define MAKE_ACCESSORS(str, name, type, field) \
  29. type av_##name##_get_##field(const str *s) { return s->field; } \
  30. void av_##name##_set_##field(str *s, type v) { s->field = v; }
  31. MAKE_ACCESSORS(AVFrame, frame, int64_t, best_effort_timestamp)
  32. MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_duration)
  33. MAKE_ACCESSORS(AVFrame, frame, int64_t, pkt_pos)
  34. MAKE_ACCESSORS(AVFrame, frame, int64_t, channel_layout)
  35. MAKE_ACCESSORS(AVFrame, frame, int, channels)
  36. MAKE_ACCESSORS(AVFrame, frame, int, sample_rate)
  37. MAKE_ACCESSORS(AVFrame, frame, AVDictionary *, metadata)
  38. MAKE_ACCESSORS(AVFrame, frame, int, decode_error_flags)
  39. MAKE_ACCESSORS(AVFrame, frame, int, pkt_size)
  40. #define CHECK_CHANNELS_CONSISTENCY(frame) \
  41. av_assert2(!(frame)->channel_layout || \
  42. (frame)->channels == \
  43. av_get_channel_layout_nb_channels((frame)->channel_layout))
  44. AVDictionary **avpriv_frame_get_metadatap(AVFrame *frame) {return &frame->metadata;};
  45. int av_frame_set_qp_table(AVFrame *f, AVBufferRef *buf, int stride, int qp_type)
  46. {
  47. av_buffer_unref(&f->qp_table_buf);
  48. f->qp_table_buf = buf;
  49. f->qscale_table = buf->data;
  50. f->qstride = stride;
  51. f->qscale_type = qp_type;
  52. return 0;
  53. }
  54. int8_t *av_frame_get_qp_table(AVFrame *f, int *stride, int *type)
  55. {
  56. *stride = f->qstride;
  57. *type = f->qscale_type;
  58. if (!f->qp_table_buf)
  59. return NULL;
  60. return f->qp_table_buf->data;
  61. }
  62. static void get_frame_defaults(AVFrame *frame)
  63. {
  64. if (frame->extended_data != frame->data)
  65. av_freep(&frame->extended_data);
  66. memset(frame, 0, sizeof(*frame));
  67. frame->pts =
  68. frame->pkt_dts =
  69. frame->pkt_pts = AV_NOPTS_VALUE;
  70. av_frame_set_best_effort_timestamp(frame, AV_NOPTS_VALUE);
  71. av_frame_set_pkt_duration (frame, 0);
  72. av_frame_set_pkt_pos (frame, -1);
  73. av_frame_set_pkt_size (frame, -1);
  74. frame->key_frame = 1;
  75. frame->sample_aspect_ratio = (AVRational){ 0, 1 };
  76. frame->format = -1; /* unknown */
  77. frame->extended_data = frame->data;
  78. }
  79. AVFrame *av_frame_alloc(void)
  80. {
  81. AVFrame *frame = av_mallocz(sizeof(*frame));
  82. if (!frame)
  83. return NULL;
  84. frame->extended_data = NULL;
  85. get_frame_defaults(frame);
  86. return frame;
  87. }
  88. void av_frame_free(AVFrame **frame)
  89. {
  90. if (!frame || !*frame)
  91. return;
  92. av_frame_unref(*frame);
  93. av_freep(frame);
  94. }
  95. static int get_video_buffer(AVFrame *frame, int align)
  96. {
  97. const AVPixFmtDescriptor *desc = av_pix_fmt_desc_get(frame->format);
  98. int ret, i;
  99. if (!desc)
  100. return AVERROR(EINVAL);
  101. if ((ret = av_image_check_size(frame->width, frame->height, 0, NULL)) < 0)
  102. return ret;
  103. if (!frame->linesize[0]) {
  104. for(i=1; i<=align; i+=i) {
  105. ret = av_image_fill_linesizes(frame->linesize, frame->format,
  106. FFALIGN(frame->width, i));
  107. if (ret < 0)
  108. return ret;
  109. if (!(frame->linesize[0] & (align-1)))
  110. break;
  111. }
  112. for (i = 0; i < 4 && frame->linesize[i]; i++)
  113. frame->linesize[i] = FFALIGN(frame->linesize[i], align);
  114. }
  115. for (i = 0; i < 4 && frame->linesize[i]; i++) {
  116. int h = FFALIGN(frame->height, 32);
  117. if (i == 1 || i == 2)
  118. h = FF_CEIL_RSHIFT(h, desc->log2_chroma_h);
  119. frame->buf[i] = av_buffer_alloc(frame->linesize[i] * h + 16);
  120. if (!frame->buf[i])
  121. goto fail;
  122. frame->data[i] = frame->buf[i]->data;
  123. }
  124. if (desc->flags & AV_PIX_FMT_FLAG_PAL || desc->flags & AV_PIX_FMT_FLAG_PSEUDOPAL) {
  125. av_buffer_unref(&frame->buf[1]);
  126. frame->buf[1] = av_buffer_alloc(1024);
  127. if (!frame->buf[1])
  128. goto fail;
  129. frame->data[1] = frame->buf[1]->data;
  130. }
  131. frame->extended_data = frame->data;
  132. return 0;
  133. fail:
  134. av_frame_unref(frame);
  135. return AVERROR(ENOMEM);
  136. }
  137. static int get_audio_buffer(AVFrame *frame, int align)
  138. {
  139. int channels = frame->channels;
  140. int planar = av_sample_fmt_is_planar(frame->format);
  141. int planes = planar ? channels : 1;
  142. int ret, i;
  143. CHECK_CHANNELS_CONSISTENCY(frame);
  144. if (!frame->linesize[0]) {
  145. ret = av_samples_get_buffer_size(&frame->linesize[0], channels,
  146. frame->nb_samples, frame->format,
  147. align);
  148. if (ret < 0)
  149. return ret;
  150. }
  151. if (planes > AV_NUM_DATA_POINTERS) {
  152. frame->extended_data = av_mallocz(planes *
  153. sizeof(*frame->extended_data));
  154. frame->extended_buf = av_mallocz((planes - AV_NUM_DATA_POINTERS) *
  155. sizeof(*frame->extended_buf));
  156. if (!frame->extended_data || !frame->extended_buf) {
  157. av_freep(&frame->extended_data);
  158. av_freep(&frame->extended_buf);
  159. return AVERROR(ENOMEM);
  160. }
  161. frame->nb_extended_buf = planes - AV_NUM_DATA_POINTERS;
  162. } else
  163. frame->extended_data = frame->data;
  164. for (i = 0; i < FFMIN(planes, AV_NUM_DATA_POINTERS); i++) {
  165. frame->buf[i] = av_buffer_alloc(frame->linesize[0]);
  166. if (!frame->buf[i]) {
  167. av_frame_unref(frame);
  168. return AVERROR(ENOMEM);
  169. }
  170. frame->extended_data[i] = frame->data[i] = frame->buf[i]->data;
  171. }
  172. for (i = 0; i < planes - AV_NUM_DATA_POINTERS; i++) {
  173. frame->extended_buf[i] = av_buffer_alloc(frame->linesize[0]);
  174. if (!frame->extended_buf[i]) {
  175. av_frame_unref(frame);
  176. return AVERROR(ENOMEM);
  177. }
  178. frame->extended_data[i + AV_NUM_DATA_POINTERS] = frame->extended_buf[i]->data;
  179. }
  180. return 0;
  181. }
  182. int av_frame_get_buffer(AVFrame *frame, int align)
  183. {
  184. if (frame->format < 0)
  185. return AVERROR(EINVAL);
  186. if (frame->width > 0 && frame->height > 0)
  187. return get_video_buffer(frame, align);
  188. else if (frame->nb_samples > 0 && frame->channel_layout)
  189. return get_audio_buffer(frame, align);
  190. return AVERROR(EINVAL);
  191. }
  192. int av_frame_ref(AVFrame *dst, AVFrame *src)
  193. {
  194. int i, ret = 0;
  195. dst->format = src->format;
  196. dst->width = src->width;
  197. dst->height = src->height;
  198. dst->channels = src->channels;
  199. dst->channel_layout = src->channel_layout;
  200. dst->nb_samples = src->nb_samples;
  201. ret = av_frame_copy_props(dst, src);
  202. if (ret < 0)
  203. return ret;
  204. /* duplicate the frame data if it's not refcounted */
  205. if (!src->buf[0]) {
  206. ret = av_frame_get_buffer(dst, 32);
  207. if (ret < 0)
  208. return ret;
  209. if (src->nb_samples) {
  210. int ch = src->channels;
  211. CHECK_CHANNELS_CONSISTENCY(src);
  212. av_samples_copy(dst->extended_data, src->extended_data, 0, 0,
  213. dst->nb_samples, ch, dst->format);
  214. } else {
  215. av_image_copy(dst->data, dst->linesize, src->data, src->linesize,
  216. dst->format, dst->width, dst->height);
  217. }
  218. return 0;
  219. }
  220. /* ref the buffers */
  221. for (i = 0; i < FF_ARRAY_ELEMS(src->buf); i++) {
  222. if (!src->buf[i])
  223. continue;
  224. dst->buf[i] = av_buffer_ref(src->buf[i]);
  225. if (!dst->buf[i]) {
  226. ret = AVERROR(ENOMEM);
  227. goto fail;
  228. }
  229. }
  230. if (src->extended_buf) {
  231. dst->extended_buf = av_mallocz(sizeof(*dst->extended_buf) *
  232. src->nb_extended_buf);
  233. if (!dst->extended_buf) {
  234. ret = AVERROR(ENOMEM);
  235. goto fail;
  236. }
  237. dst->nb_extended_buf = src->nb_extended_buf;
  238. for (i = 0; i < src->nb_extended_buf; i++) {
  239. dst->extended_buf[i] = av_buffer_ref(src->extended_buf[i]);
  240. if (!dst->extended_buf[i]) {
  241. ret = AVERROR(ENOMEM);
  242. goto fail;
  243. }
  244. }
  245. }
  246. /* duplicate extended data */
  247. if (src->extended_data != src->data) {
  248. int ch = src->channels;
  249. if (!ch) {
  250. ret = AVERROR(EINVAL);
  251. goto fail;
  252. }
  253. CHECK_CHANNELS_CONSISTENCY(src);
  254. dst->extended_data = av_malloc(sizeof(*dst->extended_data) * ch);
  255. if (!dst->extended_data) {
  256. ret = AVERROR(ENOMEM);
  257. goto fail;
  258. }
  259. memcpy(dst->extended_data, src->extended_data, sizeof(*src->extended_data) * ch);
  260. } else
  261. dst->extended_data = dst->data;
  262. memcpy(dst->data, src->data, sizeof(src->data));
  263. memcpy(dst->linesize, src->linesize, sizeof(src->linesize));
  264. return 0;
  265. fail:
  266. av_frame_unref(dst);
  267. return ret;
  268. }
  269. AVFrame *av_frame_clone(AVFrame *src)
  270. {
  271. AVFrame *ret = av_frame_alloc();
  272. if (!ret)
  273. return NULL;
  274. if (av_frame_ref(ret, src) < 0)
  275. av_frame_free(&ret);
  276. return ret;
  277. }
  278. void av_frame_unref(AVFrame *frame)
  279. {
  280. int i;
  281. for (i = 0; i < frame->nb_side_data; i++) {
  282. av_freep(&frame->side_data[i]->data);
  283. av_dict_free(&frame->side_data[i]->metadata);
  284. av_freep(&frame->side_data[i]);
  285. }
  286. av_freep(&frame->side_data);
  287. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
  288. av_buffer_unref(&frame->buf[i]);
  289. for (i = 0; i < frame->nb_extended_buf; i++)
  290. av_buffer_unref(&frame->extended_buf[i]);
  291. av_freep(&frame->extended_buf);
  292. av_dict_free(&frame->metadata);
  293. av_buffer_unref(&frame->qp_table_buf);
  294. get_frame_defaults(frame);
  295. }
  296. void av_frame_move_ref(AVFrame *dst, AVFrame *src)
  297. {
  298. *dst = *src;
  299. if (src->extended_data == src->data)
  300. dst->extended_data = dst->data;
  301. memset(src, 0, sizeof(*src));
  302. get_frame_defaults(src);
  303. }
  304. int av_frame_is_writable(AVFrame *frame)
  305. {
  306. int i, ret = 1;
  307. /* assume non-refcounted frames are not writable */
  308. if (!frame->buf[0])
  309. return 0;
  310. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf); i++)
  311. if (frame->buf[i])
  312. ret &= !!av_buffer_is_writable(frame->buf[i]);
  313. for (i = 0; i < frame->nb_extended_buf; i++)
  314. ret &= !!av_buffer_is_writable(frame->extended_buf[i]);
  315. return ret;
  316. }
  317. int av_frame_make_writable(AVFrame *frame)
  318. {
  319. AVFrame tmp;
  320. int ret;
  321. if (!frame->buf[0])
  322. return AVERROR(EINVAL);
  323. if (av_frame_is_writable(frame))
  324. return 0;
  325. memset(&tmp, 0, sizeof(tmp));
  326. tmp.format = frame->format;
  327. tmp.width = frame->width;
  328. tmp.height = frame->height;
  329. tmp.channels = frame->channels;
  330. tmp.channel_layout = frame->channel_layout;
  331. tmp.nb_samples = frame->nb_samples;
  332. ret = av_frame_get_buffer(&tmp, 32);
  333. if (ret < 0)
  334. return ret;
  335. if (tmp.nb_samples) {
  336. int ch = tmp.channels;
  337. CHECK_CHANNELS_CONSISTENCY(&tmp);
  338. av_samples_copy(tmp.extended_data, frame->extended_data, 0, 0,
  339. frame->nb_samples, ch, frame->format);
  340. } else {
  341. av_image_copy(tmp.data, tmp.linesize, frame->data, frame->linesize,
  342. frame->format, frame->width, frame->height);
  343. }
  344. ret = av_frame_copy_props(&tmp, frame);
  345. if (ret < 0) {
  346. av_frame_unref(&tmp);
  347. return ret;
  348. }
  349. av_frame_unref(frame);
  350. *frame = tmp;
  351. if (tmp.data == tmp.extended_data)
  352. frame->extended_data = frame->data;
  353. return 0;
  354. }
  355. int av_frame_copy_props(AVFrame *dst, const AVFrame *src)
  356. {
  357. int i;
  358. dst->key_frame = src->key_frame;
  359. dst->pict_type = src->pict_type;
  360. dst->sample_aspect_ratio = src->sample_aspect_ratio;
  361. dst->pts = src->pts;
  362. dst->repeat_pict = src->repeat_pict;
  363. dst->interlaced_frame = src->interlaced_frame;
  364. dst->top_field_first = src->top_field_first;
  365. dst->palette_has_changed = src->palette_has_changed;
  366. dst->sample_rate = src->sample_rate;
  367. dst->opaque = src->opaque;
  368. #if FF_API_AVFRAME_LAVC
  369. dst->type = src->type;
  370. #endif
  371. dst->pkt_pts = src->pkt_pts;
  372. dst->pkt_dts = src->pkt_dts;
  373. dst->pkt_pos = src->pkt_pos;
  374. dst->pkt_size = src->pkt_size;
  375. dst->pkt_duration = src->pkt_duration;
  376. dst->reordered_opaque = src->reordered_opaque;
  377. dst->quality = src->quality;
  378. dst->best_effort_timestamp = src->best_effort_timestamp;
  379. dst->coded_picture_number = src->coded_picture_number;
  380. dst->display_picture_number = src->display_picture_number;
  381. dst->decode_error_flags = src->decode_error_flags;
  382. av_dict_copy(&dst->metadata, src->metadata, 0);
  383. memcpy(dst->error, src->error, sizeof(dst->error));
  384. for (i = 0; i < src->nb_side_data; i++) {
  385. const AVFrameSideData *sd_src = src->side_data[i];
  386. AVFrameSideData *sd_dst = av_frame_new_side_data(dst, sd_src->type,
  387. sd_src->size);
  388. if (!sd_dst) {
  389. for (i = 0; i < dst->nb_side_data; i++) {
  390. av_freep(&dst->side_data[i]->data);
  391. av_freep(&dst->side_data[i]);
  392. av_dict_free(&dst->side_data[i]->metadata);
  393. }
  394. av_freep(&dst->side_data);
  395. return AVERROR(ENOMEM);
  396. }
  397. memcpy(sd_dst->data, sd_src->data, sd_src->size);
  398. av_dict_copy(&sd_dst->metadata, sd_src->metadata, 0);
  399. }
  400. dst->qscale_table = NULL;
  401. dst->qstride = 0;
  402. dst->qscale_type = 0;
  403. if (src->qp_table_buf) {
  404. dst->qp_table_buf = av_buffer_ref(src->qp_table_buf);
  405. if (dst->qp_table_buf) {
  406. dst->qscale_table = dst->qp_table_buf->data;
  407. dst->qstride = src->qstride;
  408. dst->qscale_type = src->qscale_type;
  409. }
  410. }
  411. return 0;
  412. }
  413. AVBufferRef *av_frame_get_plane_buffer(AVFrame *frame, int plane)
  414. {
  415. uint8_t *data;
  416. int planes, i;
  417. if (frame->nb_samples) {
  418. int channels = frame->channels;
  419. if (!channels)
  420. return NULL;
  421. CHECK_CHANNELS_CONSISTENCY(frame);
  422. planes = av_sample_fmt_is_planar(frame->format) ? channels : 1;
  423. } else
  424. planes = 4;
  425. if (plane < 0 || plane >= planes || !frame->extended_data[plane])
  426. return NULL;
  427. data = frame->extended_data[plane];
  428. for (i = 0; i < FF_ARRAY_ELEMS(frame->buf) && frame->buf[i]; i++) {
  429. AVBufferRef *buf = frame->buf[i];
  430. if (data >= buf->data && data < buf->data + buf->size)
  431. return buf;
  432. }
  433. for (i = 0; i < frame->nb_extended_buf; i++) {
  434. AVBufferRef *buf = frame->extended_buf[i];
  435. if (data >= buf->data && data < buf->data + buf->size)
  436. return buf;
  437. }
  438. return NULL;
  439. }
  440. AVFrameSideData *av_frame_new_side_data(AVFrame *frame,
  441. enum AVFrameSideDataType type,
  442. int size)
  443. {
  444. AVFrameSideData *ret, **tmp;
  445. if (frame->nb_side_data > INT_MAX / sizeof(*frame->side_data) - 1)
  446. return NULL;
  447. tmp = av_realloc(frame->side_data,
  448. (frame->nb_side_data + 1) * sizeof(*frame->side_data));
  449. if (!tmp)
  450. return NULL;
  451. frame->side_data = tmp;
  452. ret = av_mallocz(sizeof(*ret));
  453. if (!ret)
  454. return NULL;
  455. ret->data = av_malloc(size);
  456. if (!ret->data) {
  457. av_freep(&ret);
  458. return NULL;
  459. }
  460. ret->size = size;
  461. ret->type = type;
  462. frame->side_data[frame->nb_side_data++] = ret;
  463. return ret;
  464. }
  465. AVFrameSideData *av_frame_get_side_data(AVFrame *frame,
  466. enum AVFrameSideDataType type)
  467. {
  468. int i;
  469. for (i = 0; i < frame->nb_side_data; i++) {
  470. if (frame->side_data[i]->type == type)
  471. return frame->side_data[i];
  472. }
  473. return NULL;
  474. }