You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

410 lines
13KB

  1. /*
  2. * RTP JPEG-compressed Video Depacketizer, RFC 2435
  3. * Copyright (c) 2012 Samuel Pitoiset
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include "avformat.h"
  22. #include "rtpdec.h"
  23. #include "rtpdec_formats.h"
  24. #include "libavutil/intreadwrite.h"
  25. #include "libavcodec/mjpeg.h"
  26. #include "libavcodec/bytestream.h"
  27. /**
  28. * RTP/JPEG specific private data.
  29. */
  30. struct PayloadContext {
  31. AVIOContext *frame; ///< current frame buffer
  32. uint32_t timestamp; ///< current frame timestamp
  33. int hdr_size; ///< size of the current frame header
  34. uint8_t qtables[128][128];
  35. uint8_t qtables_len[128];
  36. };
  37. static const uint8_t default_quantizers[128] = {
  38. /* luma table */
  39. 16, 11, 12, 14, 12, 10, 16, 14,
  40. 13, 14, 18, 17, 16, 19, 24, 40,
  41. 26, 24, 22, 22, 24, 49, 35, 37,
  42. 29, 40, 58, 51, 61, 60, 57, 51,
  43. 56, 55, 64, 72, 92, 78, 64, 68,
  44. 87, 69, 55, 56, 80, 109, 81, 87,
  45. 95, 98, 103, 104, 103, 62, 77, 113,
  46. 121, 112, 100, 120, 92, 101, 103, 99,
  47. /* chroma table */
  48. 17, 18, 18, 24, 21, 24, 47, 26,
  49. 26, 47, 99, 66, 56, 66, 99, 99,
  50. 99, 99, 99, 99, 99, 99, 99, 99,
  51. 99, 99, 99, 99, 99, 99, 99, 99,
  52. 99, 99, 99, 99, 99, 99, 99, 99,
  53. 99, 99, 99, 99, 99, 99, 99, 99,
  54. 99, 99, 99, 99, 99, 99, 99, 99,
  55. 99, 99, 99, 99, 99, 99, 99, 99
  56. };
  57. static PayloadContext *jpeg_new_context(void)
  58. {
  59. return av_mallocz(sizeof(PayloadContext));
  60. }
  61. static inline void free_frame_if_needed(PayloadContext *jpeg)
  62. {
  63. if (jpeg->frame) {
  64. uint8_t *p;
  65. avio_close_dyn_buf(jpeg->frame, &p);
  66. av_free(p);
  67. jpeg->frame = NULL;
  68. }
  69. }
  70. static void jpeg_free_context(PayloadContext *jpeg)
  71. {
  72. free_frame_if_needed(jpeg);
  73. av_free(jpeg);
  74. }
  75. static int jpeg_create_huffman_table(PutByteContext *p, int table_class,
  76. int table_id, const uint8_t *bits_table,
  77. const uint8_t *value_table)
  78. {
  79. int i, n = 0;
  80. bytestream2_put_byte(p, table_class << 4 | table_id);
  81. for (i = 1; i <= 16; i++) {
  82. n += bits_table[i];
  83. bytestream2_put_byte(p, bits_table[i]);
  84. }
  85. for (i = 0; i < n; i++) {
  86. bytestream2_put_byte(p, value_table[i]);
  87. }
  88. return n + 17;
  89. }
  90. static void jpeg_put_marker(PutByteContext *pbc, int code)
  91. {
  92. bytestream2_put_byte(pbc, 0xff);
  93. bytestream2_put_byte(pbc, code);
  94. }
  95. static int jpeg_create_header(uint8_t *buf, int size, uint32_t type, uint32_t w,
  96. uint32_t h, const uint8_t *qtable, int nb_qtable,
  97. int dri)
  98. {
  99. PutByteContext pbc;
  100. uint8_t *dht_size_ptr;
  101. int dht_size, i;
  102. bytestream2_init_writer(&pbc, buf, size);
  103. /* Convert from blocks to pixels. */
  104. w <<= 3;
  105. h <<= 3;
  106. /* SOI */
  107. jpeg_put_marker(&pbc, SOI);
  108. /* JFIF header */
  109. jpeg_put_marker(&pbc, APP0);
  110. bytestream2_put_be16(&pbc, 16);
  111. bytestream2_put_buffer(&pbc, "JFIF", 5);
  112. bytestream2_put_be16(&pbc, 0x0201);
  113. bytestream2_put_byte(&pbc, 0);
  114. bytestream2_put_be16(&pbc, 1);
  115. bytestream2_put_be16(&pbc, 1);
  116. bytestream2_put_byte(&pbc, 0);
  117. bytestream2_put_byte(&pbc, 0);
  118. if (dri) {
  119. jpeg_put_marker(&pbc, DRI);
  120. bytestream2_put_be16(&pbc, 4);
  121. bytestream2_put_be16(&pbc, dri);
  122. }
  123. /* DQT */
  124. jpeg_put_marker(&pbc, DQT);
  125. bytestream2_put_be16(&pbc, 2 + nb_qtable * (1 + 64));
  126. for (i = 0; i < nb_qtable; i++) {
  127. bytestream2_put_byte(&pbc, i);
  128. /* Each table is an array of 64 values given in zig-zag
  129. * order, identical to the format used in a JFIF DQT
  130. * marker segment. */
  131. bytestream2_put_buffer(&pbc, qtable + 64 * i, 64);
  132. }
  133. /* DHT */
  134. jpeg_put_marker(&pbc, DHT);
  135. dht_size_ptr = pbc.buffer;
  136. bytestream2_put_be16(&pbc, 0);
  137. dht_size = 2;
  138. dht_size += jpeg_create_huffman_table(&pbc, 0, 0,avpriv_mjpeg_bits_dc_luminance,
  139. avpriv_mjpeg_val_dc);
  140. dht_size += jpeg_create_huffman_table(&pbc, 0, 1, avpriv_mjpeg_bits_dc_chrominance,
  141. avpriv_mjpeg_val_dc);
  142. dht_size += jpeg_create_huffman_table(&pbc, 1, 0, avpriv_mjpeg_bits_ac_luminance,
  143. avpriv_mjpeg_val_ac_luminance);
  144. dht_size += jpeg_create_huffman_table(&pbc, 1, 1, avpriv_mjpeg_bits_ac_chrominance,
  145. avpriv_mjpeg_val_ac_chrominance);
  146. AV_WB16(dht_size_ptr, dht_size);
  147. /* SOF0 */
  148. jpeg_put_marker(&pbc, SOF0);
  149. bytestream2_put_be16(&pbc, 17); /* size */
  150. bytestream2_put_byte(&pbc, 8); /* bits per component */
  151. bytestream2_put_be16(&pbc, h);
  152. bytestream2_put_be16(&pbc, w);
  153. bytestream2_put_byte(&pbc, 3); /* number of components */
  154. bytestream2_put_byte(&pbc, 1); /* component number */
  155. bytestream2_put_byte(&pbc, (2 << 4) | (type ? 2 : 1)); /* hsample/vsample */
  156. bytestream2_put_byte(&pbc, 0); /* matrix number */
  157. bytestream2_put_byte(&pbc, 2); /* component number */
  158. bytestream2_put_byte(&pbc, 1 << 4 | 1); /* hsample/vsample */
  159. bytestream2_put_byte(&pbc, nb_qtable == 2 ? 1 : 0); /* matrix number */
  160. bytestream2_put_byte(&pbc, 3); /* component number */
  161. bytestream2_put_byte(&pbc, 1 << 4 | 1); /* hsample/vsample */
  162. bytestream2_put_byte(&pbc, nb_qtable == 2 ? 1 : 0); /* matrix number */
  163. /* SOS */
  164. jpeg_put_marker(&pbc, SOS);
  165. bytestream2_put_be16(&pbc, 12);
  166. bytestream2_put_byte(&pbc, 3);
  167. bytestream2_put_byte(&pbc, 1);
  168. bytestream2_put_byte(&pbc, 0);
  169. bytestream2_put_byte(&pbc, 2);
  170. bytestream2_put_byte(&pbc, 17);
  171. bytestream2_put_byte(&pbc, 3);
  172. bytestream2_put_byte(&pbc, 17);
  173. bytestream2_put_byte(&pbc, 0);
  174. bytestream2_put_byte(&pbc, 63);
  175. bytestream2_put_byte(&pbc, 0);
  176. /* Return the length in bytes of the JPEG header. */
  177. return bytestream2_tell_p(&pbc);
  178. }
  179. static void create_default_qtables(uint8_t *qtables, uint8_t q)
  180. {
  181. int factor = q;
  182. int i;
  183. factor = av_clip(q, 1, 99);
  184. if (q < 50)
  185. q = 5000 / factor;
  186. else
  187. q = 200 - factor * 2;
  188. for (i = 0; i < 128; i++) {
  189. int val = (default_quantizers[i] * q + 50) / 100;
  190. /* Limit the quantizers to 1 <= q <= 255. */
  191. val = av_clip(val, 1, 255);
  192. qtables[i] = val;
  193. }
  194. }
  195. static int jpeg_parse_packet(AVFormatContext *ctx, PayloadContext *jpeg,
  196. AVStream *st, AVPacket *pkt, uint32_t *timestamp,
  197. const uint8_t *buf, int len, uint16_t seq,
  198. int flags)
  199. {
  200. uint8_t type, q, width, height;
  201. const uint8_t *qtables = NULL;
  202. uint16_t qtable_len;
  203. uint32_t off;
  204. int ret, dri = 0;
  205. if (len < 8) {
  206. av_log(ctx, AV_LOG_ERROR, "Too short RTP/JPEG packet.\n");
  207. return AVERROR_INVALIDDATA;
  208. }
  209. /* Parse the main JPEG header. */
  210. off = AV_RB24(buf + 1); /* fragment byte offset */
  211. type = AV_RB8(buf + 4); /* id of jpeg decoder params */
  212. q = AV_RB8(buf + 5); /* quantization factor (or table id) */
  213. width = AV_RB8(buf + 6); /* frame width in 8 pixel blocks */
  214. height = AV_RB8(buf + 7); /* frame height in 8 pixel blocks */
  215. buf += 8;
  216. len -= 8;
  217. if (type & 0x40) {
  218. if (len < 4) {
  219. av_log(ctx, AV_LOG_ERROR, "Too short RTP/JPEG packet.\n");
  220. return AVERROR_INVALIDDATA;
  221. }
  222. dri = AV_RB16(buf);
  223. buf += 4;
  224. len -= 4;
  225. type &= ~0x40;
  226. }
  227. /* Parse the restart marker header. */
  228. if (type > 63) {
  229. av_log(ctx, AV_LOG_ERROR,
  230. "Unimplemented RTP/JPEG restart marker header.\n");
  231. return AVERROR_PATCHWELCOME;
  232. }
  233. if (type > 1) {
  234. av_log(ctx, AV_LOG_ERROR, "Unimplemented RTP/JPEG type %d\n", type);
  235. return AVERROR_PATCHWELCOME;
  236. }
  237. /* Parse the quantization table header. */
  238. if (off == 0) {
  239. /* Start of JPEG data packet. */
  240. uint8_t new_qtables[128];
  241. uint8_t hdr[1024];
  242. if (q > 127) {
  243. uint8_t precision;
  244. if (len < 4) {
  245. av_log(ctx, AV_LOG_ERROR, "Too short RTP/JPEG packet.\n");
  246. return AVERROR_INVALIDDATA;
  247. }
  248. /* The first byte is reserved for future use. */
  249. precision = AV_RB8(buf + 1); /* size of coefficients */
  250. qtable_len = AV_RB16(buf + 2); /* length in bytes */
  251. buf += 4;
  252. len -= 4;
  253. if (precision)
  254. av_log(ctx, AV_LOG_WARNING, "Only 8-bit precision is supported.\n");
  255. if (qtable_len > 0) {
  256. if (len < qtable_len) {
  257. av_log(ctx, AV_LOG_ERROR, "Too short RTP/JPEG packet.\n");
  258. return AVERROR_INVALIDDATA;
  259. }
  260. qtables = buf;
  261. buf += qtable_len;
  262. len -= qtable_len;
  263. if (q < 255) {
  264. if (jpeg->qtables_len[q - 128] &&
  265. (jpeg->qtables_len[q - 128] != qtable_len ||
  266. memcmp(qtables, &jpeg->qtables[q - 128][0], qtable_len))) {
  267. av_log(ctx, AV_LOG_WARNING,
  268. "Quantization tables for q=%d changed\n", q);
  269. } else if (!jpeg->qtables_len[q - 128] && qtable_len <= 128) {
  270. memcpy(&jpeg->qtables[q - 128][0], qtables,
  271. qtable_len);
  272. jpeg->qtables_len[q - 128] = qtable_len;
  273. }
  274. }
  275. } else {
  276. if (q == 255) {
  277. av_log(ctx, AV_LOG_ERROR,
  278. "Invalid RTP/JPEG packet. Quantization tables not found.\n");
  279. return AVERROR_INVALIDDATA;
  280. }
  281. if (!jpeg->qtables_len[q - 128]) {
  282. av_log(ctx, AV_LOG_ERROR,
  283. "No quantization tables known for q=%d yet.\n", q);
  284. return AVERROR_INVALIDDATA;
  285. }
  286. qtables = &jpeg->qtables[q - 128][0];
  287. qtable_len = jpeg->qtables_len[q - 128];
  288. }
  289. } else { /* q <= 127 */
  290. if (q == 0 || q > 99) {
  291. av_log(ctx, AV_LOG_ERROR, "Reserved q value %d\n", q);
  292. return AVERROR_INVALIDDATA;
  293. }
  294. create_default_qtables(new_qtables, q);
  295. qtables = new_qtables;
  296. qtable_len = sizeof(new_qtables);
  297. }
  298. /* Skip the current frame in case of the end packet
  299. * has been lost somewhere. */
  300. free_frame_if_needed(jpeg);
  301. if ((ret = avio_open_dyn_buf(&jpeg->frame)) < 0)
  302. return ret;
  303. jpeg->timestamp = *timestamp;
  304. /* Generate a frame and scan headers that can be prepended to the
  305. * RTP/JPEG data payload to produce a JPEG compressed image in
  306. * interchange format. */
  307. jpeg->hdr_size = jpeg_create_header(hdr, sizeof(hdr), type, width,
  308. height, qtables,
  309. qtable_len / 64, dri);
  310. /* Copy JPEG header to frame buffer. */
  311. avio_write(jpeg->frame, hdr, jpeg->hdr_size);
  312. }
  313. if (!jpeg->frame) {
  314. av_log(ctx, AV_LOG_ERROR,
  315. "Received packet without a start chunk; dropping frame.\n");
  316. return AVERROR(EAGAIN);
  317. }
  318. if (jpeg->timestamp != *timestamp) {
  319. /* Skip the current frame if timestamp is incorrect.
  320. * A start packet has been lost somewhere. */
  321. free_frame_if_needed(jpeg);
  322. av_log(ctx, AV_LOG_ERROR, "RTP timestamps don't match.\n");
  323. return AVERROR_INVALIDDATA;
  324. }
  325. if (off != avio_tell(jpeg->frame) - jpeg->hdr_size) {
  326. av_log(ctx, AV_LOG_ERROR,
  327. "Missing packets; dropping frame.\n");
  328. return AVERROR(EAGAIN);
  329. }
  330. /* Copy data to frame buffer. */
  331. avio_write(jpeg->frame, buf, len);
  332. if (flags & RTP_FLAG_MARKER) {
  333. /* End of JPEG data packet. */
  334. uint8_t buf[2] = { 0xff, EOI };
  335. /* Put EOI marker. */
  336. avio_write(jpeg->frame, buf, sizeof(buf));
  337. /* Prepare the JPEG packet. */
  338. if ((ret = ff_rtp_finalize_packet(pkt, &jpeg->frame, st->index)) < 0) {
  339. av_log(ctx, AV_LOG_ERROR,
  340. "Error occurred when getting frame buffer.\n");
  341. return ret;
  342. }
  343. return 0;
  344. }
  345. return AVERROR(EAGAIN);
  346. }
  347. RTPDynamicProtocolHandler ff_jpeg_dynamic_handler = {
  348. .enc_name = "JPEG",
  349. .codec_type = AVMEDIA_TYPE_VIDEO,
  350. .codec_id = AV_CODEC_ID_MJPEG,
  351. .alloc = jpeg_new_context,
  352. .free = jpeg_free_context,
  353. .parse_packet = jpeg_parse_packet,
  354. .static_payload_id = 26,
  355. };