You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

533 lines
16KB

  1. /*
  2. * G.726 ADPCM audio codec
  3. * Copyright (c) 2004 Roman Shaposhnik
  4. *
  5. * This is a very straightforward rendition of the G.726
  6. * Section 4 "Computational Details".
  7. *
  8. * This file is part of FFmpeg.
  9. *
  10. * FFmpeg is free software; you can redistribute it and/or
  11. * modify it under the terms of the GNU Lesser General Public
  12. * License as published by the Free Software Foundation; either
  13. * version 2.1 of the License, or (at your option) any later version.
  14. *
  15. * FFmpeg is distributed in the hope that it will be useful,
  16. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  17. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  18. * Lesser General Public License for more details.
  19. *
  20. * You should have received a copy of the GNU Lesser General Public
  21. * License along with FFmpeg; if not, write to the Free Software
  22. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  23. */
  24. #include <limits.h>
  25. #include "libavutil/channel_layout.h"
  26. #include "libavutil/opt.h"
  27. #include "avcodec.h"
  28. #include "internal.h"
  29. #include "get_bits.h"
  30. #include "put_bits.h"
  31. /**
  32. * G.726 11-bit float.
  33. * G.726 Standard uses rather odd 11-bit floating point arithmetic for
  34. * numerous occasions. It's a mystery to me why they did it this way
  35. * instead of simply using 32-bit integer arithmetic.
  36. */
  37. typedef struct Float11 {
  38. uint8_t sign; /**< 1 bit sign */
  39. uint8_t exp; /**< 4 bits exponent */
  40. uint8_t mant; /**< 6 bits mantissa */
  41. } Float11;
  42. static inline Float11* i2f(int i, Float11* f)
  43. {
  44. f->sign = (i < 0);
  45. if (f->sign)
  46. i = -i;
  47. f->exp = av_log2_16bit(i) + !!i;
  48. f->mant = i? (i<<6) >> f->exp : 1<<5;
  49. return f;
  50. }
  51. static inline int16_t mult(Float11* f1, Float11* f2)
  52. {
  53. int res, exp;
  54. exp = f1->exp + f2->exp;
  55. res = (((f1->mant * f2->mant) + 0x30) >> 4);
  56. res = exp > 19 ? res << (exp - 19) : res >> (19 - exp);
  57. return (f1->sign ^ f2->sign) ? -res : res;
  58. }
  59. static inline int sgn(int value)
  60. {
  61. return (value < 0) ? -1 : 1;
  62. }
  63. typedef struct G726Tables {
  64. const int* quant; /**< quantization table */
  65. const int16_t* iquant; /**< inverse quantization table */
  66. const int16_t* W; /**< special table #1 ;-) */
  67. const uint8_t* F; /**< special table #2 */
  68. } G726Tables;
  69. typedef struct G726Context {
  70. AVClass *class;
  71. G726Tables tbls; /**< static tables needed for computation */
  72. Float11 sr[2]; /**< prev. reconstructed samples */
  73. Float11 dq[6]; /**< prev. difference */
  74. int a[2]; /**< second order predictor coeffs */
  75. int b[6]; /**< sixth order predictor coeffs */
  76. int pk[2]; /**< signs of prev. 2 sez + dq */
  77. int ap; /**< scale factor control */
  78. int yu; /**< fast scale factor */
  79. int yl; /**< slow scale factor */
  80. int dms; /**< short average magnitude of F[i] */
  81. int dml; /**< long average magnitude of F[i] */
  82. int td; /**< tone detect */
  83. int se; /**< estimated signal for the next iteration */
  84. int sez; /**< estimated second order prediction */
  85. int y; /**< quantizer scaling factor for the next iteration */
  86. int code_size;
  87. int little_endian; /**< little-endian bitstream as used in aiff and Sun AU */
  88. } G726Context;
  89. static const int quant_tbl16[] = /**< 16kbit/s 2 bits per sample */
  90. { 260, INT_MAX };
  91. static const int16_t iquant_tbl16[] =
  92. { 116, 365, 365, 116 };
  93. static const int16_t W_tbl16[] =
  94. { -22, 439, 439, -22 };
  95. static const uint8_t F_tbl16[] =
  96. { 0, 7, 7, 0 };
  97. static const int quant_tbl24[] = /**< 24kbit/s 3 bits per sample */
  98. { 7, 217, 330, INT_MAX };
  99. static const int16_t iquant_tbl24[] =
  100. { INT16_MIN, 135, 273, 373, 373, 273, 135, INT16_MIN };
  101. static const int16_t W_tbl24[] =
  102. { -4, 30, 137, 582, 582, 137, 30, -4 };
  103. static const uint8_t F_tbl24[] =
  104. { 0, 1, 2, 7, 7, 2, 1, 0 };
  105. static const int quant_tbl32[] = /**< 32kbit/s 4 bits per sample */
  106. { -125, 79, 177, 245, 299, 348, 399, INT_MAX };
  107. static const int16_t iquant_tbl32[] =
  108. { INT16_MIN, 4, 135, 213, 273, 323, 373, 425,
  109. 425, 373, 323, 273, 213, 135, 4, INT16_MIN };
  110. static const int16_t W_tbl32[] =
  111. { -12, 18, 41, 64, 112, 198, 355, 1122,
  112. 1122, 355, 198, 112, 64, 41, 18, -12};
  113. static const uint8_t F_tbl32[] =
  114. { 0, 0, 0, 1, 1, 1, 3, 7, 7, 3, 1, 1, 1, 0, 0, 0 };
  115. static const int quant_tbl40[] = /**< 40kbit/s 5 bits per sample */
  116. { -122, -16, 67, 138, 197, 249, 297, 338,
  117. 377, 412, 444, 474, 501, 527, 552, INT_MAX };
  118. static const int16_t iquant_tbl40[] =
  119. { INT16_MIN, -66, 28, 104, 169, 224, 274, 318,
  120. 358, 395, 429, 459, 488, 514, 539, 566,
  121. 566, 539, 514, 488, 459, 429, 395, 358,
  122. 318, 274, 224, 169, 104, 28, -66, INT16_MIN };
  123. static const int16_t W_tbl40[] =
  124. { 14, 14, 24, 39, 40, 41, 58, 100,
  125. 141, 179, 219, 280, 358, 440, 529, 696,
  126. 696, 529, 440, 358, 280, 219, 179, 141,
  127. 100, 58, 41, 40, 39, 24, 14, 14 };
  128. static const uint8_t F_tbl40[] =
  129. { 0, 0, 0, 0, 0, 1, 1, 1, 1, 1, 2, 3, 4, 5, 6, 6,
  130. 6, 6, 5, 4, 3, 2, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0 };
  131. static const G726Tables G726Tables_pool[] =
  132. {{ quant_tbl16, iquant_tbl16, W_tbl16, F_tbl16 },
  133. { quant_tbl24, iquant_tbl24, W_tbl24, F_tbl24 },
  134. { quant_tbl32, iquant_tbl32, W_tbl32, F_tbl32 },
  135. { quant_tbl40, iquant_tbl40, W_tbl40, F_tbl40 }};
  136. /**
  137. * Paragraph 4.2.2 page 18: Adaptive quantizer.
  138. */
  139. static inline uint8_t quant(G726Context* c, int d)
  140. {
  141. int sign, exp, i, dln;
  142. sign = i = 0;
  143. if (d < 0) {
  144. sign = 1;
  145. d = -d;
  146. }
  147. exp = av_log2_16bit(d);
  148. dln = ((exp<<7) + (((d<<7)>>exp)&0x7f)) - (c->y>>2);
  149. while (c->tbls.quant[i] < INT_MAX && c->tbls.quant[i] < dln)
  150. ++i;
  151. if (sign)
  152. i = ~i;
  153. if (c->code_size != 2 && i == 0) /* I'm not sure this is a good idea */
  154. i = 0xff;
  155. return i;
  156. }
  157. /**
  158. * Paragraph 4.2.3 page 22: Inverse adaptive quantizer.
  159. */
  160. static inline int16_t inverse_quant(G726Context* c, int i)
  161. {
  162. int dql, dex, dqt;
  163. dql = c->tbls.iquant[i] + (c->y >> 2);
  164. dex = (dql>>7) & 0xf; /* 4-bit exponent */
  165. dqt = (1<<7) + (dql & 0x7f); /* log2 -> linear */
  166. return (dql < 0) ? 0 : ((dqt<<dex) >> 7);
  167. }
  168. static int16_t g726_decode(G726Context* c, int I)
  169. {
  170. int dq, re_signal, pk0, fa1, i, tr, ylint, ylfrac, thr2, al, dq0;
  171. Float11 f;
  172. int I_sig= I >> (c->code_size - 1);
  173. dq = inverse_quant(c, I);
  174. /* Transition detect */
  175. ylint = (c->yl >> 15);
  176. ylfrac = (c->yl >> 10) & 0x1f;
  177. thr2 = (ylint > 9) ? 0x1f << 10 : (0x20 + ylfrac) << ylint;
  178. tr= (c->td == 1 && dq > ((3*thr2)>>2));
  179. if (I_sig) /* get the sign */
  180. dq = -dq;
  181. re_signal = (int16_t)(c->se + dq);
  182. /* Update second order predictor coefficient A2 and A1 */
  183. pk0 = (c->sez + dq) ? sgn(c->sez + dq) : 0;
  184. dq0 = dq ? sgn(dq) : 0;
  185. if (tr) {
  186. c->a[0] = 0;
  187. c->a[1] = 0;
  188. for (i=0; i<6; i++)
  189. c->b[i] = 0;
  190. } else {
  191. /* This is a bit crazy, but it really is +255 not +256 */
  192. fa1 = av_clip_intp2((-c->a[0]*c->pk[0]*pk0)>>5, 8);
  193. c->a[1] += 128*pk0*c->pk[1] + fa1 - (c->a[1]>>7);
  194. c->a[1] = av_clip(c->a[1], -12288, 12288);
  195. c->a[0] += 64*3*pk0*c->pk[0] - (c->a[0] >> 8);
  196. c->a[0] = av_clip(c->a[0], -(15360 - c->a[1]), 15360 - c->a[1]);
  197. for (i=0; i<6; i++)
  198. c->b[i] += 128*dq0*sgn(-c->dq[i].sign) - (c->b[i]>>8);
  199. }
  200. /* Update Dq and Sr and Pk */
  201. c->pk[1] = c->pk[0];
  202. c->pk[0] = pk0 ? pk0 : 1;
  203. c->sr[1] = c->sr[0];
  204. i2f(re_signal, &c->sr[0]);
  205. for (i=5; i>0; i--)
  206. c->dq[i] = c->dq[i-1];
  207. i2f(dq, &c->dq[0]);
  208. c->dq[0].sign = I_sig; /* Isn't it crazy ?!?! */
  209. c->td = c->a[1] < -11776;
  210. /* Update Ap */
  211. c->dms += (c->tbls.F[I]<<4) + ((- c->dms) >> 5);
  212. c->dml += (c->tbls.F[I]<<4) + ((- c->dml) >> 7);
  213. if (tr)
  214. c->ap = 256;
  215. else {
  216. c->ap += (-c->ap) >> 4;
  217. if (c->y <= 1535 || c->td || abs((c->dms << 2) - c->dml) >= (c->dml >> 3))
  218. c->ap += 0x20;
  219. }
  220. /* Update Yu and Yl */
  221. c->yu = av_clip(c->y + c->tbls.W[I] + ((-c->y)>>5), 544, 5120);
  222. c->yl += c->yu + ((-c->yl)>>6);
  223. /* Next iteration for Y */
  224. al = (c->ap >= 256) ? 1<<6 : c->ap >> 2;
  225. c->y = (c->yl + (c->yu - (c->yl>>6))*al) >> 6;
  226. /* Next iteration for SE and SEZ */
  227. c->se = 0;
  228. for (i=0; i<6; i++)
  229. c->se += mult(i2f(c->b[i] >> 2, &f), &c->dq[i]);
  230. c->sez = c->se >> 1;
  231. for (i=0; i<2; i++)
  232. c->se += mult(i2f(c->a[i] >> 2, &f), &c->sr[i]);
  233. c->se >>= 1;
  234. return av_clip(re_signal * 4, -0xffff, 0xffff);
  235. }
  236. static av_cold int g726_reset(G726Context *c)
  237. {
  238. int i;
  239. c->tbls = G726Tables_pool[c->code_size - 2];
  240. for (i=0; i<2; i++) {
  241. c->sr[i].mant = 1<<5;
  242. c->pk[i] = 1;
  243. }
  244. for (i=0; i<6; i++) {
  245. c->dq[i].mant = 1<<5;
  246. }
  247. c->yu = 544;
  248. c->yl = 34816;
  249. c->y = 544;
  250. return 0;
  251. }
  252. #if CONFIG_ADPCM_G726_ENCODER || CONFIG_ADPCM_G726LE_ENCODER
  253. static int16_t g726_encode(G726Context* c, int16_t sig)
  254. {
  255. uint8_t i;
  256. i = av_mod_uintp2(quant(c, sig/4 - c->se), c->code_size);
  257. g726_decode(c, i);
  258. return i;
  259. }
  260. /* Interfacing to the libavcodec */
  261. static av_cold int g726_encode_init(AVCodecContext *avctx)
  262. {
  263. G726Context* c = avctx->priv_data;
  264. c->little_endian = !strcmp(avctx->codec->name, "g726le");
  265. if (avctx->strict_std_compliance > FF_COMPLIANCE_UNOFFICIAL &&
  266. avctx->sample_rate != 8000) {
  267. av_log(avctx, AV_LOG_ERROR, "Sample rates other than 8kHz are not "
  268. "allowed when the compliance level is higher than unofficial. "
  269. "Resample or reduce the compliance level.\n");
  270. return AVERROR(EINVAL);
  271. }
  272. if (avctx->sample_rate <= 0) {
  273. av_log(avctx, AV_LOG_ERROR, "Invalid sample rate %d\n",
  274. avctx->sample_rate);
  275. return AVERROR(EINVAL);
  276. }
  277. if(avctx->channels != 1){
  278. av_log(avctx, AV_LOG_ERROR, "Only mono is supported\n");
  279. return AVERROR(EINVAL);
  280. }
  281. if (avctx->bit_rate)
  282. c->code_size = (avctx->bit_rate + avctx->sample_rate/2) / avctx->sample_rate;
  283. c->code_size = av_clip(c->code_size, 2, 5);
  284. avctx->bit_rate = c->code_size * avctx->sample_rate;
  285. avctx->bits_per_coded_sample = c->code_size;
  286. g726_reset(c);
  287. /* select a frame size that will end on a byte boundary and have a size of
  288. approximately 1024 bytes */
  289. avctx->frame_size = ((int[]){ 4096, 2736, 2048, 1640 })[c->code_size - 2];
  290. return 0;
  291. }
  292. static int g726_encode_frame(AVCodecContext *avctx, AVPacket *avpkt,
  293. const AVFrame *frame, int *got_packet_ptr)
  294. {
  295. G726Context *c = avctx->priv_data;
  296. const int16_t *samples = (const int16_t *)frame->data[0];
  297. PutBitContext pb;
  298. int i, ret, out_size;
  299. out_size = (frame->nb_samples * c->code_size + 7) / 8;
  300. if ((ret = ff_alloc_packet2(avctx, avpkt, out_size, 0)) < 0)
  301. return ret;
  302. init_put_bits(&pb, avpkt->data, avpkt->size);
  303. for (i = 0; i < frame->nb_samples; i++)
  304. if (c->little_endian) {
  305. put_bits_le(&pb, c->code_size, g726_encode(c, *samples++));
  306. } else {
  307. put_bits(&pb, c->code_size, g726_encode(c, *samples++));
  308. }
  309. if (c->little_endian) {
  310. flush_put_bits_le(&pb);
  311. } else {
  312. flush_put_bits(&pb);
  313. }
  314. avpkt->size = out_size;
  315. *got_packet_ptr = 1;
  316. return 0;
  317. }
  318. #define OFFSET(x) offsetof(G726Context, x)
  319. #define AE AV_OPT_FLAG_AUDIO_PARAM | AV_OPT_FLAG_ENCODING_PARAM
  320. static const AVOption options[] = {
  321. { "code_size", "Bits per code", OFFSET(code_size), AV_OPT_TYPE_INT, { .i64 = 4 }, 2, 5, AE },
  322. { NULL },
  323. };
  324. static const AVCodecDefault defaults[] = {
  325. { "b", "0" },
  326. { NULL },
  327. };
  328. #endif
  329. #if CONFIG_ADPCM_G726_ENCODER
  330. static const AVClass g726_class = {
  331. .class_name = "g726",
  332. .item_name = av_default_item_name,
  333. .option = options,
  334. .version = LIBAVUTIL_VERSION_INT,
  335. };
  336. AVCodec ff_adpcm_g726_encoder = {
  337. .name = "g726",
  338. .long_name = NULL_IF_CONFIG_SMALL("G.726 ADPCM"),
  339. .type = AVMEDIA_TYPE_AUDIO,
  340. .id = AV_CODEC_ID_ADPCM_G726,
  341. .priv_data_size = sizeof(G726Context),
  342. .init = g726_encode_init,
  343. .encode2 = g726_encode_frame,
  344. .capabilities = AV_CODEC_CAP_SMALL_LAST_FRAME,
  345. .sample_fmts = (const enum AVSampleFormat[]){ AV_SAMPLE_FMT_S16,
  346. AV_SAMPLE_FMT_NONE },
  347. .priv_class = &g726_class,
  348. .defaults = defaults,
  349. };
  350. #endif
  351. #if CONFIG_ADPCM_G726LE_ENCODER
  352. static const AVClass g726le_class = {
  353. .class_name = "g726le",
  354. .item_name = av_default_item_name,
  355. .option = options,
  356. .version = LIBAVUTIL_VERSION_INT,
  357. };
  358. AVCodec ff_adpcm_g726le_encoder = {
  359. .name = "g726le",
  360. .long_name = NULL_IF_CONFIG_SMALL("G.726 little endian ADPCM (\"right-justified\")"),
  361. .type = AVMEDIA_TYPE_AUDIO,
  362. .id = AV_CODEC_ID_ADPCM_G726LE,
  363. .priv_data_size = sizeof(G726Context),
  364. .init = g726_encode_init,
  365. .encode2 = g726_encode_frame,
  366. .capabilities = AV_CODEC_CAP_SMALL_LAST_FRAME,
  367. .sample_fmts = (const enum AVSampleFormat[]){ AV_SAMPLE_FMT_S16,
  368. AV_SAMPLE_FMT_NONE },
  369. .priv_class = &g726le_class,
  370. .defaults = defaults,
  371. };
  372. #endif
  373. #if CONFIG_ADPCM_G726_DECODER || CONFIG_ADPCM_G726LE_DECODER
  374. static av_cold int g726_decode_init(AVCodecContext *avctx)
  375. {
  376. G726Context* c = avctx->priv_data;
  377. if(avctx->channels > 1){
  378. avpriv_request_sample(avctx, "Decoding more than one channel");
  379. return AVERROR_PATCHWELCOME;
  380. }
  381. avctx->channels = 1;
  382. avctx->channel_layout = AV_CH_LAYOUT_MONO;
  383. c->little_endian = !strcmp(avctx->codec->name, "g726le");
  384. c->code_size = avctx->bits_per_coded_sample;
  385. if (c->code_size < 2 || c->code_size > 5) {
  386. av_log(avctx, AV_LOG_ERROR, "Invalid number of bits %d\n", c->code_size);
  387. return AVERROR(EINVAL);
  388. }
  389. g726_reset(c);
  390. avctx->sample_fmt = AV_SAMPLE_FMT_S16;
  391. return 0;
  392. }
  393. static int g726_decode_frame(AVCodecContext *avctx, void *data,
  394. int *got_frame_ptr, AVPacket *avpkt)
  395. {
  396. AVFrame *frame = data;
  397. const uint8_t *buf = avpkt->data;
  398. int buf_size = avpkt->size;
  399. G726Context *c = avctx->priv_data;
  400. int16_t *samples;
  401. GetBitContext gb;
  402. int out_samples, ret;
  403. out_samples = buf_size * 8 / c->code_size;
  404. /* get output buffer */
  405. frame->nb_samples = out_samples;
  406. if ((ret = ff_get_buffer(avctx, frame, 0)) < 0)
  407. return ret;
  408. samples = (int16_t *)frame->data[0];
  409. init_get_bits(&gb, buf, buf_size * 8);
  410. while (out_samples--)
  411. *samples++ = g726_decode(c, c->little_endian ?
  412. get_bits_le(&gb, c->code_size) :
  413. get_bits(&gb, c->code_size));
  414. if (get_bits_left(&gb) > 0)
  415. av_log(avctx, AV_LOG_ERROR, "Frame invalidly split, missing parser?\n");
  416. *got_frame_ptr = 1;
  417. return buf_size;
  418. }
  419. static void g726_decode_flush(AVCodecContext *avctx)
  420. {
  421. G726Context *c = avctx->priv_data;
  422. g726_reset(c);
  423. }
  424. #endif
  425. #if CONFIG_ADPCM_G726_DECODER
  426. AVCodec ff_adpcm_g726_decoder = {
  427. .name = "g726",
  428. .long_name = NULL_IF_CONFIG_SMALL("G.726 ADPCM"),
  429. .type = AVMEDIA_TYPE_AUDIO,
  430. .id = AV_CODEC_ID_ADPCM_G726,
  431. .priv_data_size = sizeof(G726Context),
  432. .init = g726_decode_init,
  433. .decode = g726_decode_frame,
  434. .flush = g726_decode_flush,
  435. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_CHANNEL_CONF,
  436. };
  437. #endif
  438. #if CONFIG_ADPCM_G726LE_DECODER
  439. AVCodec ff_adpcm_g726le_decoder = {
  440. .name = "g726le",
  441. .type = AVMEDIA_TYPE_AUDIO,
  442. .id = AV_CODEC_ID_ADPCM_G726LE,
  443. .priv_data_size = sizeof(G726Context),
  444. .init = g726_decode_init,
  445. .decode = g726_decode_frame,
  446. .flush = g726_decode_flush,
  447. .capabilities = AV_CODEC_CAP_DR1 | AV_CODEC_CAP_CHANNEL_CONF,
  448. .long_name = NULL_IF_CONFIG_SMALL("G.726 ADPCM little-endian"),
  449. };
  450. #endif