You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

495 lines
14KB

  1. /*
  2. * VFW capture interface
  3. * Copyright (c) 2006-2008 Ramiro Polla
  4. *
  5. * This file is part of FFmpeg.
  6. *
  7. * FFmpeg is free software; you can redistribute it and/or
  8. * modify it under the terms of the GNU Lesser General Public
  9. * License as published by the Free Software Foundation; either
  10. * version 2.1 of the License, or (at your option) any later version.
  11. *
  12. * FFmpeg is distributed in the hope that it will be useful,
  13. * but WITHOUT ANY WARRANTY; without even the implied warranty of
  14. * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
  15. * Lesser General Public License for more details.
  16. *
  17. * You should have received a copy of the GNU Lesser General Public
  18. * License along with FFmpeg; if not, write to the Free Software
  19. * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
  20. */
  21. #include "libavutil/log.h"
  22. #include "libavutil/opt.h"
  23. #include "libavutil/parseutils.h"
  24. #include <windows.h>
  25. #include <vfw.h>
  26. #include "avdevice.h"
  27. /* Defines for VFW missing from MinGW.
  28. * Remove this when MinGW incorporates them. */
  29. #define HWND_MESSAGE ((HWND)-3)
  30. /* End of missing MinGW defines */
  31. struct vfw_ctx {
  32. const AVClass *class;
  33. HWND hwnd;
  34. HANDLE mutex;
  35. HANDLE event;
  36. AVPacketList *pktl;
  37. unsigned int curbufsize;
  38. unsigned int frame_num;
  39. char *video_size; /**< A string describing video size, set by a private option. */
  40. char *framerate; /**< Set by a private option. */
  41. };
  42. static enum PixelFormat vfw_pixfmt(DWORD biCompression, WORD biBitCount)
  43. {
  44. switch(biCompression) {
  45. case MKTAG('U', 'Y', 'V', 'Y'):
  46. return PIX_FMT_UYVY422;
  47. case MKTAG('Y', 'U', 'Y', '2'):
  48. return PIX_FMT_YUYV422;
  49. case MKTAG('I', '4', '2', '0'):
  50. return PIX_FMT_YUV420P;
  51. case BI_RGB:
  52. switch(biBitCount) { /* 1-8 are untested */
  53. case 1:
  54. return PIX_FMT_MONOWHITE;
  55. case 4:
  56. return PIX_FMT_RGB4;
  57. case 8:
  58. return PIX_FMT_RGB8;
  59. case 16:
  60. return PIX_FMT_RGB555;
  61. case 24:
  62. return PIX_FMT_BGR24;
  63. case 32:
  64. return PIX_FMT_RGB32;
  65. }
  66. }
  67. return PIX_FMT_NONE;
  68. }
  69. static enum CodecID vfw_codecid(DWORD biCompression)
  70. {
  71. switch(biCompression) {
  72. case MKTAG('d', 'v', 's', 'd'):
  73. return CODEC_ID_DVVIDEO;
  74. case MKTAG('M', 'J', 'P', 'G'):
  75. case MKTAG('m', 'j', 'p', 'g'):
  76. return CODEC_ID_MJPEG;
  77. }
  78. return CODEC_ID_NONE;
  79. }
  80. #define dstruct(pctx, sname, var, type) \
  81. av_log(pctx, AV_LOG_DEBUG, #var":\t%"type"\n", sname->var)
  82. static void dump_captureparms(AVFormatContext *s, CAPTUREPARMS *cparms)
  83. {
  84. av_log(s, AV_LOG_DEBUG, "CAPTUREPARMS\n");
  85. dstruct(s, cparms, dwRequestMicroSecPerFrame, "lu");
  86. dstruct(s, cparms, fMakeUserHitOKToCapture, "d");
  87. dstruct(s, cparms, wPercentDropForError, "u");
  88. dstruct(s, cparms, fYield, "d");
  89. dstruct(s, cparms, dwIndexSize, "lu");
  90. dstruct(s, cparms, wChunkGranularity, "u");
  91. dstruct(s, cparms, fUsingDOSMemory, "d");
  92. dstruct(s, cparms, wNumVideoRequested, "u");
  93. dstruct(s, cparms, fCaptureAudio, "d");
  94. dstruct(s, cparms, wNumAudioRequested, "u");
  95. dstruct(s, cparms, vKeyAbort, "u");
  96. dstruct(s, cparms, fAbortLeftMouse, "d");
  97. dstruct(s, cparms, fAbortRightMouse, "d");
  98. dstruct(s, cparms, fLimitEnabled, "d");
  99. dstruct(s, cparms, wTimeLimit, "u");
  100. dstruct(s, cparms, fMCIControl, "d");
  101. dstruct(s, cparms, fStepMCIDevice, "d");
  102. dstruct(s, cparms, dwMCIStartTime, "lu");
  103. dstruct(s, cparms, dwMCIStopTime, "lu");
  104. dstruct(s, cparms, fStepCaptureAt2x, "d");
  105. dstruct(s, cparms, wStepCaptureAverageFrames, "u");
  106. dstruct(s, cparms, dwAudioBufferSize, "lu");
  107. dstruct(s, cparms, fDisableWriteCache, "d");
  108. dstruct(s, cparms, AVStreamMaster, "u");
  109. }
  110. static void dump_videohdr(AVFormatContext *s, VIDEOHDR *vhdr)
  111. {
  112. #ifdef DEBUG
  113. av_log(s, AV_LOG_DEBUG, "VIDEOHDR\n");
  114. dstruct(s, vhdr, lpData, "p");
  115. dstruct(s, vhdr, dwBufferLength, "lu");
  116. dstruct(s, vhdr, dwBytesUsed, "lu");
  117. dstruct(s, vhdr, dwTimeCaptured, "lu");
  118. dstruct(s, vhdr, dwUser, "lu");
  119. dstruct(s, vhdr, dwFlags, "lu");
  120. dstruct(s, vhdr, dwReserved[0], "lu");
  121. dstruct(s, vhdr, dwReserved[1], "lu");
  122. dstruct(s, vhdr, dwReserved[2], "lu");
  123. dstruct(s, vhdr, dwReserved[3], "lu");
  124. #endif
  125. }
  126. static void dump_bih(AVFormatContext *s, BITMAPINFOHEADER *bih)
  127. {
  128. av_log(s, AV_LOG_DEBUG, "BITMAPINFOHEADER\n");
  129. dstruct(s, bih, biSize, "lu");
  130. dstruct(s, bih, biWidth, "ld");
  131. dstruct(s, bih, biHeight, "ld");
  132. dstruct(s, bih, biPlanes, "d");
  133. dstruct(s, bih, biBitCount, "d");
  134. dstruct(s, bih, biCompression, "lu");
  135. av_log(s, AV_LOG_DEBUG, " biCompression:\t\"%.4s\"\n",
  136. (char*) &bih->biCompression);
  137. dstruct(s, bih, biSizeImage, "lu");
  138. dstruct(s, bih, biXPelsPerMeter, "lu");
  139. dstruct(s, bih, biYPelsPerMeter, "lu");
  140. dstruct(s, bih, biClrUsed, "lu");
  141. dstruct(s, bih, biClrImportant, "lu");
  142. }
  143. static int shall_we_drop(AVFormatContext *s)
  144. {
  145. struct vfw_ctx *ctx = s->priv_data;
  146. const uint8_t dropscore[] = {62, 75, 87, 100};
  147. const int ndropscores = FF_ARRAY_ELEMS(dropscore);
  148. unsigned int buffer_fullness = (ctx->curbufsize*100)/s->max_picture_buffer;
  149. if(dropscore[++ctx->frame_num%ndropscores] <= buffer_fullness) {
  150. av_log(s, AV_LOG_ERROR,
  151. "real-time buffer %d%% full! frame dropped!\n", buffer_fullness);
  152. return 1;
  153. }
  154. return 0;
  155. }
  156. static LRESULT CALLBACK videostream_cb(HWND hwnd, LPVIDEOHDR vdhdr)
  157. {
  158. AVFormatContext *s;
  159. struct vfw_ctx *ctx;
  160. AVPacketList **ppktl, *pktl_next;
  161. s = (AVFormatContext *) GetWindowLongPtr(hwnd, GWLP_USERDATA);
  162. ctx = s->priv_data;
  163. dump_videohdr(s, vdhdr);
  164. if(shall_we_drop(s))
  165. return FALSE;
  166. WaitForSingleObject(ctx->mutex, INFINITE);
  167. pktl_next = av_mallocz(sizeof(AVPacketList));
  168. if(!pktl_next)
  169. goto fail;
  170. if(av_new_packet(&pktl_next->pkt, vdhdr->dwBytesUsed) < 0) {
  171. av_free(pktl_next);
  172. goto fail;
  173. }
  174. pktl_next->pkt.pts = vdhdr->dwTimeCaptured;
  175. memcpy(pktl_next->pkt.data, vdhdr->lpData, vdhdr->dwBytesUsed);
  176. for(ppktl = &ctx->pktl ; *ppktl ; ppktl = &(*ppktl)->next);
  177. *ppktl = pktl_next;
  178. ctx->curbufsize += vdhdr->dwBytesUsed;
  179. SetEvent(ctx->event);
  180. ReleaseMutex(ctx->mutex);
  181. return TRUE;
  182. fail:
  183. ReleaseMutex(ctx->mutex);
  184. return FALSE;
  185. }
  186. static int vfw_read_close(AVFormatContext *s)
  187. {
  188. struct vfw_ctx *ctx = s->priv_data;
  189. AVPacketList *pktl;
  190. if(ctx->hwnd) {
  191. SendMessage(ctx->hwnd, WM_CAP_SET_CALLBACK_VIDEOSTREAM, 0, 0);
  192. SendMessage(ctx->hwnd, WM_CAP_DRIVER_DISCONNECT, 0, 0);
  193. DestroyWindow(ctx->hwnd);
  194. }
  195. if(ctx->mutex)
  196. CloseHandle(ctx->mutex);
  197. if(ctx->event)
  198. CloseHandle(ctx->event);
  199. pktl = ctx->pktl;
  200. while (pktl) {
  201. AVPacketList *next = pktl->next;
  202. av_destruct_packet(&pktl->pkt);
  203. av_free(pktl);
  204. pktl = next;
  205. }
  206. av_freep(&ctx->video_size);
  207. av_freep(&ctx->framerate);
  208. return 0;
  209. }
  210. static int vfw_read_header(AVFormatContext *s, AVFormatParameters *ap)
  211. {
  212. struct vfw_ctx *ctx = s->priv_data;
  213. AVCodecContext *codec;
  214. AVStream *st;
  215. int devnum;
  216. int bisize;
  217. BITMAPINFO *bi;
  218. CAPTUREPARMS cparms;
  219. DWORD biCompression;
  220. WORD biBitCount;
  221. int ret;
  222. AVRational fps;
  223. if (!strcmp(s->filename, "list")) {
  224. for (devnum = 0; devnum <= 9; devnum++) {
  225. char driver_name[256];
  226. char driver_ver[256];
  227. ret = capGetDriverDescription(devnum,
  228. driver_name, sizeof(driver_name),
  229. driver_ver, sizeof(driver_ver));
  230. if (ret) {
  231. av_log(s, AV_LOG_INFO, "Driver %d\n", devnum);
  232. av_log(s, AV_LOG_INFO, " %s\n", driver_name);
  233. av_log(s, AV_LOG_INFO, " %s\n", driver_ver);
  234. }
  235. }
  236. return AVERROR(EIO);
  237. }
  238. #if FF_API_FORMAT_PARAMETERS
  239. if (ap->time_base.num)
  240. fps = (AVRational){ap->time_base.den, ap->time_base.num};
  241. #endif
  242. ctx->hwnd = capCreateCaptureWindow(NULL, 0, 0, 0, 0, 0, HWND_MESSAGE, 0);
  243. if(!ctx->hwnd) {
  244. av_log(s, AV_LOG_ERROR, "Could not create capture window.\n");
  245. return AVERROR(EIO);
  246. }
  247. /* If atoi fails, devnum==0 and the default device is used */
  248. devnum = atoi(s->filename);
  249. ret = SendMessage(ctx->hwnd, WM_CAP_DRIVER_CONNECT, devnum, 0);
  250. if(!ret) {
  251. av_log(s, AV_LOG_ERROR, "Could not connect to device.\n");
  252. DestroyWindow(ctx->hwnd);
  253. return AVERROR(ENODEV);
  254. }
  255. SendMessage(ctx->hwnd, WM_CAP_SET_OVERLAY, 0, 0);
  256. SendMessage(ctx->hwnd, WM_CAP_SET_PREVIEW, 0, 0);
  257. ret = SendMessage(ctx->hwnd, WM_CAP_SET_CALLBACK_VIDEOSTREAM, 0,
  258. (LPARAM) videostream_cb);
  259. if(!ret) {
  260. av_log(s, AV_LOG_ERROR, "Could not set video stream callback.\n");
  261. goto fail_io;
  262. }
  263. SetWindowLongPtr(ctx->hwnd, GWLP_USERDATA, (LONG_PTR) s);
  264. st = av_new_stream(s, 0);
  265. if(!st) {
  266. vfw_read_close(s);
  267. return AVERROR(ENOMEM);
  268. }
  269. /* Set video format */
  270. bisize = SendMessage(ctx->hwnd, WM_CAP_GET_VIDEOFORMAT, 0, 0);
  271. if(!bisize)
  272. goto fail_io;
  273. bi = av_malloc(bisize);
  274. if(!bi) {
  275. vfw_read_close(s);
  276. return AVERROR(ENOMEM);
  277. }
  278. ret = SendMessage(ctx->hwnd, WM_CAP_GET_VIDEOFORMAT, bisize, (LPARAM) bi);
  279. if(!ret)
  280. goto fail_bi;
  281. dump_bih(s, &bi->bmiHeader);
  282. if (ctx->video_size) {
  283. ret = av_parse_video_size(&bi->bmiHeader.biWidth, &bi->bmiHeader.biHeight, ctx->video_size);
  284. if (ret < 0) {
  285. av_log(s, AV_LOG_ERROR, "Couldn't parse video size.\n");
  286. goto fail_bi;
  287. }
  288. }
  289. #if FF_API_FORMAT_PARAMETERS
  290. if (ap->width > 0)
  291. bi->bmiHeader.biWidth = ap->width;
  292. if (ap->height > 0)
  293. bi->bmiHeader.biHeight = ap->height;
  294. #endif
  295. if (0) {
  296. /* For testing yet unsupported compressions
  297. * Copy these values from user-supplied verbose information */
  298. bi->bmiHeader.biWidth = 320;
  299. bi->bmiHeader.biHeight = 240;
  300. bi->bmiHeader.biPlanes = 1;
  301. bi->bmiHeader.biBitCount = 12;
  302. bi->bmiHeader.biCompression = MKTAG('I','4','2','0');
  303. bi->bmiHeader.biSizeImage = 115200;
  304. dump_bih(s, &bi->bmiHeader);
  305. }
  306. ret = SendMessage(ctx->hwnd, WM_CAP_SET_VIDEOFORMAT, bisize, (LPARAM) bi);
  307. if(!ret) {
  308. av_log(s, AV_LOG_ERROR, "Could not set Video Format.\n");
  309. goto fail_bi;
  310. }
  311. biCompression = bi->bmiHeader.biCompression;
  312. biBitCount = bi->bmiHeader.biBitCount;
  313. av_free(bi);
  314. /* Set sequence setup */
  315. ret = SendMessage(ctx->hwnd, WM_CAP_GET_SEQUENCE_SETUP, sizeof(cparms),
  316. (LPARAM) &cparms);
  317. if(!ret)
  318. goto fail_io;
  319. dump_captureparms(s, &cparms);
  320. cparms.fYield = 1; // Spawn a background thread
  321. cparms.dwRequestMicroSecPerFrame =
  322. (fps.den*1000000) / fps.num;
  323. cparms.fAbortLeftMouse = 0;
  324. cparms.fAbortRightMouse = 0;
  325. cparms.fCaptureAudio = 0;
  326. cparms.vKeyAbort = 0;
  327. ret = SendMessage(ctx->hwnd, WM_CAP_SET_SEQUENCE_SETUP, sizeof(cparms),
  328. (LPARAM) &cparms);
  329. if(!ret)
  330. goto fail_io;
  331. codec = st->codec;
  332. codec->time_base = (AVRational){fps.den, fps.num};
  333. codec->codec_type = AVMEDIA_TYPE_VIDEO;
  334. codec->width = bi->bmiHeader.biWidth;
  335. codec->height = bi->bmiHeader.biHeight;
  336. codec->pix_fmt = vfw_pixfmt(biCompression, biBitCount);
  337. if(codec->pix_fmt == PIX_FMT_NONE) {
  338. codec->codec_id = vfw_codecid(biCompression);
  339. if(codec->codec_id == CODEC_ID_NONE) {
  340. av_log(s, AV_LOG_ERROR, "Unknown compression type. "
  341. "Please report verbose (-v 9) debug information.\n");
  342. vfw_read_close(s);
  343. return AVERROR_PATCHWELCOME;
  344. }
  345. codec->bits_per_coded_sample = biBitCount;
  346. } else {
  347. codec->codec_id = CODEC_ID_RAWVIDEO;
  348. if(biCompression == BI_RGB) {
  349. codec->bits_per_coded_sample = biBitCount;
  350. codec->extradata = av_malloc(9 + FF_INPUT_BUFFER_PADDING_SIZE);
  351. if (codec->extradata) {
  352. codec->extradata_size = 9;
  353. memcpy(codec->extradata, "BottomUp", 9);
  354. }
  355. }
  356. }
  357. av_set_pts_info(st, 32, 1, 1000);
  358. ctx->mutex = CreateMutex(NULL, 0, NULL);
  359. if(!ctx->mutex) {
  360. av_log(s, AV_LOG_ERROR, "Could not create Mutex.\n" );
  361. goto fail_io;
  362. }
  363. ctx->event = CreateEvent(NULL, 1, 0, NULL);
  364. if(!ctx->event) {
  365. av_log(s, AV_LOG_ERROR, "Could not create Event.\n" );
  366. goto fail_io;
  367. }
  368. ret = SendMessage(ctx->hwnd, WM_CAP_SEQUENCE_NOFILE, 0, 0);
  369. if(!ret) {
  370. av_log(s, AV_LOG_ERROR, "Could not start capture sequence.\n" );
  371. goto fail_io;
  372. }
  373. return 0;
  374. fail_bi:
  375. av_free(bi);
  376. fail_io:
  377. vfw_read_close(s);
  378. return AVERROR(EIO);
  379. }
  380. static int vfw_read_packet(AVFormatContext *s, AVPacket *pkt)
  381. {
  382. struct vfw_ctx *ctx = s->priv_data;
  383. AVPacketList *pktl = NULL;
  384. while(!pktl) {
  385. WaitForSingleObject(ctx->mutex, INFINITE);
  386. pktl = ctx->pktl;
  387. if(ctx->pktl) {
  388. *pkt = ctx->pktl->pkt;
  389. ctx->pktl = ctx->pktl->next;
  390. av_free(pktl);
  391. }
  392. ResetEvent(ctx->event);
  393. ReleaseMutex(ctx->mutex);
  394. if(!pktl) {
  395. if(s->flags & AVFMT_FLAG_NONBLOCK) {
  396. return AVERROR(EAGAIN);
  397. } else {
  398. WaitForSingleObject(ctx->event, INFINITE);
  399. }
  400. }
  401. }
  402. ctx->curbufsize -= pkt->size;
  403. return pkt->size;
  404. }
  405. #define OFFSET(x) offsetof(struct vfw_ctx, x)
  406. #define DEC AV_OPT_FLAG_DECODING_PARAM
  407. static const AVOption options[] = {
  408. { "video_size", "A string describing frame size, such as 640x480 or hd720.", OFFSET(video_size), FF_OPT_TYPE_STRING, {.str = NULL}, 0, 0, DEC },
  409. { "framerate", "", OFFSET(framerate), FF_OPT_TYPE_STRING, {.str = "ntsc"}, 0, 0, DEC },
  410. { NULL },
  411. };
  412. static const AVClass vfw_class = {
  413. .class_name = "VFW indev",
  414. .item_name = av_default_item_name,
  415. .option = options,
  416. .version = LIBAVUTIL_VERSION_INT,
  417. };
  418. AVInputFormat ff_vfwcap_demuxer = {
  419. "vfwcap",
  420. NULL_IF_CONFIG_SMALL("VFW video capture"),
  421. sizeof(struct vfw_ctx),
  422. NULL,
  423. vfw_read_header,
  424. vfw_read_packet,
  425. vfw_read_close,
  426. .flags = AVFMT_NOFILE,
  427. .priv_class = &vfw_class,
  428. };