/* ============================================================================== This file is part of the JUCE library. Copyright (c) 2013 - Raw Material Software Ltd. Permission is granted to use this software under the terms of either: a) the GPL v2 (or any later version) b) the Affero GPL v3 Details of these licenses can be found at: www.gnu.org/licenses JUCE is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. ------------------------------------------------------------------------------ To release a closed-source product which uses JUCE, commercial licenses are available: visit www.juce.com for more information. ============================================================================== */ #if ! JUCE_QUICKTIME #error "On the Mac, cameras use Quicktime, so if you turn on JUCE_USE_CAMERA, you also need to enable JUCE_QUICKTIME" #endif extern Image juce_createImageFromCIImage (CIImage* im, int w, int h); //============================================================================== class QTCameraDeviceInternal { public: QTCameraDeviceInternal (CameraDevice*, const int index) : input (nil), audioDevice (nil), audioInput (nil), session (nil), fileOutput (nil), imageOutput (nil), firstPresentationTime (0), averageTimeOffset (0) { JUCE_AUTORELEASEPOOL { session = [[QTCaptureSession alloc] init]; NSArray* devs = [QTCaptureDevice inputDevicesWithMediaType: QTMediaTypeVideo]; device = (QTCaptureDevice*) [devs objectAtIndex: index]; static DelegateClass cls; callbackDelegate = [cls.createInstance() init]; DelegateClass::setOwner (callbackDelegate, this); NSError* err = nil; [device retain]; [device open: &err]; if (err == nil) { input = [[QTCaptureDeviceInput alloc] initWithDevice: device]; audioInput = [[QTCaptureDeviceInput alloc] initWithDevice: device]; [session addInput: input error: &err]; if (err == nil) { resetFile(); imageOutput = [[QTCaptureDecompressedVideoOutput alloc] init]; [imageOutput setDelegate: callbackDelegate]; if (err == nil) { [session startRunning]; return; } } } openingError = nsStringToJuce ([err description]); DBG (openingError); } } ~QTCameraDeviceInternal() { [session stopRunning]; [session removeOutput: imageOutput]; [session release]; [input release]; [device release]; [audioDevice release]; [audioInput release]; [fileOutput release]; [imageOutput release]; [callbackDelegate release]; } void resetFile() { [fileOutput recordToOutputFileURL: nil]; [session removeOutput: fileOutput]; [fileOutput release]; fileOutput = [[QTCaptureMovieFileOutput alloc] init]; [session removeInput: audioInput]; [audioInput release]; audioInput = nil; [audioDevice release]; audioDevice = nil; [fileOutput setDelegate: callbackDelegate]; } void addDefaultAudioInput() { NSError* err = nil; audioDevice = [QTCaptureDevice defaultInputDeviceWithMediaType: QTMediaTypeSound]; if ([audioDevice open: &err]) [audioDevice retain]; else audioDevice = nil; if (audioDevice != nil) { audioInput = [[QTCaptureDeviceInput alloc] initWithDevice: audioDevice]; [session addInput: audioInput error: &err]; } } void addListener (CameraDevice::Listener* listenerToAdd) { const ScopedLock sl (listenerLock); if (listeners.size() == 0) [session addOutput: imageOutput error: nil]; listeners.addIfNotAlreadyThere (listenerToAdd); } void removeListener (CameraDevice::Listener* listenerToRemove) { const ScopedLock sl (listenerLock); listeners.removeFirstMatchingValue (listenerToRemove); if (listeners.size() == 0) [session removeOutput: imageOutput]; } void callListeners (CIImage* frame, int w, int h) { Image image (juce_createImageFromCIImage (frame, w, h)); const ScopedLock sl (listenerLock); for (int i = listeners.size(); --i >= 0;) { CameraDevice::Listener* const l = listeners[i]; if (l != nullptr) l->imageReceived (image); } } void captureBuffer (QTSampleBuffer* sampleBuffer) { const Time now (Time::getCurrentTime()); #if MAC_OS_X_VERSION_MAX_ALLOWED >= MAC_OS_X_VERSION_10_5 NSNumber* hosttime = (NSNumber*) [sampleBuffer attributeForKey: QTSampleBufferHostTimeAttribute]; #else NSNumber* hosttime = (NSNumber*) [sampleBuffer attributeForKey: nsStringLiteral ("hostTime")]; #endif int64 presentationTime = (hosttime != nil) ? ((int64) AudioConvertHostTimeToNanos ([hosttime unsignedLongLongValue]) / 1000000 + 40) : (([sampleBuffer presentationTime].timeValue * 1000) / [sampleBuffer presentationTime].timeScale + 50); const int64 timeDiff = now.toMilliseconds() - presentationTime; if (firstPresentationTime == 0) { firstPresentationTime = presentationTime; averageTimeOffset = timeDiff; } else { averageTimeOffset = (averageTimeOffset * 120 + timeDiff * 8) / 128; } } QTCaptureDevice* device; QTCaptureDeviceInput* input; QTCaptureDevice* audioDevice; QTCaptureDeviceInput* audioInput; QTCaptureSession* session; QTCaptureMovieFileOutput* fileOutput; QTCaptureDecompressedVideoOutput* imageOutput; NSObject* callbackDelegate; String openingError; int64 firstPresentationTime; int64 averageTimeOffset; Array listeners; CriticalSection listenerLock; private: //============================================================================== struct DelegateClass : public ObjCClass { DelegateClass() : ObjCClass ("JUCEAppDelegate_") { addIvar ("owner"); addMethod (@selector (captureOutput:didOutputVideoFrame:withSampleBuffer:fromConnection:), didOutputVideoFrame, "v@:@", @encode (CVImageBufferRef), "@@"); addMethod (@selector (captureOutput:didOutputSampleBuffer:fromConnection:), didOutputVideoFrame, "v@:@@@"); registerClass(); } static void setOwner (id self, QTCameraDeviceInternal* owner) { object_setInstanceVariable (self, "owner", owner); } static QTCameraDeviceInternal* getOwner (id self) { return getIvar (self, "owner"); } private: static void didOutputVideoFrame (id self, SEL, QTCaptureOutput*, CVImageBufferRef videoFrame, QTSampleBuffer*, QTCaptureConnection*) { QTCameraDeviceInternal* const internal = getOwner (self); if (internal->listeners.size() > 0) { JUCE_AUTORELEASEPOOL { internal->callListeners ([CIImage imageWithCVImageBuffer: videoFrame], (int) CVPixelBufferGetWidth (videoFrame), (int) CVPixelBufferGetHeight (videoFrame)); } } } static void didOutputSampleBuffer (id self, SEL, QTCaptureFileOutput*, QTSampleBuffer* sampleBuffer, QTCaptureConnection*) { getOwner (self)->captureBuffer (sampleBuffer); } }; }; //============================================================================== class QTCaptureViewerComp : public NSViewComponent { public: QTCaptureViewerComp (CameraDevice*, QTCameraDeviceInternal* internal) { JUCE_AUTORELEASEPOOL { captureView = [[QTCaptureView alloc] init]; [captureView setCaptureSession: internal->session]; setSize (640, 480); // xxx need to somehow get the movie size - how? setView (captureView); } } ~QTCaptureViewerComp() { setView (0); [captureView setCaptureSession: nil]; [captureView release]; } QTCaptureView* captureView; }; //============================================================================== CameraDevice::CameraDevice (const String& name_, int index) : name (name_) { isRecording = false; internal = new QTCameraDeviceInternal (this, index); } CameraDevice::~CameraDevice() { stopRecording(); delete static_cast (internal); internal = nullptr; } Component* CameraDevice::createViewerComponent() { return new QTCaptureViewerComp (this, static_cast (internal)); } String CameraDevice::getFileExtension() { return ".mov"; } void CameraDevice::startRecordingToFile (const File& file, int quality) { stopRecording(); QTCameraDeviceInternal* const d = static_cast (internal); d->firstPresentationTime = 0; file.deleteFile(); // In some versions of QT (e.g. on 10.5), if you record video without audio, the speed comes // out wrong, so we'll put some audio in there too.., d->addDefaultAudioInput(); [d->session addOutput: d->fileOutput error: nil]; NSEnumerator* connectionEnumerator = [[d->fileOutput connections] objectEnumerator]; for (;;) { QTCaptureConnection* connection = [connectionEnumerator nextObject]; if (connection == nil) break; QTCompressionOptions* options = nil; NSString* mediaType = [connection mediaType]; if ([mediaType isEqualToString: QTMediaTypeVideo]) options = [QTCompressionOptions compressionOptionsWithIdentifier: quality >= 1 ? nsStringLiteral ("QTCompressionOptionsSD480SizeH264Video") : nsStringLiteral ("QTCompressionOptions240SizeH264Video")]; else if ([mediaType isEqualToString: QTMediaTypeSound]) options = [QTCompressionOptions compressionOptionsWithIdentifier: nsStringLiteral ("QTCompressionOptionsHighQualityAACAudio")]; [d->fileOutput setCompressionOptions: options forConnection: connection]; } [d->fileOutput recordToOutputFileURL: [NSURL fileURLWithPath: juceStringToNS (file.getFullPathName())]]; isRecording = true; } Time CameraDevice::getTimeOfFirstRecordedFrame() const { QTCameraDeviceInternal* const d = static_cast (internal); if (d->firstPresentationTime != 0) return Time (d->firstPresentationTime + d->averageTimeOffset); return Time(); } void CameraDevice::stopRecording() { if (isRecording) { static_cast (internal)->resetFile(); isRecording = false; } } void CameraDevice::addListener (Listener* listenerToAdd) { if (listenerToAdd != nullptr) static_cast (internal)->addListener (listenerToAdd); } void CameraDevice::removeListener (Listener* listenerToRemove) { if (listenerToRemove != nullptr) static_cast (internal)->removeListener (listenerToRemove); } //============================================================================== StringArray CameraDevice::getAvailableDevices() { JUCE_AUTORELEASEPOOL { StringArray results; NSArray* devs = [QTCaptureDevice inputDevicesWithMediaType: QTMediaTypeVideo]; for (int i = 0; i < (int) [devs count]; ++i) { QTCaptureDevice* dev = (QTCaptureDevice*) [devs objectAtIndex: i]; results.add (nsStringToJuce ([dev localizedDisplayName])); } return results; } } CameraDevice* CameraDevice::openDevice (int index, int /*minWidth*/, int /*minHeight*/, int /*maxWidth*/, int /*maxHeight*/) { ScopedPointer d (new CameraDevice (getAvailableDevices() [index], index)); if (static_cast (d->internal)->openingError.isEmpty()) return d.release(); return nullptr; }