/* ============================================================================== This file is part of the JUCE library. Copyright (c) 2020 - Raw Material Software Limited JUCE is an open source library subject to commercial or open-source licensing. By using JUCE, you agree to the terms of both the JUCE 6 End-User License Agreement and JUCE Privacy Policy (both effective as of the 16th June 2020). End User License Agreement: www.juce.com/juce-6-licence Privacy Policy: www.juce.com/juce-privacy-policy Or: You may also use this code under the terms of the GPL v3 (see www.gnu.org/licenses). JUCE IS PROVIDED "AS IS" WITHOUT ANY WARRANTY, AND ALL WARRANTIES, WHETHER EXPRESSED OR IMPLIED, INCLUDING MERCHANTABILITY AND FITNESS FOR PURPOSE, ARE DISCLAIMED. ============================================================================== */ #if (defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MIN_REQUIRED >= __IPHONE_10_0) JUCE_BEGIN_IGNORE_WARNINGS_GCC_LIKE ("-Wdeprecated-declarations") #define JUCE_DEPRECATION_IGNORED 1 #endif struct CameraDevice::Pimpl { using InternalOpenCameraResultCallback = std::function; Pimpl (CameraDevice& ownerToUse, const String& cameraIdToUse, int /*index*/, int /*minWidth*/, int /*minHeight*/, int /*maxWidth*/, int /*maxHeight*/, bool useHighQuality) : owner (ownerToUse), cameraId (cameraIdToUse), captureSession (*this, useHighQuality) { } String getCameraId() const noexcept { return cameraId; } void open (InternalOpenCameraResultCallback cameraOpenCallbackToUse) { cameraOpenCallback = std::move (cameraOpenCallbackToUse); if (cameraOpenCallback == nullptr) { // A valid camera open callback must be passed. jassertfalse; return; } [AVCaptureDevice requestAccessForMediaType: AVMediaTypeVideo completionHandler: ^(BOOL granted) { // Access to video is required for camera to work, // black images will be produced otherwise! jassertquiet (granted); }]; [AVCaptureDevice requestAccessForMediaType: AVMediaTypeAudio completionHandler: ^(BOOL granted) { // Access to audio is required for camera to work, // silence will be produced otherwise! jassertquiet (granted); }]; captureSession.startSessionForDeviceWithId (cameraId); } bool openedOk() const noexcept { return captureSession.openedOk(); } void takeStillPicture (std::function pictureTakenCallbackToUse) { if (pictureTakenCallbackToUse == nullptr) { jassertfalse; return; } pictureTakenCallback = std::move (pictureTakenCallbackToUse); triggerStillPictureCapture(); } void startRecordingToFile (const File& file, int /*quality*/) { file.deleteFile(); captureSession.startRecording (file); } void stopRecording() { captureSession.stopRecording(); } Time getTimeOfFirstRecordedFrame() const { return captureSession.getTimeOfFirstRecordedFrame(); } static StringArray getAvailableDevices() { StringArray results; JUCE_CAMERA_LOG ("Available camera devices: "); for (AVCaptureDevice* device in getDevices()) { JUCE_CAMERA_LOG ("Device start----------------------------------"); printDebugCameraInfo (device); JUCE_CAMERA_LOG ("Device end----------------------------------"); results.add (nsStringToJuce (device.uniqueID)); } return results; } void addListener (CameraDevice::Listener* listenerToAdd) { const ScopedLock sl (listenerLock); listeners.add (listenerToAdd); if (listeners.size() == 1) triggerStillPictureCapture(); } void removeListener (CameraDevice::Listener* listenerToRemove) { const ScopedLock sl (listenerLock); listeners.remove (listenerToRemove); } private: static NSArray* getDevices() { #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) { std::unique_ptr, NSObjectDeleter> deviceTypes ([[NSMutableArray alloc] initWithCapacity: 2]); [deviceTypes.get() addObject: AVCaptureDeviceTypeBuiltInWideAngleCamera]; [deviceTypes.get() addObject: AVCaptureDeviceTypeBuiltInTelephotoCamera]; if (@available (iOS 10.2, *)) [deviceTypes.get() addObject: AVCaptureDeviceTypeBuiltInDualCamera]; if (@available (iOS 11.1, *)) [deviceTypes.get() addObject: AVCaptureDeviceTypeBuiltInTrueDepthCamera]; auto discoverySession = [AVCaptureDeviceDiscoverySession discoverySessionWithDeviceTypes: deviceTypes.get() mediaType: AVMediaTypeVideo position: AVCaptureDevicePositionUnspecified]; return [discoverySession devices]; } #endif return [AVCaptureDevice devicesWithMediaType: AVMediaTypeVideo]; } //============================================================================== static void printDebugCameraInfo (AVCaptureDevice* device) { auto position = device.position; String positionString = position == AVCaptureDevicePositionBack ? "Back" : position == AVCaptureDevicePositionFront ? "Front" : "Unspecified"; JUCE_CAMERA_LOG ("Position: " + positionString); JUCE_CAMERA_LOG ("Model ID: " + nsStringToJuce (device.modelID)); JUCE_CAMERA_LOG ("Localized name: " + nsStringToJuce (device.localizedName)); JUCE_CAMERA_LOG ("Unique ID: " + nsStringToJuce (device.uniqueID)); JUCE_CAMERA_LOG ("Lens aperture: " + String (device.lensAperture)); JUCE_CAMERA_LOG ("Has flash: " + String ((int)device.hasFlash)); JUCE_CAMERA_LOG ("Supports flash always on: " + String ((int)[device isFlashModeSupported: AVCaptureFlashModeOn])); JUCE_CAMERA_LOG ("Supports auto flash: " + String ((int)[device isFlashModeSupported: AVCaptureFlashModeAuto])); JUCE_CAMERA_LOG ("Has torch: " + String ((int)device.hasTorch)); JUCE_CAMERA_LOG ("Supports torch always on: " + String ((int)[device isTorchModeSupported: AVCaptureTorchModeOn])); JUCE_CAMERA_LOG ("Supports auto torch: " + String ((int)[device isTorchModeSupported: AVCaptureTorchModeAuto])); JUCE_CAMERA_LOG ("Low light boost supported: " + String ((int)device.lowLightBoostEnabled)); JUCE_CAMERA_LOG ("Supports auto white balance: " + String ((int)[device isWhiteBalanceModeSupported: AVCaptureWhiteBalanceModeAutoWhiteBalance])); JUCE_CAMERA_LOG ("Supports continuous auto white balance: " + String ((int)[device isWhiteBalanceModeSupported: AVCaptureWhiteBalanceModeContinuousAutoWhiteBalance])); JUCE_CAMERA_LOG ("Supports auto focus: " + String ((int)[device isFocusModeSupported: AVCaptureFocusModeAutoFocus])); JUCE_CAMERA_LOG ("Supports continuous auto focus: " + String ((int)[device isFocusModeSupported: AVCaptureFocusModeContinuousAutoFocus])); JUCE_CAMERA_LOG ("Supports point of interest focus: " + String ((int)device.focusPointOfInterestSupported)); JUCE_CAMERA_LOG ("Smooth auto focus supported: " + String ((int)device.smoothAutoFocusSupported)); JUCE_CAMERA_LOG ("Auto focus range restriction supported: " + String ((int)device.autoFocusRangeRestrictionSupported)); JUCE_CAMERA_LOG ("Supports auto exposure: " + String ((int)[device isExposureModeSupported: AVCaptureExposureModeAutoExpose])); JUCE_CAMERA_LOG ("Supports continuous auto exposure: " + String ((int)[device isExposureModeSupported: AVCaptureExposureModeContinuousAutoExposure])); JUCE_CAMERA_LOG ("Supports custom exposure: " + String ((int)[device isExposureModeSupported: AVCaptureExposureModeCustom])); JUCE_CAMERA_LOG ("Supports point of interest exposure: " + String ((int)device.exposurePointOfInterestSupported)); #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) { JUCE_CAMERA_LOG ("Device type: " + nsStringToJuce (device.deviceType)); JUCE_CAMERA_LOG ("Locking focus with custom lens position supported: " + String ((int)device.lockingFocusWithCustomLensPositionSupported)); } #endif #if defined (__IPHONE_11_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_11_0 if (@available (iOS 11.0, *)) { JUCE_CAMERA_LOG ("Min available video zoom factor: " + String (device.minAvailableVideoZoomFactor)); JUCE_CAMERA_LOG ("Max available video zoom factor: " + String (device.maxAvailableVideoZoomFactor)); JUCE_CAMERA_LOG ("Dual camera switch over video zoom factor: " + String (device.dualCameraSwitchOverVideoZoomFactor)); } #endif JUCE_CAMERA_LOG ("Capture formats start-------------------"); for (AVCaptureDeviceFormat* format in device.formats) { JUCE_CAMERA_LOG ("Capture format start------"); printDebugCameraFormatInfo (format); JUCE_CAMERA_LOG ("Capture format end------"); } JUCE_CAMERA_LOG ("Capture formats end-------------------"); } static void printDebugCameraFormatInfo (AVCaptureDeviceFormat* format) { JUCE_CAMERA_LOG ("Media type: " + nsStringToJuce (format.mediaType)); #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) { String colourSpaces; for (NSNumber* number in format.supportedColorSpaces) { switch ([number intValue]) { case AVCaptureColorSpace_sRGB: colourSpaces << "sRGB "; break; case AVCaptureColorSpace_P3_D65: colourSpaces << "P3_D65 "; break; default: break; } } JUCE_CAMERA_LOG ("Supported colour spaces: " + colourSpaces); } #endif JUCE_CAMERA_LOG ("Video field of view: " + String (format.videoFieldOfView)); JUCE_CAMERA_LOG ("Video max zoom factor: " + String (format.videoMaxZoomFactor)); JUCE_CAMERA_LOG ("Video zoom factor upscale threshold: " + String (format.videoZoomFactorUpscaleThreshold)); String videoFrameRateRangesString = "Video supported frame rate ranges: "; for (AVFrameRateRange* range in format.videoSupportedFrameRateRanges) videoFrameRateRangesString << frameRateRangeToString (range); JUCE_CAMERA_LOG (videoFrameRateRangesString); JUCE_CAMERA_LOG ("Video binned: " + String (int (format.videoBinned))); JUCE_CAMERA_LOG ("Video HDR supported: " + String (int (format.videoHDRSupported))); JUCE_CAMERA_LOG ("High resolution still image dimensions: " + getHighResStillImgDimensionsString (format.highResolutionStillImageDimensions)); JUCE_CAMERA_LOG ("Min ISO: " + String (format.minISO)); JUCE_CAMERA_LOG ("Max ISO: " + String (format.maxISO)); JUCE_CAMERA_LOG ("Min exposure duration: " + cmTimeToString (format.minExposureDuration)); String autoFocusSystemString; switch (format.autoFocusSystem) { case AVCaptureAutoFocusSystemPhaseDetection: autoFocusSystemString = "PhaseDetection"; break; case AVCaptureAutoFocusSystemContrastDetection: autoFocusSystemString = "ContrastDetection"; break; case AVCaptureAutoFocusSystemNone: default: autoFocusSystemString = "None"; } JUCE_CAMERA_LOG ("Auto focus system: " + autoFocusSystemString); JUCE_CAMERA_LOG ("Standard (iOS 5.0) video stabilization supported: " + String ((int) [format isVideoStabilizationModeSupported: AVCaptureVideoStabilizationModeStandard])); JUCE_CAMERA_LOG ("Cinematic video stabilization supported: " + String ((int) [format isVideoStabilizationModeSupported: AVCaptureVideoStabilizationModeCinematic])); JUCE_CAMERA_LOG ("Auto video stabilization supported: " + String ((int) [format isVideoStabilizationModeSupported: AVCaptureVideoStabilizationModeAuto])); #if defined (__IPHONE_11_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_11_0 if (@available (iOS 11.0, *)) { JUCE_CAMERA_LOG ("Min zoom factor for depth data delivery: " + String (format.videoMinZoomFactorForDepthDataDelivery)); JUCE_CAMERA_LOG ("Max zoom factor for depth data delivery: " + String (format.videoMaxZoomFactorForDepthDataDelivery)); } #endif } static String getHighResStillImgDimensionsString (CMVideoDimensions d) { return "[" + String (d.width) + " " + String (d.height) + "]"; } static String cmTimeToString (CMTime time) { CFUniquePtr timeDesc (CMTimeCopyDescription (nullptr, time)); return String::fromCFString (timeDesc.get()); } static String frameRateRangeToString (AVFrameRateRange* range) { String result; result << "[minFrameDuration: " + cmTimeToString (range.minFrameDuration); result << " maxFrameDuration: " + cmTimeToString (range.maxFrameDuration); result << " minFrameRate: " + String (range.minFrameRate); result << " maxFrameRate: " + String (range.maxFrameRate) << "] "; return result; } //============================================================================== class CaptureSession { public: CaptureSession (Pimpl& ownerToUse, bool useHighQuality) : owner (ownerToUse), captureSessionQueue (dispatch_queue_create ("JuceCameraDeviceBackgroundDispatchQueue", DISPATCH_QUEUE_SERIAL)), captureSession ([[AVCaptureSession alloc] init]), delegate (nullptr), stillPictureTaker (*this), videoRecorder (*this) { static SessionDelegateClass cls; delegate.reset ([cls.createInstance() init]); SessionDelegateClass::setOwner (delegate.get(), this); JUCE_BEGIN_IGNORE_WARNINGS_GCC_LIKE ("-Wundeclared-selector") [[NSNotificationCenter defaultCenter] addObserver: delegate.get() selector: @selector (sessionDidStartRunning:) name: AVCaptureSessionDidStartRunningNotification object: captureSession.get()]; [[NSNotificationCenter defaultCenter] addObserver: delegate.get() selector: @selector (sessionDidStopRunning:) name: AVCaptureSessionDidStopRunningNotification object: captureSession.get()]; [[NSNotificationCenter defaultCenter] addObserver: delegate.get() selector: @selector (runtimeError:) name: AVCaptureSessionRuntimeErrorNotification object: captureSession.get()]; [[NSNotificationCenter defaultCenter] addObserver: delegate.get() selector: @selector (sessionWasInterrupted:) name: AVCaptureSessionWasInterruptedNotification object: captureSession.get()]; [[NSNotificationCenter defaultCenter] addObserver: delegate.get() selector: @selector (sessionInterruptionEnded:) name: AVCaptureSessionInterruptionEndedNotification object: captureSession.get()]; JUCE_END_IGNORE_WARNINGS_GCC_LIKE dispatch_async (captureSessionQueue,^ { [captureSession.get() setSessionPreset: useHighQuality ? AVCaptureSessionPresetHigh : AVCaptureSessionPresetMedium]; }); ++numCaptureSessions; } ~CaptureSession() { [[NSNotificationCenter defaultCenter] removeObserver: delegate.get()]; stopRecording(); if (--numCaptureSessions == 0) { dispatch_async (captureSessionQueue, ^ { if (captureSession.get().running) [captureSession.get() stopRunning]; sessionClosedEvent.signal(); }); sessionClosedEvent.wait (-1); } } bool openedOk() const noexcept { return sessionStarted; } void startSessionForDeviceWithId (const String& cameraIdToUse) { dispatch_async (captureSessionQueue,^ { cameraDevice = [AVCaptureDevice deviceWithUniqueID: juceStringToNS (cameraIdToUse)]; auto audioDevice = [AVCaptureDevice defaultDeviceWithMediaType: AVMediaTypeAudio]; [captureSession.get() beginConfiguration]; // This will add just video... auto error = addInputToDevice (cameraDevice); if (error.isNotEmpty()) { MessageManager::callAsync ([weakRef = WeakReference { this }, error]() mutable { if (weakRef != nullptr) weakRef->owner.cameraOpenCallback ({}, error); }); return; } // ... so add audio explicitly here error = addInputToDevice (audioDevice); if (error.isNotEmpty()) { MessageManager::callAsync ([weakRef = WeakReference { this }, error]() mutable { if (weakRef != nullptr) weakRef->owner.cameraOpenCallback ({}, error); }); return; } [captureSession.get() commitConfiguration]; if (! captureSession.get().running) [captureSession.get() startRunning]; }); } AVCaptureVideoPreviewLayer* createPreviewLayer() { if (! openedOk()) { // A session must be started first! jassertfalse; return nullptr; } previewLayer = [AVCaptureVideoPreviewLayer layerWithSession: captureSession.get()]; return previewLayer; } void takeStillPicture() { if (! openedOk()) { // A session must be started first! jassert (openedOk()); return; } stillPictureTaker.takePicture (previewLayer.connection.videoOrientation); } void startRecording (const File& file) { if (! openedOk()) { // A session must be started first! jassertfalse; return; } if (file.existsAsFile()) { // File overwriting is not supported by iOS video recorder, the target // file must not exist. jassertfalse; return; } videoRecorder.startRecording (file, previewLayer.connection.videoOrientation); } void stopRecording() { videoRecorder.stopRecording(); } Time getTimeOfFirstRecordedFrame() const { return videoRecorder.getTimeOfFirstRecordedFrame(); } JUCE_DECLARE_WEAK_REFERENCEABLE (CaptureSession) private: String addInputToDevice (AVCaptureDevice* device) { NSError* error = nil; auto input = [AVCaptureDeviceInput deviceInputWithDevice: device error: &error]; if (error != nil) return nsStringToJuce (error.localizedDescription); if (! [captureSession.get() canAddInput: input]) return "Could not add input to camera session."; [captureSession.get() addInput: input]; return {}; } //============================================================================== struct SessionDelegateClass : public ObjCClass { SessionDelegateClass() : ObjCClass ("SessionDelegateClass_") { JUCE_BEGIN_IGNORE_WARNINGS_GCC_LIKE ("-Wundeclared-selector") addMethod (@selector (sessionDidStartRunning:), started, "v@:@"); addMethod (@selector (sessionDidStopRunning:), stopped, "v@:@"); addMethod (@selector (runtimeError:), runtimeError, "v@:@"); addMethod (@selector (sessionWasInterrupted:), interrupted, "v@:@"); addMethod (@selector (sessionInterruptionEnded:), interruptionEnded, "v@:@"); JUCE_END_IGNORE_WARNINGS_GCC_LIKE addIvar ("owner"); registerClass(); } //============================================================================== static CaptureSession& getOwner (id self) { return *getIvar (self, "owner"); } static void setOwner (id self, CaptureSession* s) { object_setInstanceVariable (self, "owner", s); } private: //============================================================================== static void started (id self, SEL, NSNotification* notification) { JUCE_CAMERA_LOG (nsStringToJuce ([notification description])); ignoreUnused (notification); dispatch_async (dispatch_get_main_queue(), ^{ getOwner (self).cameraSessionStarted(); }); } static void stopped (id, SEL, NSNotification* notification) { JUCE_CAMERA_LOG (nsStringToJuce ([notification description])); ignoreUnused (notification); } static void runtimeError (id self, SEL, NSNotification* notification) { JUCE_CAMERA_LOG (nsStringToJuce ([notification description])); dispatch_async (dispatch_get_main_queue(), ^{ NSError* error = notification.userInfo[AVCaptureSessionErrorKey]; auto errorString = error != nil ? nsStringToJuce (error.localizedDescription) : String(); getOwner (self).cameraSessionRuntimeError (errorString); }); } static void interrupted (id, SEL, NSNotification* notification) { JUCE_CAMERA_LOG (nsStringToJuce ([notification description])); ignoreUnused (notification); } static void interruptionEnded (id, SEL, NSNotification* notification) { JUCE_CAMERA_LOG (nsStringToJuce ([notification description])); ignoreUnused (notification); } }; //============================================================================== class StillPictureTaker { public: StillPictureTaker (CaptureSession& cs) : captureSession (cs), captureOutput (createCaptureOutput()), photoOutputDelegate (nullptr) { if (@available (iOS 10.0, *)) { static PhotoOutputDelegateClass cls; photoOutputDelegate.reset ([cls.createInstance() init]); PhotoOutputDelegateClass::setOwner (photoOutputDelegate.get(), this); } captureSession.addOutputIfPossible (captureOutput); } void takePicture (AVCaptureVideoOrientation orientationToUse) { if (takingPicture) { // Picture taking already in progress! jassertfalse; return; } takingPicture = true; printImageOutputDebugInfo (captureOutput); if (auto* connection = findVideoConnection (captureOutput)) { #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) { if ([captureOutput isKindOfClass: [AVCapturePhotoOutput class]]) { auto* photoOutput = (AVCapturePhotoOutput*) captureOutput; auto outputConnection = [photoOutput connectionWithMediaType: AVMediaTypeVideo]; outputConnection.videoOrientation = orientationToUse; [photoOutput capturePhotoWithSettings: [AVCapturePhotoSettings photoSettings] delegate: id (photoOutputDelegate.get())]; return; } } #endif auto* stillImageOutput = (AVCaptureStillImageOutput*) captureOutput; auto outputConnection = [stillImageOutput connectionWithMediaType: AVMediaTypeVideo]; outputConnection.videoOrientation = orientationToUse; [stillImageOutput captureStillImageAsynchronouslyFromConnection: connection completionHandler: ^(CMSampleBufferRef imageSampleBuffer, NSError* error) { takingPicture = false; if (error != nil) { JUCE_CAMERA_LOG ("Still picture capture failed, error: " + nsStringToJuce (error.localizedDescription)); jassertfalse; return; } NSData* imageData = [AVCaptureStillImageOutput jpegStillImageNSDataRepresentation: imageSampleBuffer]; auto image = ImageFileFormat::loadFrom (imageData.bytes, (size_t) imageData.length); callListeners (image); MessageManager::callAsync ([this, image] { notifyPictureTaken (image); }); }]; } else { // Could not find a connection of video type jassertfalse; } } private: static AVCaptureOutput* createCaptureOutput() { #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) return [AVCapturePhotoOutput new]; #endif return [AVCaptureStillImageOutput new]; } static void printImageOutputDebugInfo (AVCaptureOutput* captureOutput) { #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) { if ([captureOutput isKindOfClass: [AVCapturePhotoOutput class]]) { auto* photoOutput = (AVCapturePhotoOutput*) captureOutput; String typesString; for (AVVideoCodecType type in photoOutput.availablePhotoCodecTypes) typesString << nsStringToJuce (type) << " "; JUCE_CAMERA_LOG ("Available image codec types: " + typesString); JUCE_CAMERA_LOG ("Still image stabilization supported: " + String ((int) photoOutput.stillImageStabilizationSupported)); JUCE_CAMERA_LOG ("Dual camera fusion supported: " + String ((int) photoOutput.dualCameraFusionSupported)); JUCE_CAMERA_LOG ("Supports flash: " + String ((int) [photoOutput.supportedFlashModes containsObject: @(AVCaptureFlashModeOn)])); JUCE_CAMERA_LOG ("Supports auto flash: " + String ((int) [photoOutput.supportedFlashModes containsObject: @(AVCaptureFlashModeAuto)])); JUCE_CAMERA_LOG ("Max bracketed photo count: " + String (photoOutput.maxBracketedCapturePhotoCount)); JUCE_CAMERA_LOG ("Lens stabilization during bracketed capture supported: " + String ((int) photoOutput.lensStabilizationDuringBracketedCaptureSupported)); JUCE_CAMERA_LOG ("Live photo capture supported: " + String ((int) photoOutput.livePhotoCaptureSupported)); #if defined (__IPHONE_11_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_11_0 if (@available (iOS 11.0, *)) { typesString.clear(); for (AVFileType type in photoOutput.availablePhotoFileTypes) typesString << nsStringToJuce (type) << " "; JUCE_CAMERA_LOG ("Available photo file types: " + typesString); typesString.clear(); for (AVFileType type in photoOutput.availableRawPhotoFileTypes) typesString << nsStringToJuce (type) << " "; JUCE_CAMERA_LOG ("Available RAW photo file types: " + typesString); typesString.clear(); for (AVFileType type in photoOutput.availableLivePhotoVideoCodecTypes) typesString << nsStringToJuce (type) << " "; JUCE_CAMERA_LOG ("Available live photo video codec types: " + typesString); JUCE_CAMERA_LOG ("Dual camera dual photo delivery supported: " + String ((int) photoOutput.dualCameraDualPhotoDeliverySupported)); JUCE_CAMERA_LOG ("Camera calibration data delivery supported: " + String ((int) photoOutput.cameraCalibrationDataDeliverySupported)); JUCE_CAMERA_LOG ("Depth data delivery supported: " + String ((int) photoOutput.depthDataDeliverySupported)); } #endif return; } } #endif auto* stillImageOutput = (AVCaptureStillImageOutput*) captureOutput; String typesString; for (AVVideoCodecType type in stillImageOutput.availableImageDataCodecTypes) typesString << nsStringToJuce (type) << " "; JUCE_CAMERA_LOG ("Available image codec types: " + typesString); JUCE_CAMERA_LOG ("Still image stabilization supported: " + String ((int) stillImageOutput.stillImageStabilizationSupported)); JUCE_CAMERA_LOG ("Automatically enables still image stabilization when available: " + String ((int) stillImageOutput.automaticallyEnablesStillImageStabilizationWhenAvailable)); JUCE_CAMERA_LOG ("Output settings for image output: " + nsStringToJuce ([stillImageOutput.outputSettings description])); } //============================================================================== static AVCaptureConnection* findVideoConnection (AVCaptureOutput* output) { for (AVCaptureConnection* connection in output.connections) for (AVCaptureInputPort* port in connection.inputPorts) if ([port.mediaType isEqual: AVMediaTypeVideo]) return connection; return nullptr; } //============================================================================== JUCE_BEGIN_IGNORE_WARNINGS_GCC_LIKE ("-Wunguarded-availability", "-Wunguarded-availability-new") class PhotoOutputDelegateClass : public ObjCClass { public: PhotoOutputDelegateClass() : ObjCClass ("PhotoOutputDelegateClass_") { addMethod (@selector (captureOutput:willBeginCaptureForResolvedSettings:), willBeginCaptureForSettings, "v@:@@"); addMethod (@selector (captureOutput:willCapturePhotoForResolvedSettings:), willCaptureForSettings, "v@:@@"); addMethod (@selector (captureOutput:didCapturePhotoForResolvedSettings:), didCaptureForSettings, "v@:@@"); addMethod (@selector (captureOutput:didFinishCaptureForResolvedSettings:error:), didFinishCaptureForSettings, "v@:@@@"); #if defined (__IPHONE_11_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_11_0 if (@available (iOS 11.0, *)) { addMethod (@selector (captureOutput:didFinishProcessingPhoto:error:), didFinishProcessingPhoto, "v@:@@@"); } else #endif { addMethod (@selector (captureOutput:didFinishProcessingPhotoSampleBuffer:previewPhotoSampleBuffer:resolvedSettings:bracketSettings:error:), didFinishProcessingPhotoSampleBuffer, "v@:@@@@@@"); } addIvar ("owner"); registerClass(); } //============================================================================== static StillPictureTaker& getOwner (id self) { return *getIvar (self, "owner"); } static void setOwner (id self, StillPictureTaker* t) { object_setInstanceVariable (self, "owner", t); } private: static void willBeginCaptureForSettings (id, SEL, AVCapturePhotoOutput*, AVCaptureResolvedPhotoSettings*) { JUCE_CAMERA_LOG ("willBeginCaptureForSettings()"); } static void willCaptureForSettings (id, SEL, AVCapturePhotoOutput*, AVCaptureResolvedPhotoSettings*) { JUCE_CAMERA_LOG ("willCaptureForSettings()"); } static void didCaptureForSettings (id, SEL, AVCapturePhotoOutput*, AVCaptureResolvedPhotoSettings*) { JUCE_CAMERA_LOG ("didCaptureForSettings()"); } static void didFinishCaptureForSettings (id, SEL, AVCapturePhotoOutput*, AVCaptureResolvedPhotoSettings*, NSError* error) { String errorString = error != nil ? nsStringToJuce (error.localizedDescription) : String(); ignoreUnused (errorString); JUCE_CAMERA_LOG ("didFinishCaptureForSettings(), error = " + errorString); } static void didFinishProcessingPhoto (id self, SEL, AVCapturePhotoOutput*, AVCapturePhoto* capturePhoto, NSError* error) { getOwner (self).takingPicture = false; String errorString = error != nil ? nsStringToJuce (error.localizedDescription) : String(); ignoreUnused (errorString); JUCE_CAMERA_LOG ("didFinishProcessingPhoto(), error = " + errorString); if (error != nil) { JUCE_CAMERA_LOG ("Still picture capture failed, error: " + nsStringToJuce (error.localizedDescription)); jassertfalse; return; } auto* imageOrientation = (NSNumber *) capturePhoto.metadata[(NSString*) kCGImagePropertyOrientation]; auto* uiImage = getImageWithCorrectOrientation ((CGImagePropertyOrientation) imageOrientation.unsignedIntValue, [capturePhoto CGImageRepresentation]); auto* imageData = UIImageJPEGRepresentation (uiImage, 0.f); auto image = ImageFileFormat::loadFrom (imageData.bytes, (size_t) imageData.length); getOwner (self).callListeners (image); MessageManager::callAsync ([self, image]() { getOwner (self).notifyPictureTaken (image); }); } static UIImage* getImageWithCorrectOrientation (CGImagePropertyOrientation imageOrientation, CGImageRef imageData) { auto origWidth = CGImageGetWidth (imageData); auto origHeight = CGImageGetHeight (imageData); auto targetSize = getTargetImageDimensionFor (imageOrientation, imageData); UIGraphicsBeginImageContext (targetSize); CGContextRef context = UIGraphicsGetCurrentContext(); switch (imageOrientation) { case kCGImagePropertyOrientationUp: CGContextScaleCTM (context, 1.0, -1.0); CGContextTranslateCTM (context, 0.0, -targetSize.height); break; case kCGImagePropertyOrientationRight: CGContextRotateCTM (context, 90 * MathConstants::pi / 180); CGContextScaleCTM (context, targetSize.height / origHeight, -targetSize.width / origWidth); break; case kCGImagePropertyOrientationDown: CGContextTranslateCTM (context, targetSize.width, 0.0); CGContextScaleCTM (context, -1.0, 1.0); break; case kCGImagePropertyOrientationLeft: CGContextRotateCTM (context, -90 * MathConstants::pi / 180); CGContextScaleCTM (context, targetSize.height / origHeight, -targetSize.width / origWidth); CGContextTranslateCTM (context, -targetSize.width, -targetSize.height); break; case kCGImagePropertyOrientationUpMirrored: case kCGImagePropertyOrientationDownMirrored: case kCGImagePropertyOrientationLeftMirrored: case kCGImagePropertyOrientationRightMirrored: default: // Not implemented. jassertfalse; break; } CGContextDrawImage (context, CGRectMake (0, 0, targetSize.width, targetSize.height), imageData); UIImage* correctedImage = UIGraphicsGetImageFromCurrentImageContext(); UIGraphicsEndImageContext(); return correctedImage; } static CGSize getTargetImageDimensionFor (CGImagePropertyOrientation imageOrientation, CGImageRef imageData) { auto width = CGImageGetWidth (imageData); auto height = CGImageGetHeight (imageData); switch (imageOrientation) { case kCGImagePropertyOrientationUp: case kCGImagePropertyOrientationUpMirrored: case kCGImagePropertyOrientationDown: case kCGImagePropertyOrientationDownMirrored: return CGSizeMake ((CGFloat) width, (CGFloat) height); case kCGImagePropertyOrientationRight: case kCGImagePropertyOrientationRightMirrored: case kCGImagePropertyOrientationLeft: case kCGImagePropertyOrientationLeftMirrored: return CGSizeMake ((CGFloat) height, (CGFloat) width); } jassertfalse; return CGSizeMake ((CGFloat) width, (CGFloat) height); } static void didFinishProcessingPhotoSampleBuffer (id self, SEL, AVCapturePhotoOutput*, CMSampleBufferRef imageBuffer, CMSampleBufferRef imagePreviewBuffer, AVCaptureResolvedPhotoSettings*, AVCaptureBracketedStillImageSettings*, NSError* error) { getOwner (self).takingPicture = false; String errorString = error != nil ? nsStringToJuce (error.localizedDescription) : String(); ignoreUnused (errorString); JUCE_CAMERA_LOG ("didFinishProcessingPhotoSampleBuffer(), error = " + errorString); if (error != nil) { JUCE_CAMERA_LOG ("Still picture capture failed, error: " + nsStringToJuce (error.localizedDescription)); jassertfalse; return; } NSData* origImageData = [AVCapturePhotoOutput JPEGPhotoDataRepresentationForJPEGSampleBuffer: imageBuffer previewPhotoSampleBuffer: imagePreviewBuffer]; auto origImage = [UIImage imageWithData: origImageData]; auto imageOrientation = uiImageOrientationToCGImageOrientation (origImage.imageOrientation); auto* uiImage = getImageWithCorrectOrientation (imageOrientation, origImage.CGImage); auto* imageData = UIImageJPEGRepresentation (uiImage, 0.f); auto image = ImageFileFormat::loadFrom (imageData.bytes, (size_t) imageData.length); getOwner (self).callListeners (image); MessageManager::callAsync ([self, image]() { getOwner (self).notifyPictureTaken (image); }); } static CGImagePropertyOrientation uiImageOrientationToCGImageOrientation (UIImageOrientation orientation) { switch (orientation) { case UIImageOrientationUp: return kCGImagePropertyOrientationUp; case UIImageOrientationDown: return kCGImagePropertyOrientationDown; case UIImageOrientationLeft: return kCGImagePropertyOrientationLeft; case UIImageOrientationRight: return kCGImagePropertyOrientationRight; case UIImageOrientationUpMirrored: return kCGImagePropertyOrientationUpMirrored; case UIImageOrientationDownMirrored: return kCGImagePropertyOrientationDownMirrored; case UIImageOrientationLeftMirrored: return kCGImagePropertyOrientationLeftMirrored; case UIImageOrientationRightMirrored: return kCGImagePropertyOrientationRightMirrored; } } }; JUCE_END_IGNORE_WARNINGS_GCC_LIKE //============================================================================== void callListeners (const Image& image) { captureSession.callListeners (image); } void notifyPictureTaken (const Image& image) { captureSession.notifyPictureTaken (image); } CaptureSession& captureSession; AVCaptureOutput* captureOutput; std::unique_ptr photoOutputDelegate; bool takingPicture = false; }; //============================================================================== // NB: FileOutputRecordingDelegateClass callbacks can be called from any thread (incl. // the message thread), so waiting for an event when stopping recording is not an // option and VideoRecorder must be alive at all times in order to get stopped // recording callback. class VideoRecorder { public: VideoRecorder (CaptureSession& session) : movieFileOutput ([AVCaptureMovieFileOutput new]), delegate (nullptr) { static FileOutputRecordingDelegateClass cls; delegate.reset ([cls.createInstance() init]); FileOutputRecordingDelegateClass::setOwner (delegate.get(), this); session.addOutputIfPossible (movieFileOutput); } ~VideoRecorder() { stopRecording(); // Shutting down a device while recording will stop the recording // abruptly and the recording will be lost. jassert (! recordingInProgress); } void startRecording (const File& file, AVCaptureVideoOrientation orientationToUse) { #if defined (__IPHONE_10_0) && __IPHONE_OS_VERSION_MAX_ALLOWED >= __IPHONE_10_0 if (@available (iOS 10.0, *)) printVideoOutputDebugInfo (movieFileOutput); #endif auto url = [NSURL fileURLWithPath: juceStringToNS (file.getFullPathName()) isDirectory: NO]; auto outputConnection = [movieFileOutput connectionWithMediaType: AVMediaTypeVideo]; outputConnection.videoOrientation = orientationToUse; [movieFileOutput startRecordingToOutputFileURL: url recordingDelegate: delegate.get()]; } void stopRecording() { [movieFileOutput stopRecording]; } Time getTimeOfFirstRecordedFrame() const { return Time (firstRecordedFrameTimeMs.get()); } private: static void printVideoOutputDebugInfo (AVCaptureMovieFileOutput* output) { ignoreUnused (output); JUCE_CAMERA_LOG ("Available video codec types:"); #if JUCE_CAMERA_LOG_ENABLED for (AVVideoCodecType type in output.availableVideoCodecTypes) JUCE_CAMERA_LOG (nsStringToJuce (type)); #endif JUCE_CAMERA_LOG ("Output settings per video connection:"); #if JUCE_CAMERA_LOG_ENABLED for (AVCaptureConnection* connection in output.connections) JUCE_CAMERA_LOG (nsStringToJuce ([[output outputSettingsForConnection: connection] description])); #endif } //============================================================================== struct FileOutputRecordingDelegateClass : public ObjCClass> { FileOutputRecordingDelegateClass() : ObjCClass> ("FileOutputRecordingDelegateClass_") { addMethod (@selector (captureOutput:didStartRecordingToOutputFileAtURL:fromConnections:), started, "v@:@@@"); addMethod (@selector (captureOutput:didFinishRecordingToOutputFileAtURL:fromConnections:error:), stopped, "v@:@@@@"); addIvar ("owner"); registerClass(); } //============================================================================== static VideoRecorder& getOwner (id self) { return *getIvar (self, "owner"); } static void setOwner (id self, VideoRecorder* r) { object_setInstanceVariable (self, "owner", r); } private: static void started (id self, SEL, AVCaptureFileOutput*, NSURL*, NSArray*) { JUCE_CAMERA_LOG ("Started recording"); getOwner (self).firstRecordedFrameTimeMs.set (Time::getCurrentTime().toMilliseconds()); getOwner (self).recordingInProgress = true; } static void stopped (id self, SEL, AVCaptureFileOutput*, NSURL*, NSArray*, NSError* error) { String errorString; bool recordingPlayable = true; // There might have been an error in the recording, yet there may be a playable file... if ([error code] != noErr) { id value = [[error userInfo] objectForKey: AVErrorRecordingSuccessfullyFinishedKey]; if (value != nil && ! [value boolValue]) recordingPlayable = false; errorString = nsStringToJuce (error.localizedDescription) + ", playable: " + String ((int) recordingPlayable); } JUCE_CAMERA_LOG ("Stopped recording, error = " + errorString); getOwner (self).recordingInProgress = false; } }; AVCaptureMovieFileOutput* movieFileOutput; std::unique_ptr, NSObjectDeleter> delegate; bool recordingInProgress = false; Atomic firstRecordedFrameTimeMs { 0 }; }; //============================================================================== void addOutputIfPossible (AVCaptureOutput* output) { dispatch_async (captureSessionQueue,^ { if ([captureSession.get() canAddOutput: output]) { [captureSession.get() beginConfiguration]; [captureSession.get() addOutput: output]; [captureSession.get() commitConfiguration]; return; } // Can't add output to camera session! jassertfalse; }); } //============================================================================== void cameraSessionStarted() { sessionStarted = true; owner.cameraSessionStarted(); } void cameraSessionRuntimeError (const String& error) { owner.cameraSessionRuntimeError (error); } void callListeners (const Image& image) { owner.callListeners (image); } void notifyPictureTaken (const Image& image) { owner.notifyPictureTaken (image); } Pimpl& owner; dispatch_queue_t captureSessionQueue; std::unique_ptr captureSession; std::unique_ptr delegate; StillPictureTaker stillPictureTaker; VideoRecorder videoRecorder; AVCaptureDevice* cameraDevice = nil; AVCaptureVideoPreviewLayer* previewLayer = nil; bool sessionStarted = false; WaitableEvent sessionClosedEvent; static int numCaptureSessions; }; //============================================================================== void cameraSessionStarted() { JUCE_CAMERA_LOG ("cameraSessionStarted()"); cameraOpenCallback (cameraId, {}); } void cameraSessionRuntimeError (const String& error) { JUCE_CAMERA_LOG ("cameraSessionRuntimeError(), error = " + error); if (! notifiedOfCameraOpening) { cameraOpenCallback ({}, error); } else { if (owner.onErrorOccurred != nullptr) owner.onErrorOccurred (error); } } void callListeners (const Image& image) { const ScopedLock sl (listenerLock); listeners.call ([=] (Listener& l) { l.imageReceived (image); }); if (listeners.size() == 1) triggerStillPictureCapture(); } void notifyPictureTaken (const Image& image) { JUCE_CAMERA_LOG ("notifyPictureTaken()"); if (pictureTakenCallback != nullptr) pictureTakenCallback (image); } //============================================================================== void triggerStillPictureCapture() { captureSession.takeStillPicture(); } //============================================================================== CameraDevice& owner; String cameraId; InternalOpenCameraResultCallback cameraOpenCallback; CriticalSection listenerLock; ListenerList listeners; std::function pictureTakenCallback; CaptureSession captureSession; bool notifiedOfCameraOpening = false; friend struct CameraDevice::ViewerComponent; //============================================================================== JUCE_DECLARE_NON_COPYABLE (Pimpl) }; int CameraDevice::Pimpl::CaptureSession::numCaptureSessions = 0; //============================================================================== struct CameraDevice::ViewerComponent : public UIViewComponent { //============================================================================== struct JuceCameraDeviceViewerClass : public ObjCClass { JuceCameraDeviceViewerClass() : ObjCClass ("JuceCameraDeviceViewerClass_") { addMethod (@selector (layoutSubviews), layoutSubviews, "v@:"); registerClass(); } private: static void layoutSubviews (id self, SEL) { sendSuperclassMessage (self, @selector (layoutSubviews)); UIView* asUIView = (UIView*) self; updateOrientation (self); if (auto* previewLayer = getPreviewLayer (self)) previewLayer.frame = asUIView.bounds; } static AVCaptureVideoPreviewLayer* getPreviewLayer (id self) { UIView* asUIView = (UIView*) self; if (asUIView.layer.sublayers != nil && [asUIView.layer.sublayers count] > 0) if ([asUIView.layer.sublayers[0] isKindOfClass: [AVCaptureVideoPreviewLayer class]]) return (AVCaptureVideoPreviewLayer*) asUIView.layer.sublayers[0]; return nil; } static void updateOrientation (id self) { if (auto* previewLayer = getPreviewLayer (self)) { UIDeviceOrientation o = [UIDevice currentDevice].orientation; if (UIDeviceOrientationIsPortrait (o) || UIDeviceOrientationIsLandscape (o)) { if (previewLayer.connection != nil) previewLayer.connection.videoOrientation = (AVCaptureVideoOrientation) o; } } } }; ViewerComponent (CameraDevice& device) { static JuceCameraDeviceViewerClass cls; // Initial size that can be overriden later. setSize (640, 480); auto view = [cls.createInstance() init]; setView (view); auto* previewLayer = device.pimpl->captureSession.createPreviewLayer(); previewLayer.frame = view.bounds; UIInterfaceOrientation statusBarOrientation = [UIApplication sharedApplication].statusBarOrientation; AVCaptureVideoOrientation videoOrientation = statusBarOrientation != UIInterfaceOrientationUnknown ? (AVCaptureVideoOrientation) statusBarOrientation : AVCaptureVideoOrientationPortrait; previewLayer.connection.videoOrientation = videoOrientation; [view.layer addSublayer: previewLayer]; } }; //============================================================================== String CameraDevice::getFileExtension() { return ".mov"; } #if JUCE_DEPRECATION_IGNORED JUCE_END_IGNORE_WARNINGS_GCC_LIKE #endif