US12489874B2 - Stereoscopic capture using cameras with different fields of view - Google Patents
Stereoscopic capture using cameras with different fields of viewInfo
- Publication number
- US12489874B2 US12489874B2 US18/645,357 US202418645357A US12489874B2 US 12489874 B2 US12489874 B2 US 12489874B2 US 202418645357 A US202418645357 A US 202418645357A US 12489874 B2 US12489874 B2 US 12489874B2
- Authority
- US
- United States
- Prior art keywords
- image
- rectified
- stabilized
- view
- camera
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/261—Image signal generators with monoscopic-to-stereoscopic image conversion
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
- G06T7/85—Stereo camera calibration
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/122—Improving the 3D impression of stereoscopic images by modifying image signal contents, e.g. by filtering or adding monoscopic depth cues
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/161—Encoding, multiplexing or demultiplexing different image signal components
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/167—Synchronising or controlling image signals
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/239—Image signal generators using stereoscopic image cameras using two 2D image sensors having a relative position equal to or related to the interocular distance
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/246—Calibration of cameras
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/296—Synchronisation thereof; Control thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/57—Mechanical or electrical details of cameras or camera modules specially adapted for being embedded in other devices
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/681—Motion detection
- H04N23/6812—Motion detection based on additional sensors, e.g. acceleration sensors
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
- H04N23/683—Vibration or motion blur correction performed by a processor, e.g. controlling the readout of an image memory
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/68—Control of cameras or camera modules for stable pick-up of the scene, e.g. compensating for camera body vibrations
- H04N23/682—Vibration or motion blur correction
- H04N23/685—Vibration or motion blur correction performed by mechanical compensation
- H04N23/687—Vibration or motion blur correction performed by mechanical compensation by shifting the lens or sensor position
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/90—Arrangement of cameras or camera modules, e.g. multiple cameras in TV studios or sports stadiums
Definitions
- This relates generally to electronic devices, and, more particularly, to electronic devices with cameras.
- Electronic devices can include one or more cameras for capturing an image or a video feed of scene.
- Electronic devices can include a wide camera having a first field of view, an ultrawide camera having a second field of view that is greater than the first field of view, and a telephoto camera having a third field of view that is smaller than the first field of view.
- An electronic device may include one or more cameras for capturing an image or a video feed of a real-world environment.
- the electronic device can include a wide image sensor configured to capture corresponding wide images, a ultra wide image sensor configured to capture corresponding ultra wide images, and control circuitry for processing and combining the wide images with the ultra wide images to generate stereoscopic images.
- An aspect of the disclosure provides a method of operating an electronic device that includes capturing a first image using a first image sensor having a first field of view, capturing a second image using a second image sensor having a second field of view different than the first field of view, and outputting stereoscopic content based on the first image captured using the first image sensor with the first field of view and the second image captured using the second image sensor.
- the method can include processing the first image using a first image signal processor to output a first processed image and processing the second image using a second image signal processor to output a second processed image.
- the method can include performing stereo rectification and image stabilization operations on the first and second processed images to output corresponding first and second rectified and stabilized images and compressing the first and second rectified and stabilized images using a multiview video encoding scheme to generate a stereoscope video stream.
- the method can include obtaining stabilization information for the first image sensor and synchronizing image stabilization between the first and second image sensors by applying the stabilization information for the first image sensor to the second processed image captured using the second image sensor.
- the method can include obtaining a rotation matrix associated with the first image sensor, calculating a rectified and stabilized pose based on the rotation matrix and motion data associated with the first image sensor, rectifying a pose of the first image sensor, and stabilizing the rectified posed of the first image sensor using a temporal filter.
- the method can further include calculating a first homography based on the rectified and stabilized pose, the motion data associated with the first image sensor, intrinsic data associated with the first image sensor, and intrinsic data associated with a target image sensor and calculating a second homography based on the rectified and stabilized pose, motion data associated with the second image sensor, intrinsic data associated with the second image sensor, and the intrinsic data associated with the target image sensor.
- the method can include warping the first processed image using the first homography to generate the first rectified and stabilized image and warping the second processed image using the second homography to generate the second rectified and stabilized image.
- the method can further include generating a still stereoscopic pair.
- An aspect of the disclosure provides an electronic device that includes a first camera having a first field of view and configured to capture a first image, a second camera having a second field of view different than the first field of view and configured to capture a second image, and control circuitry configured to output stereoscopic content based on the first image captured from the first camera having the first field of view and the second image from the second camera having the second field of view different than the first field of view.
- the control circuitry can include: a first image signal processor configured to receive the first image captured by the first camera and to output a corresponding first processed image, a second image signal processor configured to receive the second image captured by the second camera and to output a corresponding second processed image, a first additional processor configured to receive the first processed image, motion data associated with the first camera, and first calibration data and further configured to generate a first rectified and stabilized image, a second additional processor configured to receive the second processed image, motion data associated with the second camera, and second calibration data and further configured to generate a second rectified and stabilized image, and a video compression block configured to receive the first and second rectified and stabilized images and to generate a corresponding stereoscopic video stream.
- the electronic device can further include circuitry for outputting still images.
- An aspect of the disclosure provides a method of operating an electronic device that includes capturing a first image using a first camera with a first field of view, capturing a second image using a second camera with a second field of view different than the first field of view, generating stereoscopic content based on the first and second captured images, and outputting a user notification to improve a quality of the stereoscopic content.
- the method can include detecting an incorrect stereoscopic capture orientation of the electronic device and then outputting a user notification to switch to a correct stereoscopic capture orientation.
- the method can include detecting whether the first camera is occluded and then outputting a user notification that the first camera is occluded.
- the method can include detecting a lighting condition of the first or second image and then outputting a user notification that the lighting condition is below the threshold.
- the method can include detecting an external object within the field of view of the first camera and then outputting a user notification that the external object is within the field of view of the first camera.
- the method can include detecting motion or jitter of the first and second cameras and then outputting a user notification to stop moving or hold still.
- FIG. 1 is a schematic diagram of an illustrative electronic device in accordance with some embodiments.
- FIG. 2 is a front perspective view of an electronic device of the type shown in FIG. 1 in accordance with some embodiments.
- FIG. 3 is a rear perspective view of an electronic device of the type shown in FIGS. 1 - 2 in accordance with some embodiments.
- FIG. 4 is a diagram of an illustrative electronic device having image sensors of different focal lengths that are used to generate a stereoscopic video stream in accordance with some embodiments.
- FIG. 5 is a diagram of illustrative operations for warping images from two image sensors in accordance with some embodiments.
- FIG. 6 is a diagram showing additional subsystems configured to generate a still stereoscopic image in accordance with some embodiments.
- FIG. 7 is a diagram showing additional subsystems coupled to image signal processors and configured to generate a still stereoscopic image in accordance with some embodiments.
- FIG. 8 is a flow chart of illustrative steps for operating an electronic device of the type shown in FIGS. 1 - 7 in accordance with some embodiments.
- FIG. 1 A schematic diagram of an illustrative electronic device 10 configured to capture stereoscopic content is shown in FIG. 1 .
- Device 10 of FIG. 1 may be operated as a stand-alone device and/or the resources of device 10 may be used to communicate with external electronic equipment.
- communications circuitry in device 10 may be used to transmit user input information, sensor information, and/or other information to external electronic devices (e.g., wirelessly or via wired connections) and/or may be used to receive such information from external electronic devices.
- Each of these external devices may include components of the type shown by device 10 of FIG. 1 .
- control circuitry 14 may include storage such as storage circuitry 16 .
- Storage circuitry 16 may include hard disk drive storage, nonvolatile memory (e.g., flash memory or other electrically-programmable-read-only memory configured to form a solid-state drive), volatile memory (e.g., static or dynamic random-access-memory), etc.
- Storage circuitry 16 may include storage that is integrated within device 10 and/or removable storage media.
- Control circuitry 14 may include processing circuitry such as processing circuitry 18 .
- Processing circuitry 18 may be used to control the operation of device 10 .
- Processing circuitry 18 may include on one or more microprocessors, microcontrollers, digital signal processors, host processors, baseband processor integrated circuits, application specific integrated circuits, central processing units (CPUs), power management units, audio chips, etc.
- Control circuitry 14 may be configured to perform operations in device 10 using hardware (e.g., dedicated hardware or circuitry), firmware, and/or software.
- Software code for performing operations in device 10 may be stored on storage circuitry 16 (e.g., storage circuitry 16 may include non-transitory (tangible) computer readable storage media that stores the software code).
- the software code may sometimes be referred to as program instructions, software, data, instructions, or code.
- Software code stored on storage circuitry 16 may be executed by processing circuitry 18 .
- Control circuitry 14 may be used to run software on device 10 such as satellite navigation applications, internet browsing applications, voice-over-internet-protocol (VOIP) telephone call applications, email applications, media playback applications, operating system functions, etc. To support interactions with external equipment, control circuitry 14 may be used in implementing communications protocols.
- VOIP voice-over-internet-protocol
- Communications protocols that may be implemented using control circuitry 14 include internet protocols, wireless local area network (WLAN) protocols (e.g., IEEE 802.11 protocols-sometimes referred to as Wi-Fi®), protocols for other short-range wireless communications links such as the Bluetooth® protocol or other wireless personal area network (WPAN) protocols, IEEE 802.11ad protocols (e.g., ultra-wideband protocols), cellular telephone protocols (e.g., 3G protocols, 4G (LTE) protocols, 5G protocols, etc.), antenna diversity protocols, satellite navigation system protocols (e.g., global positioning system (GPS) protocols, global navigation satellite system (GLONASS) protocols, etc.), antenna-based spatial ranging protocols (e.g., radio detection and ranging (RADAR) protocols or other desired range detection protocols for signals conveyed at millimeter and centimeter wave frequencies), or any other desired communications protocols.
- Each communications protocol may be associated with a corresponding radio access technology (RAT) that specifies the physical connection methodology used in implementing the protocol.
- RAT radio access technology
- control circuitry 14 may communicate using communications circuitry 20 .
- Communications circuitry 20 may include antennas, radio-frequency transceiver circuitry, and other wireless communications circuitry and/or wired communications circuitry.
- Communications circuitry 20 which may sometimes be referred to as part of control circuitry and/or control and communications circuitry, may support bidirectional wireless communications between device 10 and external equipment (e.g., a companion device such as a computer, cellular telephone, or other electronic device, an accessory such as a point device, computer stylus, or other input device, speakers or other output devices, etc.) over a wireless link.
- a companion device such as a computer, cellular telephone, or other electronic device
- an accessory such as a point device, computer stylus, or other input device, speakers or other output devices, etc.
- Communications circuitry 20 may transmit and/or receive radio-frequency signals within a corresponding frequency band at radio frequencies (sometimes referred to herein as a communications band or simply as a “band”).
- the frequency bands handled by communications circuitry 20 may include wireless local area network (WLAN) frequency bands (e.g., Wi-Fi® (IEEE 802.11) or other WLAN communications bands) such as a 2.4 GHz WLAN band (e.g., from 2400 to 2480 MHz), a 5 GHz WLAN band (e.g., from 5180 to 5825 MHz), a Wi-Fi® 6E band (e.g., from 5925-7125 MHZ), and/or other Wi-Fi® bands (e.g., from 1875-5160 MHZ), wireless personal area network (WPAN) frequency bands such as the 2.4 GHz Bluetooth® band or other WPAN communications bands, cellular telephone frequency bands (e.g., bands from about 600 MHz to about 5 GHz, 3G bands, 4G LTE bands, 5G New Radio Frequency
- Device 10 may, if desired, include power circuits for transmitting and/or receiving wired and/or wireless power and may include batteries or other energy storage devices.
- device 10 may include a wireless power coil and a rectifier to receive wireless power that is provided to other circuitry in device 10 .
- Device 10 may include input-output devices such as input-output devices 22 .
- Electronic components such as input-output devices 22 may be used in gathering user input, in gathering information on the environment surrounding the user, and/or in providing a user with output.
- Input-output devices 22 may include one or more displays such as display 24 .
- Display 24 may include one or more display devices such as organic light-emitting diode display panels (panels with organic light-emitting diode pixels formed on polymer substrates or silicon substrates that contain pixel control circuitry), liquid crystal display panels, microelectromechanical systems displays (e.g., two-dimensional mirror arrays or scanning mirror display devices), display panels having pixel arrays formed from crystalline semiconductor light-emitting diode dies (sometimes referred to as microLEDs), and/or other display devices.
- display devices such as organic light-emitting diode display panels (panels with organic light-emitting diode pixels formed on polymer substrates or silicon substrates that contain pixel control circuitry), liquid crystal display panels, microelectromechanical systems displays (e.g., two-dimensional mirror arrays or scanning mirror display devices), display panels having pixel arrays formed from crystalline semiconductor light-emitting diode dies (sometimes referred to as microLEDs), and/or other display devices.
- display devices such as organic light-emit
- Input-output devices 22 may also include sensors 26 .
- Sensors 26 in input-output devices 22 may include image sensors (e.g., visible light cameras, infrared cameras, cameras that are sensitive at multiple wavelengths, three-dimensional camera systems such as depth sensors, structured light sensors and/or depth sensors based on stereo imaging devices that capture three-dimensional images, time-of-flight cameras, etc.), force sensors (e.g., strain gauges, capacitive force sensors, resistive force sensors, etc.), audio sensors such as microphones, touch and/or proximity sensors such as capacitive sensors such as a touch sensor that forms a button, trackpad, or other input device), and other sensors.
- image sensors e.g., visible light cameras, infrared cameras, cameras that are sensitive at multiple wavelengths, three-dimensional camera systems such as depth sensors, structured light sensors and/or depth sensors based on stereo imaging devices that capture three-dimensional images, time-of-flight cameras, etc.
- force sensors e.g., strain gauges, capacitive force
- sensors 26 may include optical sensors such as optical sensors that emit and detect light, ultrasonic sensors, optical touch sensors, optical proximity sensors, and/or other touch sensors and/or proximity sensors, monochromatic and color ambient light sensors, fingerprint sensors, iris scanning sensors, retinal scanning sensors, and other biometric sensors, temperature sensors, sensors for measuring three-dimensional non-contact gestures (“air gestures”), pressure sensors, sensors for detecting position, orientation, and/or motion (e.g., accelerometers, magnetic sensors such as compass sensors, gyroscopes, and/or inertial measurement units that contain some or all of these sensors), health sensors such as blood oxygen sensors, heart rate sensors, blood flow sensors, and/or other health sensors, radio-frequency sensors, optical sensors such as self-mixing sensors and light detection and ranging (lidar) sensors, humidity sensors, moisture sensors, gaze tracking sensors, electromyography sensors to sense muscle activation, facial sensors, interferometric sensors, time-of-flight sensors, magnetic sensors, resistive sensors, distance sensors, angle sensors, and/or other sensors, fingerprint
- device 10 may use sensors 26 and/or other input-output devices 22 to gather user input.
- input-output devices 22 such as buttons may be used to gather button press input
- touch sensors overlapping displays can be used for gathering user touch screen input
- touch pads may be used in gathering touch input
- microphones may be used for gathering audio input (e.g., voice commands)
- accelerometers may be used in monitoring when a finger contacts an input surface and may therefore be used to gather finger press input, etc.
- Input-output devices 22 may include optical components such as depth sensors (e.g., structured light sensors or other sensors that gather three-dimensional image data), optical proximity sensors, ambient light sensors (e.g., color ambient light sensors), optical time-of-flight sensors and other sensors 16 that are sensitive to visible and/or infrared light and that may emit visible and/or infrared light (e.g., devices 22 may contain optical sensors that emit and/or detect light).
- a visible light image sensor in a camera may have a visible light flash or an associated infrared flood illuminator to provide illumination while the image sensor captures a two-dimensional and/or three-dimensional image.
- An infrared camera such as an infrared structured light camera that captures three-dimensional infrared images may have an infrared flood illuminator that emits infrared flood illumination and/or may have a dot projector the emits an array of infrared light beams.
- Infrared proximity sensors may emit infrared light and detect the infrared light after the infrared light has reflected from a target object.
- electronic device 10 may include additional components (see, e.g., other devices 28 in input-output devices 22 ).
- the additional components may include haptic output devices, actuators for moving movable structures in device 10 , audio output devices such as speakers, light-emitting diodes for status indicators, light sources such as light-emitting diodes that illuminate portions of a housing and/or display structure, other optical output devices, and/or other circuitry for gathering input and/or providing output.
- Device 10 may also include a battery or other energy storage device, connector ports for supporting wired communication with ancillary equipment and for receiving wired power, and other circuitry.
- FIG. 2 is a front perspective view of electronic device 10 of the type shown in FIG. 1 .
- Electronic device 10 may be a computing device such as a laptop computer, a computer monitor containing an embedded computer, a tablet computer, a cellular telephone, a media player, or other handheld or portable electronic device, a smaller device such as a wristwatch device (e.g., a watch with a wrist strap), a pendant device, a headphone or earpiece device, a device embedded in eyeglasses or other equipment worn on a user's head, or other wearable or miniature device, a television, a computer display that does not contain an embedded computer, a gaming device, a navigation device, an embedded system such as a system in which electronic equipment with a display is mounted in a kiosk or automobile, equipment that implements the functionality of two or more of these devices, or other electronic equipment.
- a computing device such as a laptop computer, a computer monitor containing an embedded computer, a tablet computer, a cellular telephone, a media player,
- device 10 is a portable device such as a cellular telephone, media player, tablet computer, wrist device, or other portable computing device. Other configurations may be used for device 10 if desired.
- FIG. 2 is merely illustrative.
- device 10 includes a display such as display 14 mounted in housing 12 .
- display 14 (or a transparent cover layer that covers display 14 ) may be mounted in housing 12 and form at least a portion of a front surface of device 10 .
- device 10 may also have an opposing rear surface formed by housing 12 .
- Housing 12 which may sometimes be referred to as an enclosure or case, may be formed of plastic, glass, ceramics, fiber composites, metal (e.g., stainless steel, aluminum, titanium, gold, etc.), other suitable materials, or a combination of any two or more of these materials.
- Housing 12 may be formed using a unibody configuration in which some or all of housing 12 is machined or molded as a single structure or may be formed using multiple structures (e.g., an internal frame structure, one or more structures that form exterior housing surfaces, etc.).
- Display 14 may be a touch screen display that incorporates a layer of conductive capacitive touch sensor electrodes or other touch sensor components (e.g., resistive touch sensor components, acoustic touch sensor components, force-based touch sensor components, light-based touch sensor components, etc.) or may be a display that is not touch-sensitive. Capacitive touch screen electrodes may be formed from an array of indium tin oxide pads or other transparent conductive structures. Display 14 may include an array of pixels formed from liquid crystal display (LCD) components, an array of electrophoretic pixels, an array of plasma pixels, an array of organic light-emitting diode pixels or other light-emitting diodes, an array of electrowetting pixels, or pixels based on other display technologies.
- LCD liquid crystal display
- a display cover layer for display 14 may planar or curved and may have a rectangular outline, a circular outline, or outlines of other shapes. If desired, openings may be formed in the display cover layer. For example, an opening may be formed in the display cover layer to accommodate a button, a speaker port, a sensor, or other component. Openings may be formed in housing 12 to form communications ports (e.g., an audio jack port, a digital data port, etc.), to form openings for buttons, or to form audio ports (e.g., openings for speakers and/or microphones), as examples.
- communications ports e.g., an audio jack port, a digital data port, etc.
- audio ports e.g., openings for speakers and/or microphones
- FIG. 3 is a rear perspective view of electronic device 10 of the type shown in FIGS. 1 - 2 .
- multiple cameras including a first image sensor 30 - 1 and a second image sensor 30 - 2 can be disposed at the rear side R of device 10 .
- Image sensors 30 - 1 and 30 - 2 are therefore sometimes referred to as rear-facing cameras.
- Image sensors 30 - 1 and 30 - 2 can both be color image sensors (e.g., cameras configured to capture color images).
- Image sensors 30 - 1 and 30 - 2 (and corresponding lenses) can be configured with different fields of view (FoVs).
- image sensor 30 - 1 can have a first FoV equivalent to a first focal length
- image sensor 30 - 2 can have a second FoV equivalent to a second focal length that is less than the first focal length (e.g., the second FoV of camera 30 - 2 can also be greater than the first FoV of camera 30 - 1 ).
- the first camera 30 - 1 can be referred to as a wide (“W”) image sensor
- the second camera 30 - 2 with an even wider field of view (and shorter focal length) can be referred to as an ultrawide (“UW”) image sensor.
- W wide
- UW ultrawide
- Electronic device 10 can have a rectangular shape having a lengthwise dimension that is elongated along a longitudinal axis 32 .
- Image sensors 30 - 1 and 30 - 2 can be disposed at different points along the longitudinal axis 32 .
- device 10 When device 10 is held upright (e.g., when device 10 is oriented such that its longitudinal axis 32 is orthogonal to the ground), device 10 can be referred to as being operated in a “portrait” orientation such that image sensors 30 - 1 and 30 - 2 are positioned on top of or vertically with respect to each other.
- device 10 When device 10 is held sideways (e.g., when device 10 is oriented in a way such that its longitudinal axis 32 is parallel to the ground), device 10 can be referred to as being operated in a “landscape” orientation such that image sensors 30 - 1 and 30 - 2 are positioned laterally or horizontally with respect to each other.
- device 10 may include only one rear-facing camera, two or more rear-facing cameras, three or more rear-facing cameras, four or more rear-facing cameras, four to ten rear-facing cameras, or more than ten rear-facing cameras, each with the same or different fields of view.
- image sensors 30 - 1 and 30 - 2 can alternatively be disposed along different points of a latitudinal axis that is perpendicular to longitudinal axis 32 across the rear side R of device 10 .
- device 10 may also include a plurality of cameras with the same or different fields or view and focal lengths on the front side of the device (see, e.g., FIG. 2 ). Configurations in which device 10 includes at least a first wide image sensor 30 - 1 and a second ultra wide image sensor 30 - 2 configured in the way shown in FIG. 3 are sometimes described herein as an example.
- the multiple cameras on device 10 can be employed to capture stereoscopic content.
- Stereoscopic content can refer to visual media (e.g., videos or a still image) having a sense of depth and dimensionality by presenting two slightly different perspectives of the same scene to the eyes of a user.
- FIG. 4 is a diagram showing hardware and/or software subsystems within device 10 that can be used to capture a stereoscope video stream. As shown in FIG.
- device 10 can include first image sensor 30 - 1 , second image sensor 30 - 2 , image signal processors (ISPs) such as a first image signal processor 50 - 1 and a second image signal processor 50 - 2 , a first processor downstream of and associated with ISP 50 - 1 such as first processor 60 - 1 , a second processor downstream of and associated with ISP 50 - 2 such as second processor 60 - 2 , a codec (coder decoder) block such as codec 70 , memory devices such as memory 40 , and sensors such as motion sensor 42 .
- ISPs image signal processors
- codec coder decoder
- the first image sensor 30 - 1 (camera A) can be a “wide” camera having a first field of view and a first focal length.
- the second image sensor 30 - 2 (camera B) can be an “ultra wide” camera having a second field of view wider than the first field of view and a second focal length that is shorter than the first focal length.
- This example in which image sensors 30 - 1 and 30 - 2 are wide and ultra wide cameras, respectively, is illustrative and is not intended to limit the scope of the present embodiments.
- two or more image sensors 30 having the same or different fields of view can be used to generate stereoscopic content using the techniques described herein.
- Image sensor 30 - 1 may output a first raw (unprocessed) image to image signal processor 50 - 1
- image sensor 30 - 2 may output a second raw (unprocessed) image to image signal processor 50 - 2 .
- Each image signal processor 50 can be configured to perform classic image signal processing functions that only rely on the input of the live camera feed itself.
- each ISP block 50 may be configured to perform automatic exposure (AE), automatic color correction (sometimes referred to as automatic white balancing), tone mapping (e.g., global and/or local tone mapping), gamma correction, shading correction, noise reduction, black level adjustment, demosaicing, image sharpening, high dynamic range (HDR) correction, color space conversion, and/or other image signal processing functions (just to name a few).
- AE automatic exposure
- automatic color correction sometimes referred to as automatic white balancing
- tone mapping e.g., global and/or local tone mapping
- gamma correction e.g., shading correction, noise reduction, black level adjustment, demosaicing, image sharpening, high dynamic range (HDR) correction, color space conversion, and/or other image signal processing functions (just to name a few).
- HDR high dynamic range
- image signal processor 50 - 1 can receive a first raw image from image sensor 30 - 1 , perform one or more ISP functions described above to obtain a first processed image, and use an internal crop/scaling subblock 52 to crop or downscale the first processed image.
- image signal processor 50 - 2 can receive a second raw image from image sensor 30 - 2 , perform one or more ISP functions described above to obtain a second processed image, and use an internal crop/scaling subblock 52 to crop or downscale the second processed image.
- the first processed image can be conveyed to first downstream processor 60 - 1 .
- the second processed image can be conveyed to second downstream processor 60 - 2 .
- the first and second processed images can have different resolutions and different fields of view.
- Processors 60 - 1 and 60 - 2 can each be implemented as a central processing unit (CPU), a graphics processing unit (GPU), a neural processing unit (NPU), a digital signal processor (DSP), a programmable processing unit such as a field programmable gate array (FPGA), and/or other types of processors.
- processors 60 - 1 and 60 - 2 can be configured to perform stereo rectification operations and image stabilization operations on the processed images. As shown in FIG. 4 , processors 60 - 1 and 60 - 2 can each include subblocks 64 and 66 for performing the stereo rectification and stabilization operations. In one example, the stereo rectification operations can be performed prior to the image stabilization operations. In another example, the stereo rectification operations can be performed after the image stabilization operations. In yet another example, the stereo rectification operations can be performed in parallel (simultaneously) with the image stabilization operations.
- Stereo rectification may refer to a process of transforming a pair of stereo images or videos so that corresponding points in the two views lie on the same horizontal scan line.
- the stereo rectification operations or functions performed by subblocks 64 can produce epipolar alignment (e.g., so that epipolar lines are horizontal).
- Stereo rectification (sometimes referred to as “stereo alignment”) may refer to the process of aligning a pair of images or videos so that corresponding points in the two views so that the corresponding points in the two views line on the same horizontal scan line. Stereo alignment can be accomplished by finding the translation and/or rotation between the two views and then using this information to warp both of the images.
- processors 60 - 1 and 60 - 2 can receive stereo rectification calibration data that is stored on memory 40 .
- Memory 40 may be nonvolatile memory, volatile memory, and/or other types of storage that is part of circuitry 16 in FIG. 1 .
- the stereo rectification calibration data can include extrinsic calibration data of image sensor 30 - 1 and/or image sensor 30 - 2 .
- Extrinsic” calibration data may refer to and be defined herein as data associated with the six degrees of freedom (6 DOF) of an image sensor in the 3D space.
- the extrinsic calibration data can include information relating to the rotational orientation of an image sensor (e.g., the pitch, roll, and yaw) and the translational positioning of an image sensor (e.g., the forward/backward displacement, up/down displacement, and left/right displacement).
- an image sensor e.g., the pitch, roll, and yaw
- the translational positioning of an image sensor e.g., the forward/backward displacement, up/down displacement, and left/right displacement
- intrinsic data can refer and be defined herein as data relating to how world coordinates are being projected onto an image sensor (e.g., how 3D data is translated to 2D data).
- intrinsic camera data can include the focal length of an image sensor, an optical center of an image sensor, a skew associated with an image sensor, and/other intrinsic metrics that can potentially vary from frame-to-frame.
- intrinsic data relating to image sensor 30 - 1 and/or 30 - 2 can also be stored in memory 40 .
- memory 40 can store and provide extrinsic calibration data (e.g., stereo rectification calibration data) associated with image sensor 30 - 1 and/or image sensor 30 - 2 and intrinsic calibration sensor data of both image sensors 30 - 1 and 30 - 2 to processors 60 - 1 and 60 - 2 .
- extrinsic calibration data e.g., stereo rectification calibration data
- the stabilization (and pose) information associated with the first image sensor 30 - 1 can be used to stabilize the image output from image sensor 30 - 2 to ensure that stabilization between the two cameras are synchronized.
- stabilization information for the first image sensor 30 - 1 can be applied to the second processed image captured using second image sensor 30 - 2 .
- the stabilization/pose information associated with sensor 30 - 1 can be applied to the output of sensor 30 - 2 .
- the stabilization/pose information from the ultra wide sensor 30 - 2 can alternatively be applied to the output of sensor 30 - 1 .
- each image sensor 30 can rely on its own image stabilization/pose information. Device configurations in which the stabilization/pose information associated with image sensor 30 - 1 is applied to the output of image sensor 30 - 2 are sometimes described herein as an example.
- the pose information that is used for the image stabilization algorithms in subblocks 66 may be obtained using one or more motion sensor 42 .
- Motion sensor(s) 42 may output motion data associated with image sensor 30 - 1 and/or 30 - 2 .
- Motion sensor(s) 42 may be considered part of sensors 26 in FIG. 1 .
- motion sensor(s) 42 can include visual inertial odometry (VIO) sensors for gathering information used to track an orientation and position of device 10 .
- VIO sensors can include inertial measurement units (e.g., gyroscopes, gyrocompasses, accelerometers, magnetometers, and/or other inertial sensors), one or more tracking cameras, and/or other position and motion sensors.
- Motion sensor 42 may directly determine the pose, movement, yaw, pitch, roll, etc. of an image sensor.
- Motion sensor 42 can also be used to determine a current orientation and position of device 10 within the environment. Sensors 42 is thus sometime also referred to as a position sensor. As an example, a first motion sensor 42 can provide motion data associated with first image sensor 30 - 1 to stabilization subblock 66 in processor 60 - 1 , whereas a second motion sensor 42 can provide motion data associated with second image sensor 30 - 2 to stabilization subblock 66 in processor 60 - 2 .
- the motion data output by such types of motion sensors 42 can also be considered “extrinsic” (non-calibration) camera data.
- Processors 60 - 1 and 60 - 2 configured to perform stereo rectification and image stabilization operations in this way are sometimes referred to as stereo rectification (alignment) and stabilization processors.
- Processor 60 - 1 may output a first rectified and stabilized image (labeled Image_out 1 in FIG. 4 ) to a first input of codec block 70 .
- Processor 60 - 2 may output a second rectified and stabilized image (labeled Image_out 2 ) to a second input of codec block 70 .
- the first and second rectified (aligned) and stabilized images generated at the outputs of processors 60 - 1 and 60 - 2 can exhibit the same resolution and the same field of view (e.g., the images provided to the inputs of the codec block 70 can have the same focal length without any camera or lens distortion).
- Codec block 70 can, for example, be configured to implement MV-HEVC (Multiview High Efficiency Video Coding), which is a video compression protocol designed to provide efficient data compression for multi-view video content. Codec block 70 can therefore generate a MV-HEVC video stream based on a series of first and second rectified and stabilized images received from processors 60 - 1 and 60 - 2 .
- MV-HEVC is an extension to the HEVC standard that supports the encoding of multiple views of a captured scene into a single data stream.
- MV-HEVC achieves efficient compression by exploiting similarities between different views of a scene. For example, an interview prediction scheme can allow the motion and texture information from one view to predict the motion and texture of another view. This prediction scheme can reduce redundancy in the data and can improve compression efficiency.
- Codec block 70 is therefore sometimes referred to as a video compression block.
- codec block 70 implements MV-HEVC is illustrative. If desired, codec 70 can be configured to implement other types of multi-view encoding schemes to generate a stereoscopic video stream.
- the stereoscopic video stream can optionally be stored on a cloud in a remote or local database and played back on a device having one or more displays capable of presenting stereoscopic video content. As an example, the stereoscopic video stream can be played back on a head-mounted device having one or more displays for creating a sense of depth and 3D experience to a user.
- the various blocks shown in FIG. 4 such as blocks 50 - 1 , 50 - 2 , 60 - 1 , 60 - 2 , and 70 can be referred to collectively as control circuitry (see, e.g., control circuitry 14 in FIG. 1 ).
- Device 10 e.g., a cellular phone
- a head-mounted device can have different viewing conditions.
- device 10 can be operated in a non-immersive viewing condition
- a head-mounted device can be operated in an immersive viewing condition that tends to be much darker.
- the stereoscopic video stream captured by device 10 can be provided with metadata including a chromatic adaptation matrix that is adapted to the non-immersive viewing condition. Since the viewing condition of the head-mounted device is different than that of device 10 , the head-mounted device can optionally color correct the stereoscopic content captured using device 10 using a modified version of the chromatic adaptation matrix included in the metadata. If the stereoscopic content captured using device 10 is viewed on another device in the non-immersive viewing condition, that device can simply apply the same chromatic adaptation matrix included in the metadata to the stereoscopic content being displayed.
- FIG. 5 is a diagram showing illustrative operations that can be performed using the stereo rectification (alignment) and image stabilization processors 60 - 1 and 60 - 2 .
- a stereo rectification 3D rotation matrix can be computed for image sensor 30 - 1 .
- the stereo rectification 3D rotation matrix may be a fixed matrix that is computed based on extrinsic calibration data associated with image sensor 30 - 1 and/or image sensor 30 - 2 .
- a rectified and stabilized pose can be calculated by (1) rectifying the pose from image sensor 30 - 1 and then (2) stabilizing the rectified pose of image sensor 30 - 1 (e.g., using a temporal filter).
- This rectified and stabilized pose (defined herein as “R T ”) can be computed on a frame-by-frame basis.
- the rectified and stabilized pose can be computed based on the rotation matrix computed from block 80 and based on rotation data associated with image sensor 30 - 1 , defined herein as “R A ” and optionally expressed in world coordinates, as provided from motion sensor(s) 42 .
- the operations of block 80 and 82 can be performed by processor 60 - 1 .
- a first stabilization homography H 1 can be computed for image sensor 30 - 1 .
- a homography can generally refer to a mathematical transformation that maps points in one plane to corresponding points in another plane.
- a homography can be used to describe the relationship between two images of the same scene taken from different angle or perspective.
- a homography can sometimes be expressed as a matrix that is used to project a point in one image onto a corresponding point in another image based on intrinsic and extrinsic parameters of an image sensor.
- the first stabilization homography H 1 can be calculated based on the rectified and stabilized pose R T computed from block 82 , rotation or motion data associated with image sensor 30 - 1 (defined herein as “RA”) as provided by motion sensor(s) 42 , intrinsic calibration sensor data associated with image sensor 30 - 1 (defined herein as “K A ”) as provided from memory 40 , and intrinsic sensor data associated with a target image sensor (defined herein as “K T ” and sometimes referred to as intrinsic target sensor data) as provided from memory 40 .
- the intrinsic calibration sensor data K A can, for example, include the focal length and/or optical center point of image sensor 30 - 1 .
- the intrinsic target sensor data K T can be equal to or different than the intrinsic calibration sensor data K A of sensor 30 - 1 .
- the optical center of the target camera can be dynamically set equal to the optical center of image sensor 30 - 1 or can be statically set to the center of an image.
- the first stabilization homography H 1 can be computed as follows:
- H 1 K A ⁇ R A ( R T ) - 1 ⁇ ( K T ) - 1 ( 1 )
- the first stabilization homography H 1 computed from block 84 - 1 can be used as a warp mesh to perform a first warping function at block 86 - 1 .
- the warping function can, for example, be a bicubic warping function for interpolating between related pixels.
- the warping operation 86 - 1 can take Image_in 1 as an input (e.g., the processed image received from the output of image signal processor 50 - 1 ) and then warp Image_in 1 using homography H 1 to generate a corresponding first rectified and stabilized image Image_out 1 . Configured in this way, the rectification (alignment) and stabilization are applied together in a single operation to warp the image.
- the operations of blocks 84 - 1 and 86 - 1 can be performed by processor 60 - 1 .
- the second stabilization homography H 2 can be calculated based on the rectified and stabilized pose R T computed from block 82 , rotation data associated with image sensor 30 - 2 (defined herein as “R B ”) as provided by motion sensor(s) 42 , intrinsic calibration sensor data associated with image sensor 30 - 2 (defined herein as “K B ”) as provided from memory 40 , and intrinsic sensor data K T associated with the target image sensor as provided from memory 40 .
- the intrinsic calibration sensor data K B can, for example, include the focal length and/or optical center point of image sensor 30 - 2 .
- the second stabilization homography H 2 can be computed as follows:
- H 2 K B ⁇ R B ( R T ) - 1 ⁇ ( K T ) - 1 ( 2 )
- the second stabilization homography H 2 computed from block 84 - 2 can be used as a warp mesh to perform a second warping function at block 86 - 2 .
- the warping function can, for example, be a bicubic warping function for interpolating between related pixels.
- the warping operation 86 - 2 can take Image_in 2 as an input (e.g., the processed image received from the output of image signal processor 50 - 2 ) and then warp Image_in 2 using homography H 2 to generate a corresponding second rectified and stabilized image Image_out 2 . Configured in this way, the rectification (alignment) and stabilization are applied together in a single operation to warp the image.
- the operations of blocks 84 - 2 and 86 - 2 can be performed by processor 60 - 2 .
- the first and second rectified (aligned) and stabilized images Image_out 1 and Image_out 2 generated at the outputs of processors 60 - 1 and 60 - 2 can exhibit the same resolution and the same field of view (e.g., the warped images can have the same focal length without any camera or lens distortion).
- FIGS. 4 - 5 showing hardware and software subsystems configured to generate a stereoscopic video stream are illustrative.
- FIGS. 6 and 7 show other embodiments that include additional circuitry for outputting stereoscopic still images (e.g., an image that captures a single moment in time rather than a sequence of images).
- electronic device 10 can further include additional processing blocks such as a first denoising block 90 - 1 , a second denoising block 90 - 2 , and codec block 71 .
- First denoising block 90 - 1 may have an input coupled to the output of processor 60 - 1 via data path 94 - 1 , can perform a software based multiband noise reduction (as an example) to produce a first still image Still 1 .
- Second denoising block 90 - 2 may have an input coupled to the output of processor 60 - 2 via data path 94 - 2 , can perform a software based multiband noise reduction (as an example) to produce a second still image Still 2 . If the image resolution of Still 1 and Still 2 are insufficient, additional upscaling or super-resolution processing blocks can be included to increase the resolution of Still 1 and Still 2 . Images Still 1 and Still 2 can be fed to inputs of codec block 71 . Denoising blocks 90 - 1 and 90 - 2 in FIG. 6 are optional. In general, blocks 90 - 1 and 90 - 2 can be omitted or can be replaced with other image enhance block(s) configured to perform denoising and upscaling operations.
- Codec block 71 can, for example, be configured to implement HEVC (High Efficiency Video Coding), which is a video compression protocol designed to provide efficient data compression for video content.
- Codec block 71 can, for example, generate a corresponding still stereoscopic pair (or stereoscopic image pair) using the HEIC (High Efficiency Image Format). This is merely illustrative. If desired, codec block 71 can generate a still stereoscopic image pair compressed using the JPEG (Joint Photographics Expert Group) image format, PNG (Portable Network Graphics) image format, GIF (Graphics Interchange Format), TIFF (Tagged Image File Format), and/or other image formats. Codec 71 is therefore sometimes referred to as an image compression block.
- JPEG Joint Photographics Expert Group
- PNG Portable Network Graphics
- GIF Graphics Interchange Format
- TIFF Tagged Image File Format
- codec 71 can be configured to implement other types of encoding schemes to generate a stereoscopic image pair.
- the still stereoscopic pair can optionally be stored on a cloud in a remote or local database and played back on a device having one or more displays capable of presenting stereoscopic video content.
- the stereoscopic video stream can be played back on a head-mounted device having one or more displays for displaying slightly different content to create a sense of depth and 3D experience to a user.
- FIG. 6 shows the example of FIG. 6 in which the stills generation circuitry (e.g., blocks 90 - 1 , 90 - 2 , and codec 71 ) taps into the rectified and stabilized images output from processors 60 - 1 and 60 - 2 is illustrative.
- FIG. 7 shows another example in which the stills generation circuitry is coupled directly to the image signal processors 50 .
- device 10 can further include a first denoising block 90 - 1 , a second denoising block 90 - 2 , a third processor 60 - 3 , and a codec block 71 .
- the first denoising block 90 - 1 may have an input coupled to ISP 50 - 1 via data path 92 - 1 and can perform a software based multiband noise reduction (as an example).
- the second denoising block 90 - 2 may have an input coupled to the output of ISP 50 - 2 via data path 92 - 2 and can perform a software based multiband noise reduction (as an example).
- the images (e.g., still images) output on paths 92 - 1 and 92 - 2 can be generated on demand by ISPs 50 - 1 and 50 - 2 .
- Denoising blocks 90 - 1 and 90 - 2 of FIG. 7 are optional. In general, blocks 90 - 1 and 90 - 2 can be omitted or can be replaced with other image enhancement block(s) configured to provide denoising and upscaling functions.
- Processor 60 - 3 may have inputs coupled to denoising blocks 90 - 1 and 90 - 2 .
- Processors 60 - 3 can be implemented as a central processing unit (CPU), a graphics processing unit (GPU), a neural processing unit (NPU), a digital signal processor (DSP), a programmable processing unit such as a field programmable gate array (FPGA), and/or other types of processors.
- processors 60 - 1 and 60 - 2 which can be configured to perform both stereo rectification and image stabilization functions
- processor 60 - 3 can be configured to perform only stereo rectification functions (see subblock 64 ).
- Processor 60 - 3 need not perform any image stabilization.
- Processor 60 - 3 can perform stereo rectification (e.g., using a rotation matrix associated with image sensor 30 - 1 and/or other extrinsic calibration data) based on the output from denoising blocks 90 - 1 and 90 - 2 to output still images Still 1 and Still 2 . Since the images output from ISPs 50 - 1 and 50 - 2 are relatively high resolution, no additional upscaling or super-resolution processing blocks are needed. If desired, however, upscaling or super-resolution processing blocks can be included to boost the resolution of Still 1 and Still 2 . Images Still 1 and Still 2 can be fed to inputs of codec block 71 .
- Codec block 71 can, for example, be configured to implement HEVC (High Efficiency Video Coding), which is a video compression protocol designed to provide efficient data compression for video content.
- Codec block 71 can, for example, generate a corresponding still stereoscopic pair (or stereoscopic image pair) using the HEIC (High Efficiency Image Format). This is merely illustrative. If desired, codec block 71 can generate a still stereoscopic image pair compressed using the JPEG (Joint Photographics Expert Group) image format, PNG (Portable Network Graphics) image format, GIF (Graphics Interchange Format), TIFF (Tagged Image File Format), and/or other image formats.
- JPEG Joint Photographics Expert Group
- PNG Portable Network Graphics
- GIF Graphics Interchange Format
- TIFF Tagged Image File Format
- codec 71 can be configured to implement other types of encoding schemes to generate a stereoscopic image pair.
- the still stereoscopic pair can optionally be stored on a cloud in a remote or local database and played back on a device having one or more displays capable of presenting stereoscopic video content.
- the stereoscopic video stream can be played back on a head-mounted device having one or more displays for displaying slightly different content to create a sense of depth and 3D experience to a user.
- control circuitry such as blocks 50 - 1 , 50 - 2 , 60 - 1 , 60 - 2 , 60 - 3 , 70 , 71 , 90 - 1 and 90 - 2 can be referred to collectively as control circuitry (see, e.g., control circuitry 14 in FIG. 1 )
- FIG. 8 is a flow chart of illustrative steps for operating an electronic device 10 of the type shown in FIGS. 1 - 7 in accordance with some embodiments.
- device 10 can be operated to start a stereoscopic capture.
- device 10 can be operated to capture a stereoscopic video stream (e.g., using the circuitry and processes described in connection with FIGS. 4 - 5 ) and/or to capture a stereoscopic still image (e.g., using the circuitry and processes described in connection with FIGS. 6 - 7 ).
- device 10 can optionally detect an incorrect stereoscopic capture orientation.
- device 10 might use one or more motion sensors or inertial measurement units to determine whether device 10 is currently held in a portrait orientation or a landscape orientation. Assuming the rear-facing cameras 30 - 1 and 30 - 2 are disposed in the way shown in FIG. 3 , stereoscopic capture should be carried out when device 10 is in the landscape orientation.
- device 10 can issue a notification to guide the user to use the correct capture orientation (e.g., to output a user alert to switch to the landscape orientation, as shown by the operations of block 104 ). If device 10 is already in the correct (landscape) orientation, block 104 can be skipped.
- device 10 can optionally detect whether one of the image sensors used for the stereoscopic capture is currently occluded. For example, device 10 can analyze and compare thumbnail information, integration time, local and/or global brightness information, color information, focus information, and/or other image statistics between the images captured by the two cameras to determine whether one of the cameras is occluded. In response to determining that one of the cameras is currently occluded, device 10 can notify the user that one of the image sensors is occluded (see operations of block 108 ), and the user can be provided with an opportunity to remove that occlusion.
- block 108 can be skipped.
- FIG. 8 in which blocks 106 / 108 are performed after blocks 102 / 104 is merely illustrative. If desired, blocks 106 / 108 can be performed before or in parallel with the operations of blocks 102 / 104 .
- device 10 can optionally detect a low light condition.
- device 10 can analyze the brightness information gathered by one of image sensors 30 - 1 and 30 - 2 , by an ambient light sensor, and/or by other optical sensor to determine an ambient light level of the scene being captured. If the ambient light (lux) level of the scene being captured is below a certain threshold, device 10 can notify the user of the low light condition (see block 112 ). The user can be provided with an opportunity to add additional lighting, to move to an area with better lighting conditions, and/or to otherwise improve the lighting condition to improve the quality of the stereoscopic content being captured. If the ambient light level of the scene being captured is greater than the threshold, block 112 can be skipped.
- blocks 110 / 112 are performed after blocks 106 / 108 is merely illustrative. If desired, blocks 110 / 112 can be performed before or in parallel with the operations of blocks 106 / 108 or blocks 102 / 104 .
- device 10 can optionally detect a near-field (close) object that is within a field of view of one of image sensors 30 - 1 and 30 - 2 .
- device 10 can output an alert to the user to move further away or to remove the nearby object so that the object is no longer in the field of view of one of image sensors 30 - 1 and 30 - 2 (see block 116 ). If there is no near-field object detected, then block 116 can be skipped.
- FIG. 8 in which blocks 114 / 116 are performed after blocks 110 / 112 is merely illustrative. If desired, blocks 114 / 116 can be performed before or in parallel with the operations of blocks 110 / 112 , 106 / 108 , or blocks 102 / 104 .
- device 10 can optionally detect if it is moving excessively or if there is an elevated level of camera jitter. For example, the detected movement or camera jitter can be compared to a threshold level. In response to detecting excessive motion or excessive camera jitter, device 10 can notify the user to stop moving as much or to hold still (see operations of block 120 ).
- FIG. 8 in which blocks 118 / 120 are performed after blocks 114 / 116 is merely illustrative. If desired, blocks 118 / 120 can be performed before or in parallel with the operations of blocks 114 / 116 , 110 / 112 , 106 / 108 , or blocks 102 / 104 .
- the operations of FIG. 8 are merely illustrative.
- the various blocks 102 - 120 can all serve to improve the quality of the stereoscopic content. If desired, other steps for improving the quality of the stereoscopic video or image can be employed. If desired, additional steps for detection a far-field object can be included. If desired, additional steps for detecting camera jitter or shake (and associated alert prompting the user to hold still) can be included.
- one or more of the described operations may be modified, replaced, or omitted. In some embodiments, one or more of the described operations may be performed in parallel. In some embodiments, additional processes may be added or inserted between the described operations. If desired, the order of certain operations may be reversed or altered and/or the timing of the described operations may be adjusted so that they occur at slightly different times. In some embodiments, the described operations may be distributed in a larger system.
- the methods and operations described above in connection with FIGS. 1 - 8 may be performed by the components of device 10 using software, firmware, and/or hardware (e.g., dedicated circuitry or hardware).
- Software code for performing these operations may be stored on non-transitory computer readable storage media (e.g., tangible computer readable storage media) stored on one or more of the components of device 10 (e.g., the storage circuitry within control circuitry 14 of FIG. 1 ).
- the software code may sometimes be referred to as software, data, instructions, program instructions, or code.
- the non-transitory computer readable storage media may include drives, non-volatile memory such as non-volatile random-access memory (NVRAM), removable flash drives or other removable media, other types of random-access memory, etc.
- NVRAM non-volatile random-access memory
- Software stored on the non-transitory computer readable storage media may be executed by processing circuitry on one or more of the components of device 10 (e.g., one or more processors in control circuitry 14 ).
- the processing circuitry may include microprocessors, application processors, digital signal processors, central processing units (CPUs), application-specific integrated circuits with processing circuitry, or other processing circuitry.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Studio Devices (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
Description
Claims (25)
Priority Applications (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US18/645,357 US12489874B2 (en) | 2023-05-31 | 2024-04-24 | Stereoscopic capture using cameras with different fields of view |
| EP24177323.3A EP4472193A1 (en) | 2023-05-31 | 2024-05-22 | Stereoscopic capture using cameras with different fields of view |
| CN202410668228.1A CN119071462A (en) | 2023-05-31 | 2024-05-28 | Stereo capture using cameras with different fields of view |
| KR1020240069276A KR20240172093A (en) | 2023-05-31 | 2024-05-28 | Stereoscopic capture using cameras with different fields of view |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202363505350P | 2023-05-31 | 2023-05-31 | |
| US18/645,357 US12489874B2 (en) | 2023-05-31 | 2024-04-24 | Stereoscopic capture using cameras with different fields of view |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20240406361A1 US20240406361A1 (en) | 2024-12-05 |
| US12489874B2 true US12489874B2 (en) | 2025-12-02 |
Family
ID=91226895
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US18/645,357 Active US12489874B2 (en) | 2023-05-31 | 2024-04-24 | Stereoscopic capture using cameras with different fields of view |
Country Status (4)
| Country | Link |
|---|---|
| US (1) | US12489874B2 (en) |
| EP (1) | EP4472193A1 (en) |
| KR (1) | KR20240172093A (en) |
| CN (1) | CN119071462A (en) |
Citations (16)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020163581A1 (en) * | 2000-07-10 | 2002-11-07 | Ricoh Company, Limited | Imaging apparatus, and method and device for shake correction in imaging apparatus |
| US20110249100A1 (en) * | 2010-04-09 | 2011-10-13 | Sankar Jayaram | Apparatus and Method for Capturing Images |
| US8274552B2 (en) | 2010-12-27 | 2012-09-25 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
| US8896667B2 (en) | 2010-10-25 | 2014-11-25 | Aptina Imaging Corporation | Stereoscopic imaging systems with convergence control for reducing conflicts between accomodation and convergence |
| US8982181B2 (en) | 2006-06-13 | 2015-03-17 | Newbery Revocable Trust Indenture | Digital stereo photographic system |
| US20150215599A1 (en) * | 2012-09-06 | 2015-07-30 | S.I.Sv.El Societa' Italiana Per Lo Sviluppo Dell 'elettronica S.P.A. | Method for generating, transmitting and receiving stereoscopic images and relating devices |
| US9445081B1 (en) | 2015-09-25 | 2016-09-13 | Intel Corporation | Method and system of 3D image capture with dynamic cameras |
| US9762800B2 (en) | 2013-03-26 | 2017-09-12 | Canon Kabushiki Kaisha | Image processing apparatus and method, and image capturing apparatus for predicting motion of camera |
| US20190251394A1 (en) * | 2016-07-29 | 2019-08-15 | Samsung Electronics Co., Ltd. | Method and apparatus for continuously displaying images on basis of similarity of images |
| US10523917B2 (en) | 2017-06-29 | 2019-12-31 | Archermind Technology (Nanjing) Co., Ltd. | Method and apparatus for acquiring three-dimensional image using two cameras |
| US10911737B2 (en) | 2010-12-27 | 2021-02-02 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
| US20210272285A1 (en) * | 2018-07-03 | 2021-09-02 | Covidien Lp | Systems, methods, and computer-readable media for detecting image degradation during surgical procedures |
| CN113596322A (en) | 2021-07-09 | 2021-11-02 | Oppo广东移动通信有限公司 | Image processing method, image processing device, electronic equipment and storage medium |
| US11190689B1 (en) * | 2020-07-29 | 2021-11-30 | Google Llc | Multi-camera video stabilization |
| US20220350225A1 (en) * | 2020-01-31 | 2022-11-03 | Fujifilm Corporation | Lens device, imaging apparatus, optical member, imaging method, and imaging program |
| US20230141515A1 (en) * | 2021-11-08 | 2023-05-11 | Foresight Automotive Ltd. | System and method for stereoscopic image analysis |
-
2024
- 2024-04-24 US US18/645,357 patent/US12489874B2/en active Active
- 2024-05-22 EP EP24177323.3A patent/EP4472193A1/en active Pending
- 2024-05-28 CN CN202410668228.1A patent/CN119071462A/en active Pending
- 2024-05-28 KR KR1020240069276A patent/KR20240172093A/en active Pending
Patent Citations (16)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020163581A1 (en) * | 2000-07-10 | 2002-11-07 | Ricoh Company, Limited | Imaging apparatus, and method and device for shake correction in imaging apparatus |
| US8982181B2 (en) | 2006-06-13 | 2015-03-17 | Newbery Revocable Trust Indenture | Digital stereo photographic system |
| US20110249100A1 (en) * | 2010-04-09 | 2011-10-13 | Sankar Jayaram | Apparatus and Method for Capturing Images |
| US8896667B2 (en) | 2010-10-25 | 2014-11-25 | Aptina Imaging Corporation | Stereoscopic imaging systems with convergence control for reducing conflicts between accomodation and convergence |
| US10911737B2 (en) | 2010-12-27 | 2021-02-02 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
| US8274552B2 (en) | 2010-12-27 | 2012-09-25 | 3Dmedia Corporation | Primary and auxiliary image capture devices for image processing and related methods |
| US20150215599A1 (en) * | 2012-09-06 | 2015-07-30 | S.I.Sv.El Societa' Italiana Per Lo Sviluppo Dell 'elettronica S.P.A. | Method for generating, transmitting and receiving stereoscopic images and relating devices |
| US9762800B2 (en) | 2013-03-26 | 2017-09-12 | Canon Kabushiki Kaisha | Image processing apparatus and method, and image capturing apparatus for predicting motion of camera |
| US9445081B1 (en) | 2015-09-25 | 2016-09-13 | Intel Corporation | Method and system of 3D image capture with dynamic cameras |
| US20190251394A1 (en) * | 2016-07-29 | 2019-08-15 | Samsung Electronics Co., Ltd. | Method and apparatus for continuously displaying images on basis of similarity of images |
| US10523917B2 (en) | 2017-06-29 | 2019-12-31 | Archermind Technology (Nanjing) Co., Ltd. | Method and apparatus for acquiring three-dimensional image using two cameras |
| US20210272285A1 (en) * | 2018-07-03 | 2021-09-02 | Covidien Lp | Systems, methods, and computer-readable media for detecting image degradation during surgical procedures |
| US20220350225A1 (en) * | 2020-01-31 | 2022-11-03 | Fujifilm Corporation | Lens device, imaging apparatus, optical member, imaging method, and imaging program |
| US11190689B1 (en) * | 2020-07-29 | 2021-11-30 | Google Llc | Multi-camera video stabilization |
| CN113596322A (en) | 2021-07-09 | 2021-11-02 | Oppo广东移动通信有限公司 | Image processing method, image processing device, electronic equipment and storage medium |
| US20230141515A1 (en) * | 2021-11-08 | 2023-05-11 | Foresight Automotive Ltd. | System and method for stereoscopic image analysis |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20240172093A (en) | 2024-12-09 |
| CN119071462A (en) | 2024-12-03 |
| EP4472193A1 (en) | 2024-12-04 |
| US20240406361A1 (en) | 2024-12-05 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12477231B2 (en) | Apparatus and methods for image encoding using spatially weighted encoding quality parameters | |
| US11450296B2 (en) | Fade-in user interface display based on finger distance or hand proximity | |
| EP3067746B1 (en) | Photographing method for dual-camera device and dual-camera device | |
| US9860448B2 (en) | Method and electronic device for stabilizing video | |
| AU2020250124A1 (en) | Image processing method and head mounted display device | |
| US9794472B2 (en) | Imaging device and focusing control method | |
| KR101675567B1 (en) | Apparatus and system for acquiring panoramic images, method using it, computer program and computer readable recording medium for acquiring panoramic images | |
| KR102746351B1 (en) | Separable distortion mismatch determination | |
| CN114531539B (en) | Photography methods and electronic equipment | |
| US10367989B2 (en) | Imaging device and focusing control method | |
| US10572764B1 (en) | Adaptive stereo rendering to reduce motion sickness | |
| US11776089B2 (en) | Apparatus and methods for non-uniform downsampling of captured panoramic images | |
| WO2022206595A1 (en) | Image processing method and related device | |
| CN106612392A (en) | Image shooting method and device based on double cameras | |
| CN117135454A (en) | Image processing methods, equipment and storage media | |
| CN106919246A (en) | The display methods and device of a kind of application interface | |
| CN115004685B (en) | Electronic device and method for displaying an image at the electronic device | |
| US12489874B2 (en) | Stereoscopic capture using cameras with different fields of view | |
| WO2020044916A1 (en) | Information processing device, information processing method, and program | |
| US8964058B2 (en) | On-board camera system for monitoring an area around a vehicle | |
| JP2023019059A (en) | Imaging apparatus and imaging system | |
| CN117218490A (en) | Gaze position prediction model pre-training method and super-resolution model pre-training method | |
| KR20170047150A (en) | Method and device for generating an image |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| FEPP | Fee payment procedure |
Free format text: ENTITY STATUS SET TO UNDISCOUNTED (ORIGINAL EVENT CODE: BIG.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| AS | Assignment |
Owner name: APPLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RAPAPORT, GUY;WANG, ZEJING;WETZLER, AARON;AND OTHERS;SIGNING DATES FROM 20240407 TO 20240417;REEL/FRAME:067297/0055 Owner name: APPLE INC., CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNOR'S INTEREST;ASSIGNORS:RAPAPORT, GUY;WANG, ZEJING;WETZLER, AARON;AND OTHERS;SIGNING DATES FROM 20240407 TO 20240417;REEL/FRAME:067297/0055 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: ALLOWED -- NOTICE OF ALLOWANCE NOT YET MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NOTICE OF ALLOWANCE MAILED -- APPLICATION RECEIVED IN OFFICE OF PUBLICATIONS |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT RECEIVED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: PUBLICATIONS -- ISSUE FEE PAYMENT VERIFIED |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |