WO2025098400A1 - Camera parameter calibration method and related device - Google Patents
Camera parameter calibration method and related device Download PDFInfo
- Publication number
- WO2025098400A1 WO2025098400A1 PCT/CN2024/130269 CN2024130269W WO2025098400A1 WO 2025098400 A1 WO2025098400 A1 WO 2025098400A1 CN 2024130269 W CN2024130269 W CN 2024130269W WO 2025098400 A1 WO2025098400 A1 WO 2025098400A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- camera
- cameras
- relationship
- lens barrel
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
Definitions
- the present disclosure relates to the field of extended reality technology, and in particular to a camera parameter calibration method and related equipment.
- Extended Reality refers to the combination of reality and virtuality through computers to create a virtual environment for human-computer interaction.
- XR extended reality
- XR extended reality
- AR augmented reality
- VR virtual reality
- MR mixed reality
- the extended reality system provides a wearable device for the user to realize human-computer interaction
- the wearable device may be a head-mounted wearable device.
- the wearable device may realize the line of sight tracking or pupil distance estimation function by collecting human eye images for calculation.
- the camera for collecting images is usually set outside the lens barrel, which limits the accuracy of pupil distance estimation and gaze tracking algorithms.
- the present disclosure provides a method for calibrating camera parameters, comprising:
- the wearable device further comprises a binocular display module, the lens barrel is arranged on a light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component;
- the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
- a wearable device comprising:
- Two lens barrels are arranged on the light-emitting side of the binocular display module, and at least one of the two lens barrels includes at least two cameras and an optical component arranged inside the lens barrel for collecting human eye images, and the at least two cameras are located between the binocular display module and the optical component.
- a camera parameter calibration device comprising:
- the receiving module is configured to: receive a plurality of images obtained by photographing a calibration reference object by at least two cameras;
- the wearable device further comprises a binocular display module, the lens barrel is arranged on a light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component;
- a first determination module is configured to: determine a projection relationship between pixel points of the plurality of images and the calibration reference object;
- a second determination module is configured to: determine a first parameter set and a second parameter set of the at least two cameras according to the projection relationship;
- the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
- a wearable device comprising:
- a lens barrel wherein a display module, an optical component, and at least two cameras are arranged in the lens barrel, wherein the at least two cameras and the optical component are located at a light-emitting side of the display module, and the at least two cameras are located between the optical component and the display module;
- a processing module is electrically coupled to the at least two cameras and is configured to: obtain a first parameter set and a second parameter set obtained by the method of the first aspect and an image taken by the camera; and use the first parameter set, the second parameter set and the image to solve the position of the target area in the image in space.
- a computer device comprising one or more processors, a memory; and one or more programs, wherein the one or more programs are stored in the memory and executed by the one or more processors, and the program includes instructions for executing the method described in the first aspect.
- a non-volatile computer-readable storage medium containing a computer program is provided.
- the processors execute the method described in the first aspect.
- a computer program product comprising computer program instructions, which, when executed on a computer, cause the computer to execute the method described in the first aspect.
- FIG. 1A shows a schematic diagram of an exemplary system provided by an embodiment of the present disclosure.
- FIG. 1B shows a schematic diagram of an exemplary head-mounted wearable device.
- 1C and 1D are schematic diagrams showing exemplary human eye images.
- FIG. 2A shows a schematic diagram of an exemplary wearable device provided by an embodiment of the present disclosure.
- FIG. 2B shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
- FIG. 2C shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
- FIG. 2D shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
- FIG. 3A shows a schematic flow chart of an exemplary method provided in an embodiment of the present disclosure.
- FIG. 3B shows a flowchart of an exemplary method for determining a projection relationship according to an embodiment of the present disclosure.
- FIG4A shows a schematic diagram of an exemplary image collection scenario according to an embodiment of the present disclosure.
- FIG4B shows a schematic diagram of another exemplary image collection scenario according to an embodiment of the present disclosure.
- FIG. 4C is a schematic diagram showing three selected target images.
- FIG4D is a schematic diagram showing a projection relationship between a pixel point and a calibration reference object according to an embodiment of the present disclosure.
- FIG. 4E shows a schematic diagram of an exemplary first parameter according to an embodiment of the present disclosure.
- FIG. 4F shows a schematic diagram of a calibration principle of a second parameter of an exemplary camera in a different-side lens barrel according to an embodiment of the present disclosure.
- FIG. 4G is a schematic diagram showing an image obtained after the captured image is binarized in an embodiment of the present disclosure.
- FIG5 shows a schematic diagram of an exemplary wearable device provided by an embodiment of the present disclosure.
- FIG. 6 shows a schematic diagram of the hardware structure of an exemplary computer device provided in an embodiment of the present disclosure.
- FIG. 7 shows a schematic diagram of an exemplary device provided by an embodiment of the present disclosure.
- a prompt message is sent to the user to clearly prompt the user that the operation requested to be performed will require obtaining and using the user's personal information.
- the user can autonomously choose whether to provide personal information to software or hardware such as an electronic device, application, server, or storage medium that performs the operation of the technical solution of the present disclosure according to the prompt message.
- the prompt information in response to receiving an active request from the user, may be sent to the user in the form of a pop-up window, in which the prompt information may be presented in text form.
- the pop-up window may also carry a selection control for the user to choose "agree” or “disagree” to provide personal information to the electronic device.
- FIG. 1A shows a schematic diagram of an exemplary extended reality system 100 provided by an embodiment of the present disclosure.
- Extended Reality refers to the combination of reality and virtuality through computers to create a virtual environment for human-computer interaction.
- XR extended reality
- XR extended reality
- AR augmented reality
- VR virtual reality
- MR mixed reality
- the system 100 may include various types of wearable devices, such as a head-mounted wearable device (e.g., VR/AR glasses or a head-mounted display (HMD)) 104, an operating handle 108, etc.
- a camera/camera 110 for taking photos of the operator (user) 102 may also be provided.
- the system 100 may also include an external control device 112 for providing a processing function.
- the control device 112 for example, may be a computer device such as a mobile phone or a computer.
- any of the aforementioned devices when any of the aforementioned devices is used as a control device or a main control device, it may communicate with other devices in the system 100 through wired or wireless communication methods to achieve information interaction.
- the user 102 can use the head-mounted wearable device 104 and the operating handle 108 to interact with the extended reality system 100.
- the system 100 can use the images captured by the camera/camera 110 to recognize the posture and gesture of the user 102, and then complete the interaction with the user 102 based on the recognized posture and gesture.
- the user 130 can also implement gesture input through bare hands, and the head-mounted wearable device 104 can collect the front image in real time through the camera or camera set in front of the head-mounted wearable device 104, and recognize the gesture of the user 130 by recognizing the image.
- the system 100 may also communicate with a server 114 and may obtain data, such as pictures, audio, video, etc., from the server 114 and may output these data through the head-mounted wearable device 104, such as displaying pictures or videos on the display screen of the head-mounted wearable device 104, playing audio and video carried by audio using the speakers of the head-mounted wearable device 104, etc.
- the server 114 may retrieve required data, such as pictures, audio, video, etc., from a database server 116 for storing data.
- a collection unit for collecting information may be provided on the head-mounted wearable device 104.
- the types of the collection unit may be various.
- the acquisition unit may further include an environment acquisition unit and a positioning tracking unit, wherein the environment acquisition unit may be used to acquire environment information around (for example, in front of) the wearable device 104, and the positioning tracking unit may be used to position and track the wearable device 104.
- the environment acquisition unit may include but is not limited to a three-color camera (for example, an RGB camera), a depth camera, a binocular camera, a laser and other photosensitive elements
- the positioning tracking unit may include but is not limited to visual simultaneous positioning and mapping (visual SLAM), an inertial measurement unit (IMU), a global positioning system (GPS), an ultra-wideband wireless communication technology (UWB), a laser and other modules.
- visual SLAM visual simultaneous positioning and mapping
- IMU inertial measurement unit
- GPS global positioning system
- UWB ultra-wideband wireless communication technology
- the head-mounted wearable device 104 may also be provided with a speed sensor, an acceleration sensor, an angular velocity sensor (e.g., a gyroscope), etc., for collecting speed information or acceleration information of the head-mounted wearable device 104.
- the operating handle 108 may also be provided with a speed sensor, an acceleration sensor, an angular velocity sensor (e.g., a gyroscope), etc., for collecting speed information or acceleration information of the operating handle 108.
- the aforementioned collection unit may also be provided on a body part of the interactive user 102 directly by attachment without relying on a hardware device, so as to collect relevant information of the body part, such as speed, acceleration, or angular velocity information, or information collected by other sensors or collection units.
- the head-mounted wearable device 104 may also be provided with a camera or a camera for taking photos of the operator (user) 102 (eg, photos of the hands or feet) and images of the environment.
- the system 100 can identify the posture and gestures of the user 102 through the collected information, and then can perform corresponding interactions based on the identified user posture and gestures.
- FIG. 1B shows a schematic diagram of an exemplary head-mounted wearable device 104 .
- the head-mounted wearable device 104 may include a lens barrel 1042, and a display screen 1044 for displaying an image and an optical component 1046 for processing an optical path may be provided inside the lens barrel 1042.
- the optical component 1046 may further include a plurality of lenses (e.g., lenses 1046A and 1046B), and the combination of the plurality of lenses may project light emitted by the display screen 1044 into the human eye 1022, so that the human eye 1022 may see the image displayed on the display screen 1044.
- FIG. 1B only exemplarily shows a single-sided structure of the head-mounted wearable device 104.
- the head-mounted wearable device 104 may include two lens barrel structures arranged in parallel.
- the head-mounted wearable device 104 may also be provided with a camera 1048 for capturing images of the human eye, and the camera 1048 may be a charge coupled device (CCD) image sensor, a complementary metal oxide semiconductor (CMOS) image sensor, or the like.
- CCD charge coupled device
- CMOS complementary metal oxide semiconductor
- the camera 1048 can be an eye tracking (ET) camera, and the human eye images collected by the camera can be used to realize functions such as pupil distance estimation and eye tracking.
- E eye tracking
- the camera 1048 is usually arranged outside the lens barrel and usually only one camera is arranged corresponding to each lens barrel.
- the common deployment position of the camera 1048 is generally at the outer corner of the eye or the nose wing. Referring to FIG1B , if the camera 1048 is close to the outside of the device, the camera deployment position shown in FIG1B is the outer corner of the eye, and if the camera 1048 is close to the inside of the device, the camera deployment position shown in FIG1B is the nose wing.
- the camera installation method in the related art easily causes the camera 1048 to be installed at a large inclination angle relative to the human eye 1022, resulting in a large angle ⁇ between the direction of the camera 1048 and the normal viewing direction of the human eye 1022, making it difficult for the collected human eye image to reflect the image at the normal viewing angle of the human eye, as shown in Figures 1C and 1D.
- the user may need to wear glasses before using the head-mounted wearable device 104.
- the camera 1048 is arranged outside the lens barrel 1042, the camera 1048 is higher than the lens barrel 1042, which is easy to squeeze the glasses, affecting the wearing comfort of the head-mounted wearable device 104.
- the imaging of the camera 1048 is easily affected by the edge of the glasses. The light refracts through the edge of the glasses, which reduces the image clarity of the camera 1048 and forms many refracted light spots in the image, thereby affecting the accuracy of the subsequent algorithm.
- when there is only one camera corresponding to the lens barrel such a problem will be further aggravated.
- an embodiment of the present disclosure provides a wearable device, in which at least two cameras are arranged inside the lens barrel, which can solve or partially solve the above-mentioned problems to a certain extent.
- FIG. 2A shows a schematic diagram of an exemplary wearable device 200 provided in an embodiment of the present disclosure.
- the wearable device 200 may also include a lens barrel 202, and a display module 204 and an optical component 206 disposed inside the lens barrel 202.
- the optical component 206 may further include a plurality of lenses (e.g., lenses 2062 and 2064), and the combination of the plurality of lenses may project the light emitted by the display module 204 into the human eye 1022, so that the human eye 1022 may view the image displayed by the display module 204.
- the wearable device 200 includes two cameras 208A and 208B, and both cameras 208A and 208B are arranged inside the lens barrel 202. Since the cameras 208A and 208B are placed inside the lens barrel 202, the wearing of the glasses will not be affected, thereby improving the comfort of the wearable device 200. At the same time, as shown in FIG.
- the distance between the cameras 208A and 208B and the human eye 1022 is extended, so that the installation inclination angle of the cameras 208A and 208B relative to the human eye 1022 becomes smaller, and then the angle ⁇ between the orientation of the cameras 208A and 208B and the normal viewing direction of the human eye 1022 is smaller than the angle ⁇ , so that the cameras 208A and 208B have a better observation angle, and the collected human eye image can better reflect the image of the normal viewing angle of the human eye, and the imaging quality is better.
- the glasses will not interfere with the imaging of the cameras 208A and 208B, further improving the imaging quality.
- the improved imaging quality also improves the accuracy of algorithms such as pupil distance estimation or line of sight tracking.
- two cameras 208A and 208B are used, more observation information can be provided (more images can be collected), which can also improve the accuracy of algorithms such as pupil distance estimation or line of sight tracking.
- two cameras 208A and 208B may be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2A ) of the lens barrel 202. In this way, the images captured by the two cameras 208A and 208B may be symmetrical, and the processing efficiency of subsequent algorithms may be further improved.
- both cameras 208A and 208B face the light-emitting side of the lens barrel 202 and the angles between the camera 208A and 208B and the axis of the lens barrel 202 (the central dotted line of FIG2A ) are equal, for example, the angles are both ⁇ .
- the images collected by the two cameras 208A and 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- FIG. 2B shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
- the wearable device 200 may further include two reflective structures 210A and 210B, which may be structures having a reflective surface such as a reflective film or a reflective mirror.
- the reflective structures 210A and 210B may correspond to cameras 208A and 208B, respectively. Both cameras 208A and 208B face the display module 204, and the reflective structure 210A is used to reflect the light from the human eye 1022 into the camera 208A, and the reflective structure 210B is used to reflect the light from the human eye 1022 into the camera 208B.
- the cameras 208A and 208B can still collect images of the human eye.
- the observation angle ⁇ is further reduced, and the cameras 208A and 208B can better perform imaging.
- two cameras 208A and 208B may be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2B ) of the lens barrel 202, and the reflective structures 210A and 210B may also be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2B ) of the lens barrel 202.
- the images collected by the two cameras 208A and 208B may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
- both cameras 208A and 208B face the display module 204 and the angles between the two cameras 208A and 208B and the axis of the lens barrel 202 (the central dotted line of FIG2B ) are equal, for example, the angles are both ⁇ ; at the same time, the angles between the reflection structures 210A and 210B and the axis of the lens barrel 202 (the central dotted line of FIG2B ) are also equal.
- the images collected by the two cameras 208A and 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- FIG. 2A and FIG. 2B only exemplarily show a single-side structure of the wearable device 200 . It can be understood that in order to achieve binocular display, the wearable device 200 may include two lens barrel structures arranged in parallel.
- FIG. 2C shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
- the wearable device 200 may include a binocular display module, which may further include a first display module 204A and a second display module 204B.
- the first display module 204A and the second display module 204B may respectively display an image for observation by the first eye 1022A (e.g., right eye) and an image for observation by the second eye 1022B (e.g., left eye).
- the wearable device 200 may further include two lens barrels disposed on the light-emitting side of the binocular display module, for example, a first lens barrel 202A and a second lens barrel 202B. At least one of the two lens barrels may further include at least two cameras disposed inside the lens barrel for collecting human eye images, thereby providing more observation data and improving the accuracy of subsequent algorithms.
- the first lens barrel 202A can be arranged on the light-emitting side of the first display module 204A, and the first lens barrel 202A can further be provided with a first camera 208A and a second camera 208B for collecting the human eye image of the first eye 1022A.
- the first optical assembly 206A can also be arranged in the first lens barrel 202A, for projecting the image displayed by the first display module 204A into the first eye 1022A through optical principles.
- the second lens barrel 202B can be arranged on the light-emitting side of the second display module 204A, and the third camera 208C and the fourth camera 208D for collecting the human eye image of the second eye 1022B can be further arranged in the second lens barrel 202A.
- the human eye image collected by the second eye 1022B for example, the left eye
- a more accurate calculation result can be obtained when it is used for pupil distance calculation or line of sight tracking in the future.
- the second optical component 206B can also be arranged in the second lens barrel 202B, for projecting the image displayed by the second display module 204B into the second eye 1022B through optical principles.
- the second optical component 206B can further include a third lens 2062B and a fourth lens 2064B, and the parameters of the third lens 2062B and the fourth lens 2064B can be different or the same, and can be designed according to actual needs. It can be understood that the type and number of lenses in the second optical component 206B are variable, and the specific type and number of lenses used can be designed according to actual needs.
- the first camera 208A and the second camera 208B may be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A. In this way, the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B may be symmetrical, and the processing efficiency of the subsequent algorithm may be further improved.
- the third camera 208C and the fourth camera 208D are symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B.
- the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can also be symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- the first camera 208A and the second camera 208B are both oriented toward the light-emitting side of the first display module 204A, and the first camera 208A and the second camera 208B are oriented at the same angle to the axis of the first lens barrel 202A, for example, the angles are both ⁇ .
- the human eye images of the first eye 1022A captured by the first camera 208A and the second camera 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- the third camera 208C and the fourth camera 208D are both facing the light-emitting side of the second display module 204B, and the angles between the third camera 208C and the fourth camera 208D and the axis of the second lens barrel 202B are equal, for example, the angles are both ⁇ .
- the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- FIG. 2D shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
- the first camera 208A and the second camera 208B are both facing the first display module 204A, and the first reflection structure 210A corresponding to the first camera 208A and the second reflection structure 210B corresponding to the second camera 208B are provided in the first lens barrel 202A.
- the first reflection structure 210A and the second reflection structure 210B may be structures with reflection surfaces such as reflection films or reflection mirrors.
- the first reflection structure 210A and the second reflection structure 210B may correspond to the first camera 208A and the second camera 208B respectively.
- the first reflection structure 210A is used to reflect the light from the first eye 1022A to the first camera 208A
- the second reflection structure 210B is used to reflect the light from the first eye 1022A to the second camera 208B.
- the first camera 208A and the second camera 208B can still collect the human eye image of the first eye 1022A.
- the observation angle ⁇ is further reduced, and the first camera 208A and the second camera 208B can better image the first eye 1022A.
- the third camera 208C and the fourth camera 208D are both facing the second display module 204B, and the second lens barrel 202B is provided with a third reflective structure 210C corresponding to the third camera 208C and a fourth reflective structure 210D corresponding to the fourth camera 208D, and the third reflective structure 210C and the fourth reflective structure 210D can be structures with a reflective surface such as a reflective film or a reflective mirror.
- the third reflective structure 210C and the fourth reflective structure 210D can correspond to the third camera 208C and the fourth camera 208D respectively, and the third reflective structure 210C is used to reflect the light from the second eye 1022B to the third camera 208C, and the fourth reflective structure 210D is used to reflect the light from the second eye 1022B to the fourth camera 208D.
- the reflection of the light by the third reflective structure 210C and the fourth reflective structure 210D enables the third camera 208C and the fourth camera 208D to still capture the human eye image of the second eye 1022B.
- the observation angle ⁇ is further reduced, and the third camera 208C and the fourth camera 208D can better image the second eye 1022B.
- the first camera 208A and the second camera 208B may be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A, and the first reflection structure 210A and the second reflection structure 210B may also be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A.
- the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
- the third camera 208C and the fourth camera 208D may be symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B, and the third reflection structure 210C and the fourth reflection structure 210D may also be symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B.
- the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
- the first camera 208A and the second camera 208B are both facing the first display module 204A, and the angles between the first camera 208A and the second camera 208B and the axis of the first lens barrel 202A are equal, for example, the angles are both ⁇ ; at the same time, the angles between the first reflective structure 210A and the second reflective structure 210B and the axis of the first lens barrel 202A are also equal.
- the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- the third camera 208C and the fourth camera 208D are both facing the second display module 204B and the angles between the third camera 208C and the fourth camera 208D and the axis of the second lens barrel 202B are equal, for example, the angles are both ⁇ ; at the same time, the angles between the third reflective structure 210C and the fourth reflective structure 210D and the axis of the second lens barrel 202B are also equal.
- the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
- the above embodiments are all described by setting two cameras in two lens barrels respectively. It can be understood that one of the two lens barrels can be set with two or more cameras, and the other lens barrel can be set with one camera. In this case, the algorithm accuracy can still be improved to a certain extent, which should also fall within the protection scope of the present disclosure.
- the structures shown in Figures 2A and 2B can also be matched in the same wearable device 200.
- the lens barrel corresponding to the first eye has the structure shown in Figure 2A
- the lens barrel corresponding to the second eye has the structure shown in Figure 2B.
- the structures corresponding to the first eye and the second eye can also be swapped.
- the embodiments of the present disclosure can obtain better imaging quality and improve the comfort of the wearable device 200 by setting at least two cameras 208 inside at least one lens barrel of the wearable device 200.
- the two cameras can provide more observation information and are less affected by occlusion, so the accuracy of algorithms such as pupil distance estimation or line of sight tracking can be improved.
- placing the camera inside the lens barrel will affect the camera imaging not only by the camera module itself but also by the optical components in the lens barrel.
- FIGS. 2A and 2B relative to the camera 1048 of FIG. 1B , in the optical path from the cameras 208A and 208B to the human eye 1022, there is an optical component 206 or a portion of the optical component 206 in the lens barrel (depending on the relative position relationship between the cameras 208A and 208B and the lenses in the optical component 206, for example, in addition to the positions shown in FIGS.
- the cameras 208A and 208B may also be arranged between the lens 2062 and the lens 2064), so that when the cameras 208A and 208B image the human eye 1022, the optical component 206 or a portion of the optical component 206 affects or changes the optical path from the cameras 208A and 208B to the human eye 1022, resulting in that the distortion of the images obtained by the imaging of the cameras 208A and 208B may no longer be symmetrical.
- the cameras 208A and 208B do not have a unified projection center, which makes it impossible to use a parameterized camera model to fit the projection process of the cameras in the lens barrel to process the distortion, making it difficult to calibrate the camera parameters.
- the embodiments of the present disclosure also provide a camera parameter calibration method, which can use a non-parametric camera model to implement parameter calibration of the camera in the lens barrel.
- a camera extrinsic calibration algorithm based on non-common viewpoints is proposed, thereby solving the problem of difficult calibration of cameras in different lens barrels.
- FIG. 3A is a schematic flow chart of an exemplary method 300 provided in an embodiment of the present disclosure.
- the method 300 can be applied to any computer device with data processing capabilities, and can be used to calibrate the parameters of the camera 208A, 208B, 208C or 208D in the wearable device 200 shown in Figure 2A, Figure 2B, Figure 2C or Figure 2D. As shown in Figure 3A, the method 300 can further include the following steps.
- step 302 multiple images obtained by photographing a calibration reference object by at least two cameras may be received.
- the multiple images may be images captured by different cameras.
- a first parameter e.g., an intrinsic parameter
- a second parameter e.g., an extrinsic parameter
- multiple images may be captured for intrinsic parameter calibration and multiple images may be captured for extrinsic parameter calibration, respectively.
- FIG4A shows a schematic diagram of an exemplary image collection scenario 400 according to an embodiment of the present disclosure.
- a calibration reference object 402 may be set at a fixed position.
- the calibration reference object 402 may be a calibration plate or a chart, and some black and white checkerboards are set on the calibration reference object 402.
- the corresponding camera parameters may be calculated according to the correspondence between the checkerboards in the image and the checkerboards of the calibration reference object 402.
- the camera to be calibrated e.g., camera 208A
- the calibration reference object 402 can be continuously captured during the movement of the camera, thereby obtaining multiple images captured by the camera in different postures, which are used as data for subsequent calculation of camera parameters. Since the captured images are obtained by the camera performing image capture in multiple postures, the camera parameters calculated subsequently are more robust and can be applied to a wider range.
- the at least two cameras provided in the wearable device are cameras of different types or cameras with different camera parameters, multiple images need to be collected in the above manner for each camera or camera with each parameter, respectively, for calibrating the first parameter (e.g., internal parameter) of the corresponding camera.
- the camera parameters of the at least two cameras provided in the wearable device are consistent, a camera can be selected to implement the aforementioned image acquisition operation, so as to calibrate the internal parameters of the camera, and the calibrated internal parameters can be applied to other cameras in the wearable device.
- FIG4B shows a schematic diagram of another exemplary image collection scenario 410 according to an embodiment of the present disclosure.
- a calibration reference object 402 may also be set at a fixed position.
- the at least two cameras can be installed in the lens barrel of the wearable device 200 and the wearable device 200 can be fully assembled.
- the assembled wearable device 200 or its prototype can be used to perform external parameter image collection, and then the corresponding external parameters (for example, the pose relationship between different cameras) can be calculated based on the collected images.
- the wearable device 200 when capturing images (or capturing images) of the calibration reference object 402, the wearable device 200 can be controlled to move along a certain motion trajectory, and the calibration reference object 402 can be continuously captured by four cameras during the movement of the wearable device 200, so that when the relative posture relationship of the four cameras remains unchanged, multiple images captured by the four cameras in different postures are obtained as image data for subsequent calculation of the second parameter. Since the captured images are obtained by the camera capturing images in multiple postures, the camera parameters calculated subsequently have better robustness and can be applied to a wider range.
- each camera can capture multiple images.
- the images captured by each camera can also be used to calculate its own internal parameters. Therefore, in some embodiments, only the image acquisition scene 410 can be used to capture images, and the first parameter and the second parameter of the camera can be calibrated based on these images.
- the computer device may receive a plurality of images captured in the aforementioned scene 400 and/or scene 410 for subsequent processing.
- the camera can be calibrated based on the multiple images.
- the parameter calibration can include calibrating a first parameter and a second parameter, wherein the first parameter can be an intrinsic parameter of the camera and the second parameter can be an extrinsic parameter of the camera.
- the intrinsic parameters of the camera may be calibrated first.
- the camera imaging is not only affected by the module itself, but also by the lens effect of the lens barrel, so that the imaging distortion is no longer symmetrical about the image principal point, and the camera does not have a unified projection center, which makes it impossible to use a parametric camera model to fit the projection process of the camera in the lens barrel to handle the asymmetric distortion. Therefore, in some embodiments, a non-parametric camera model is provided to calibrate the intrinsic parameters of the camera.
- the projection relationship between the pixels of the multiple images and the calibration reference object may be determined in step 304.
- the projection relationship between the image and the calibration reference object 402 may be established based on the acquired image, thereby establishing a corresponding relationship between the pixel and the projection direction (pixel-ray).
- the following takes the first camera 208A as an example to perform projection relationship calculation.
- step 304 of determining the projection relationship between the pixels of the plurality of images and the calibration reference object may further include the following steps:
- step 3042 a first number of target images are selected from the plurality of images.
- a projection relationship can be first established by selecting a certain number of target images from the multiple first images (images captured by the first camera 208A) in the multiple images, and the remaining first images can be used to supplement the parts for which the projection relationship is not established.
- the first number is not specifically limited as long as it is sufficient to proceed with the subsequent steps.
- the first number can be 3.
- FIG. 4C shows three selected target images 412, 414, and 416. As shown in FIG. 4C, the target images 412, 414, and 416 respectively show images obtained by the first camera 208A taking the calibration reference object 402 in different postures. It can be understood that by calibrating the camera parameters of the images captured in different camera postures, the calculated camera parameters can be made more robust.
- a homography transformation matrix with the calibration reference object 402 can be constructed for each local area of each target image to establish a corresponding relationship (pixel-coordinate) between each pixel point contained in the local area and the coordinate system of the calibration reference object 402.
- the coordinate system of the position of the calibration reference object 402 is known, and the checkerboard on the calibration reference object 402 corresponds to the checkerboard image in the target image.
- the coordinates of the four vertices corresponding to the local area in the camera coordinate system of the target image are also known. Based on these known information, the homography transformation matrix H between the local area and the calibration reference object 402 can be obtained.
- step 3046 the projection relationship between the pixel points included in the local area and the calibration reference object is determined according to the homography transformation matrix.
- FIG4D is a schematic diagram showing a projection relationship between a pixel point and a calibration reference object according to an embodiment of the present disclosure.
- the homography transformation matrix can be used to calculate the projection relationship between each pixel point contained in the local area in the target image and the corresponding point on the calibration reference object.
- step 3048 the first number of target images are converted into a reference coordinate system, and the projection relationship between the pixel points of the multiple images and the calibration reference object is determined based on the area with the established projection relationship corresponding to the local area.
- a method similar to the previous one can be used to continue selecting local areas of new target images in the remaining first image to construct a homography transformation matrix until the projection relationship between each pixel point in the entire image and the calibration reference object is completed.
- some pixels may correspond to multiple projection direction data. Therefore, these data can be averaged to obtain a projection direction as the projection direction of the pixel to obtain better projection direction data.
- only the projection direction after the average processing can be stored in subsequent storage, thereby saving storage space.
- the aforementioned method can be used to establish the projection relationship between the image pixels and the calibration reference object 402 for the second camera 208B, the third camera 208C, the fourth camera 208D and possibly more cameras (depending on the number of cameras set in the wearable device), and will not be repeated here.
- the first parameter set and the second parameter set of the at least two cameras may be determined according to the projection relationship in step 306.
- the first parameter set includes a plurality of first parameters, which are used to characterize the target pixel points in the image captured by the camera to be calibrated and the projection direction corresponding to the target pixel points
- the second parameter set includes at least one set of second parameters, which are used to indicate the posture relationship between the at least two cameras.
- a first parameter set of the at least two cameras may be determined first according to the projection relationship, and then a second parameter set may be determined.
- FIG. 4E shows a schematic diagram of an exemplary first parameter according to an embodiment of the present disclosure.
- Such a combination of pixel points and projection directions can be used as a first parameter (the camera parameter can be considered as an internal parameter of the camera), and the camera parameter is used to realize the conversion from the coordinate system of the image captured by the camera to the camera coordinate system, so that it can be used to calculate the position information of certain features (for example, pupil) in the image captured by the camera in three-dimensional space (camera coordinate system).
- the camera parameter can be considered as an internal parameter of the camera
- the camera parameter is used to realize the conversion from the coordinate system of the image captured by the camera to the camera coordinate system, so that it can be used to calculate the position information of certain features (for example, pupil) in the image captured by the camera in three-dimensional space (camera coordinate system).
- the projection direction corresponding to the pixel point obtained already includes the correction of the image distortion at the position, and no additional distortion correction is required.
- the previously obtained data may be simplified and used as the first parameters of the camera, thereby saving space for storing the first parameters.
- a spline surface may be used to fit the initial set of first parameters obtained by the aforementioned calibration, and then the control points of the fitted spline surface may be optimized using a bundle adjustment (BA) algorithm, and the first parameters corresponding to all the control points of the optimized spline surface may be used as the first parameter set of the at least two cameras.
- BA bundle adjustment
- a complete set of first parameters can be obtained through an interpolation algorithm based on the fitting spline surface and the optimized control points.
- the above method only provides a calculation method for the first parameter of a single camera. For each camera in the wearable device 200, the above method can be used to calculate the corresponding first parameter, which will not be repeated here.
- the second parameter set may be further calibrated.
- the wearable device 200 includes a first camera 208A and a second camera 208B disposed in the first lens barrel 202A for collecting human eye images of the first eye 1022A, and a third camera 208C and a fourth camera 208D disposed in the second lens 202B for collecting human eye images of the second eye 1022B. Since the four cameras are located at different positions in the wearable device 200, in order to determine the relative relationship of the images collected by the four cameras, it is necessary to know the position relationship between the four cameras as the second parameter, thereby completing the calibration of the camera extrinsic parameters.
- the following takes the calibration of four cameras in the wearable device 200 as an example to illustrate how to calculate the second parameter.
- the two cameras arranged in the same lens barrel of the wearable device 200 need to capture images of the same human eye, generally, the two cameras have a common viewpoint (or an intersection of optical paths). Therefore, a binocular positioning algorithm can be used to obtain the posture relationship between the two cameras in the same lens barrel.
- the second parameter may include the pose relationship between the first camera 208A and the second camera 208B and the pose relationship between the third camera 208C and the fourth camera 208D, and the pose relationship between the first camera and the second camera and the pose relationship between the third camera and the fourth camera are calculated according to the projection relationship based on a binocular positioning algorithm.
- step 306 of determining the first parameter set and the second parameter set of the at least two cameras according to the projection relationship may further include: determining the posture relationship between the first camera 208A and the second camera 208B and the posture relationship between the third camera 208C and the fourth camera 208D according to the projection relationship using a binocular positioning algorithm.
- an embodiment of the present disclosure provides a method for calculating the pose relationship between cameras in different lens barrels.
- FIG. 4F shows a schematic diagram of a calibration principle of a second parameter of an exemplary camera in a different-side lens barrel according to an embodiment of the present disclosure.
- FIG4F taking the calculation of the pose relationship between the first camera 208A and the fourth camera 208D as an example, the two cameras have no common viewpoint.
- the pose relationship of the first camera 208A relative to the calibration reference object 402 is T 1,i
- the pose relationship of the fourth camera 208D relative to the calibration reference object 402 is T 4,i
- the pose relationship between the first camera 208A and the fourth camera 208D is T 14 .
- a BA optimization problem can be constructed, and the pose relationship T 14 can be obtained by solving the problem.
- the second parameter may also include a pose relationship between the first camera and the third camera, a pose relationship between the first camera and the fourth camera, a pose relationship between the second camera and the third camera, and a pose relationship between the second camera and the fourth camera.
- step 306 of determining the first parameter set and the second parameter set of the at least two cameras according to the projection relationship may further include: determining the posture relationship between the first camera and the third camera, the posture relationship between the first camera and the fourth camera, the posture relationship between the second camera and the third camera, and the posture relationship between the second camera and the fourth camera according to the multiple images and the projection relationship.
- an optimization function may be constructed first.
- the optimization function may include a first formula for representing the error between a detection point in the first image captured by the first camera 208A and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using a first parameter of the first camera 208A corresponding to the detection point.
- the first formula is expressed as:
- f cam1,1 ⁇ 1(T 1,i ,P cam1 )-d cam1
- ⁇ 1 is the projection function corresponding to the first camera 208A (i.e., the projection relationship between the pixel points obtained previously and the calibration reference object 402)
- T 1,i is the external parameter of the first camera 208A and the calibration reference object 402 corresponding to the i-th first image (which can be calculated based on the spatial coordinates of the first image and the calibration reference object 402)
- P cam1 is the 3D point corresponding to the i-th first image of the first camera 208A on the calibration reference object 402 (i.e., the spatial coordinates of the points corresponding to each pixel point of the first image on the calibration reference object 402)
- d cam1 is all the detection points corresponding to the first camera 208A.
- the optimization function may further include a second formula for expressing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the fourth camera 208A corresponding to the detection point.
- the second formula is expressed as:
- f cam4,2 ⁇ 4(T 4,i ,P cam4 )-d cam4
- ⁇ 4 is the projection function corresponding to the fourth camera 208D (that is, the projection relationship between the pixel points obtained previously and the calibration reference object 402)
- T 4,i is the external parameter of the fourth camera 208D and the calibration reference object 402 corresponding to the i-th fourth image (which can be calculated according to the coordinate information of the fourth image and the calibration reference object 402)
- P cam4 is the 3D point corresponding to the i-th fourth image of the fourth camera 208D on the calibration reference object 402 (that is, the spatial coordinates of the points corresponding to each pixel point of the fourth image on the calibration reference object 402)
- d cam4 is all the detection points corresponding to the fourth camera 208D.
- the optimization function may further include a third formula for representing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the first camera 208A corresponding to the detection point and the posture relationship between the first camera and the fourth camera.
- the third formula is expressed as:
- f cam4,1 ⁇ 4(T 14 *T 1,i ,P cam4 )-d cam4
- T14 is the posture relationship between the first camera 208A and the fourth camera 208D, that is, the external parameters of the first camera and the fourth camera required to be obtained in the embodiment of the present disclosure.
- the optimization function may further include a fourth formula for representing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the fourth camera corresponding to the detection point and the posture relationship between the first camera and the fourth camera.
- the fourth formula is expressed as:
- the detection points d cam1 and d cam4 may be calculated in the following manner.
- FIG. 4G is a schematic diagram showing an image obtained after the captured image is binarized in an embodiment of the present disclosure.
- the pixels on the image are either black or white.
- the vertices of the checkerboard grid can be identified. These identifiable vertices can be used as detection points d.
- the detection points detected on the i-th image can be represented as d i .
- the optimization function may further determine an error function based on the first formula, the second formula, the third formula, and the fourth formula.
- the error function is expressed as:
- f fcam1,1 + fcam4,1 + fcam1,2 + fcam4,2
- an optimization function may be constructed based on the error function to determine the final external parameters of the first camera 208A and the fourth camera 208D.
- the optimization function is expressed as follows:
- ⁇ is the loss function
- f is the error function O i is all the information contained in the i-th image
- I is the number of input images.
- the loss function res( ⁇ , T) is a function of posture.
- the optimal solution obtained is the final posture relationship T 14 .
- the pose relationship between the first camera and the third camera, the pose relationship between the first camera and the fourth camera, the pose relationship between the second camera and the third camera, and the pose relationship between the second camera and the fourth camera include: determining the pose relationship between the first camera and the fourth camera according to the multiple images and the projection relationship; determining the pose relationship between the first camera and the third camera according to the pose relationship between the third camera and the fourth camera and the pose relationship between the first camera and the fourth camera; determining the pose relationship between the second camera and the third camera according to the pose relationship between the first camera and the second camera and the pose relationship between the first camera and the third camera; determining the pose relationship between the second camera and the fourth camera according to the pose relationship between the third camera and the fourth camera and the pose relationship between the second camera and the third camera.
- the posture relationships of other arrangements and combinations can be obtained through data conversion, thereby saving the amount of calculation.
- the above method provides a calculation method that can be used when there is no common viewpoint between the two cameras in the different-side barrels. However, when there is a common viewpoint between the two cameras in the different-side barrels, the binocular positioning algorithm can still be used to calculate the pose relationship.
- the above example only takes the example of setting two cameras in the two lens barrels of the wearable device 200 respectively. It can be understood that the number of cameras in the lens barrel can be less or more according to actual needs. However, in any case, based on the inventive concept of the embodiment provided by the present disclosure, the corresponding camera parameters can still be calculated, which will not be repeated here.
- the camera parameter calibration method provided by the embodiment of the present disclosure may include the steps of image acquisition, internal parameter calibration, external parameter calibration, and optimizing and saving parameters, so as to obtain better camera parameters for subsequent algorithm calculation.
- the camera parameter calibration method provided by the embodiments of the present disclosure provides a feasible calibration scheme for the intrinsic and extrinsic parameters of the camera for scenes where the imaging distortion of the camera in the lens barrel is asymmetric and there is no unified projection center.
- the method of the embodiment of the present disclosure can be performed by a single device, such as a computer or a server.
- the method of the present embodiment can also be applied in a distributed scenario and completed by multiple devices cooperating with each other.
- one of the multiple devices can only perform one or more steps in the method of the embodiment of the present disclosure, and the multiple devices will interact with each other to complete the described method.
- the embodiment of the present disclosure also provides a wearable device.
- FIG. 5 shows a schematic diagram of an exemplary wearable device 500 provided by an embodiment of the present disclosure.
- the wearable device 500 includes a lens barrel 202 , in which a display module 204 , cameras 208A, 208B, and an optical component 206 are disposed.
- the cameras 208A, 208B and the optical component 206 are located on the light emitting side of the display module 204 , and the cameras 208A, 208B are located between the optical component 206 and the display module 204 .
- the shooting direction of the cameras 208A and 208B is toward the optical assembly 206, as shown in FIG2A.
- a reflective structure 210A and 210B is disposed between the display module 204 and the cameras 208A and 208B, the reflective surface of the reflective structure 210A and 210B is toward the light emitting side of the display module 204, and the shooting direction of the cameras 208A and 208B is toward the reflective structure 210A and 210B.
- the wearable device 500 also includes a processing module 502, which is electrically coupled to the camera 208 and the display module 204, respectively, and is configured to: obtain the first parameter set and the second parameter set obtained by the aforementioned method 300 and the images taken by the cameras 208A and 208B; use the first parameter set and the second parameter set and the image to solve the position of the target area in the image in space.
- a processing module 502 which is electrically coupled to the camera 208 and the display module 204, respectively, and is configured to: obtain the first parameter set and the second parameter set obtained by the aforementioned method 300 and the images taken by the cameras 208A and 208B; use the first parameter set and the second parameter set and the image to solve the position of the target area in the image in space.
- the first parameter set includes camera parameters in combination of pixel points and projection directions (the camera parameters can be considered as the internal parameters of the camera), and the camera parameters are used to realize the conversion from the coordinate system of the image captured by the camera to the camera coordinate system, so that it can be used to calculate the position information of certain features (for example, pupil) in the image captured by the camera in the three-dimensional space (camera coordinate system).
- the calculation process of the camera parameters includes the correction of image distortion, when the camera parameters are used for coordinate system conversion calculation, the distortion can be directly corrected without the need for additional distortion correction.
- the second parameter set includes a second parameter representing the position and posture relationship between different cameras, and the image information collected by different cameras can be unified (for example, unified into the same camera coordinate system) according to the second parameter.
- the wearable device provided by the embodiment of the present disclosure sets the camera inside the lens barrel, which is less affected by occlusion, and can improve the accuracy of pupil distance estimation or line of sight tracking algorithm. Furthermore, by setting the camera to at least two, more observation information can be provided, further improving the accuracy of the algorithm.
- the camera parameter calibration method and related equipment provided by the embodiment of the present disclosure provide a feasible camera parameter calibration solution for scenes where the camera imaging distortion in the lens barrel is asymmetric and there is no unified projection center.
- FIG6 shows a schematic diagram of the hardware structure of an exemplary computer device 600 provided in the embodiment of the present disclosure.
- the computer device 600 can be used to implement the head-mounted wearable device 104 of FIG1A, the wearable device 200 of FIG2A to FIG2D, and can also be used to implement the external device 112 of FIG1A, and can also be used to implement the server 114 of FIG1A.
- the computer device 600 can also be used to implement the database server 116 of FIG1A.
- computer device 600 may include: processor 602, memory 604, network module 606, peripheral interface 608 and bus 610.
- processor 602, memory 604, network module 606 and peripheral interface 608 are connected to each other in communication within computer device 600 via bus 610.
- Processor 602 may be a central processing unit (CPU), an image processor, a neural network processor (NPU), a microcontroller (MCU), a programmable logic device, a digital signal processor (DSP), an application specific integrated circuit (ASIC), or one or more integrated circuits.
- processor 602 may be used to perform functions related to the technology described in the present disclosure.
- processor 602 may also include multiple processors integrated into a single logical component. For example, as shown in FIG. 6 , processor 602 may include multiple processors 602a, 602b, and 602c.
- the memory 604 may be configured to store data (e.g., instructions, computer codes, etc.). As shown in Figure 6, the data stored in the memory 604 may include program instructions (e.g., program instructions for implementing the method 300 or 500 of the embodiment of the present disclosure) and data to be processed (e.g., the memory may store configuration files of other modules, etc.). The processor 602 may also access the program instructions and data stored in the memory 604, and execute the program instructions to operate on the data to be processed.
- the memory 604 may include a volatile storage device or a non-volatile storage device.
- the memory 604 may include a random access memory (RAM), a read-only memory (ROM), an optical disk, a magnetic disk, a hard disk, a solid-state drive (SSD), a flash memory, a memory stick, etc.
- RAM random access memory
- ROM read-only memory
- SSD solid-state drive
- flash memory a memory stick, etc.
- the network interface 606 can be configured to provide the computer device 600 with communication with other external devices via a network.
- the network can be any wired or wireless network capable of transmitting and receiving data.
- the network can be a wired network, a local wireless network (e.g., Bluetooth, WiFi, near field communication (NFC)), a cellular network, the Internet, or a combination thereof. It is understood that the type of network is not limited to the above specific examples.
- the peripheral interface 608 can be configured to connect the computer device 600 to one or more peripheral devices to achieve information input and output.
- the peripheral devices can include input devices such as a keyboard, a mouse, a touch pad, a touch screen, a microphone, and various sensors, and output devices such as a display, a speaker, a vibrator, and an indicator light.
- the bus 610 may be configured to transmit information between various components of the computer device 600 (e.g., the processor 602, the memory 604, the network interface 606, and the peripheral interface 608), such as an internal bus (e.g., a processor-memory bus), an external bus (USB port, PCI-E bus), etc.
- an internal bus e.g., a processor-memory bus
- an external bus USB port, PCI-E bus
- the architecture of the above-mentioned computer device 600 only shows the processor 602, the memory 604, the network interface 606, the peripheral interface 608 and the bus 610, in the specific implementation process, the architecture of the computer device 600 may also include other components necessary for normal operation.
- the architecture of the above-mentioned computer device 600 may also only include the components necessary for implementing the embodiments of the present disclosure, and does not necessarily include all the components shown in the figure.
- FIG. 7 shows a schematic diagram of an exemplary device 700 provided by the embodiment of the present disclosure. As shown in Figure 7, the device 700 can be used to implement the method 300, and can further include the following modules.
- the receiving module 702 is configured to: receive multiple images obtained by photographing a calibrated reference object by at least two cameras; the wearable device also includes a binocular display module, the lens barrel is arranged on the light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component; the first determination module 704 is configured to: determine the projection relationship between the pixel points of the multiple images and the calibrated reference object; the second determination module 706 is configured to: determine the first parameter set and the second parameter set of the at least two cameras according to the projection relationship.
- the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
- the at least two cameras include a first camera and a second camera for capturing a first eye image of a human eye and a third camera and a fourth camera for capturing a second eye image of a human eye
- the first camera and the second camera are arranged in a first lens barrel of the wearable device
- the third camera and the fourth camera are arranged in a second lens barrel of the wearable device
- the second parameter includes a posture relationship between the first camera and the second camera and a posture relationship between the third camera and the fourth camera, and the posture relationship between the first camera and the second camera and the posture relationship between the third camera and the fourth camera are calculated based on a binocular positioning algorithm according to the projection relationship.
- the second parameters also include a posture relationship between the first camera and the third camera, a posture relationship between the first camera and the fourth camera, a posture relationship between the second camera and the third camera, and a posture relationship between the second camera and the fourth camera.
- the second determination module 706 is configured to: determine the posture relationship between the first camera and the fourth camera based on the multiple images and the projection relationship; determine the posture relationship between the first camera and the third camera based on the posture relationship between the third camera and the fourth camera and the posture relationship between the first camera and the fourth camera; determine the posture relationship between the second camera and the third camera based on the posture relationship between the first camera and the second camera and the posture relationship between the first camera and the third camera; determine the posture relationship between the second camera and the fourth camera based on the posture relationship between the third camera and the fourth camera and the posture relationship between the second camera and the third camera.
- the first determination module 704 is configured to: select a first number of target images from the multiple images; select at least one local area in each of the target images, and establish a homography transformation matrix between each of the local areas and the calibrated reference object; determine the projection relationship between the pixel points contained in the local area and the calibrated reference object according to the homography transformation matrix; convert the first number of target images to a reference coordinate system, and determine the projection relationship between the pixel points of the multiple images and the calibrated reference object according to the area with the established projection relationship corresponding to the local area.
- the second determination module 706 is configured to: determine an initial first parameter set of the at least two cameras according to the projection relationship, the initial first parameter set including multiple first parameters corresponding one-to-one to multiple pixel points of the image; fit the initial first parameter set using a spline surface to obtain a fitted spline surface; and optimize the control points of the fitted spline surface based on a bundle adjustment algorithm to obtain the first parameter set of the at least two cameras.
- the above device is described by dividing it into various modules according to its functions.
- the functions of each module can be implemented in the same or multiple software and/or hardware.
- the device of the above embodiment is used to implement the corresponding method 300 in any of the above embodiments, and has the beneficial effects of the corresponding method embodiment, which will not be described in detail here.
- the present disclosure also provides a non-transitory computer-readable storage medium, wherein the non-transitory computer-readable storage medium stores computer instructions, and the computer instructions are used to enable the computer to execute method 300 described in any of the above embodiments.
- the computer-readable medium of this embodiment includes permanent and non-permanent, removable and non-removable media, and information storage can be implemented by any method or technology.
- Information can be computer-readable instructions, data structures, modules of programs, or other data.
- Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, read-only compact disk read-only memory (CD-ROM), digital versatile disk (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices or any other non-transmission media that can be used to store information that can be accessed by a computing device.
- PRAM phase change memory
- SRAM static random access memory
- DRAM dynamic random access memory
- RAM random access memory
- ROM read-only memory
- EEPROM electrically erasable programmable
- the computer instructions stored in the storage medium of the above embodiment are used to enable the computer to execute the method 300 described in any of the above embodiments, and have the beneficial effects of the corresponding method embodiments, which will not be repeated here.
- the present disclosure further provides a computer program product, which includes computer program instructions.
- the computer program instructions can be executed by one or more processors of a computer so that the computer and/or the processor execute the method 300.
- the processor that executes the corresponding step can belong to the corresponding execution subject.
- the computer program product of the above embodiment is used to enable the computer and/or the processor to execute the method 300 described in any of the above embodiments, and has the beneficial effects of the corresponding method embodiments, which will not be described in detail here.
- the known power/ground connections to the integrated circuit (IC) chips and other components may or may not be shown in the provided figures.
- the device can be shown in the form of a block diagram to avoid making the embodiments of the present disclosure difficult to understand, and this also takes into account the fact that the details of the implementation of these block diagram devices are highly dependent on the platform on which the embodiments of the present disclosure will be implemented (that is, these details should be fully within the scope of understanding of those skilled in the art).
- DRAM dynamic RAM
Landscapes
- Engineering & Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Studio Devices (AREA)
Abstract
Description
本申请要求2023年11月8日递交的、标题为“相机参数的标定方法及相关设备”、申请号为202311484874.4的中国发明专利申请的优先权,该申请的全部内容通过引用结合在本申请中。This application claims priority to the Chinese invention patent application entitled “Camera parameter calibration method and related equipment” filed on November 8, 2023 and application number 202311484874.4, the entire contents of which are incorporated by reference into this application.
本公开涉及扩展现实技术领域,尤其涉及一种相机参数的标定方法及相关设备。The present disclosure relates to the field of extended reality technology, and in particular to a camera parameter calibration method and related equipment.
扩展现实(Extended Reality,简称XR),是指通过计算机将真实与虚拟相结合,打造一个可人机交互的虚拟环境。XR(扩展现实)技术可以进一步包含增强现实(AR)、虚拟现实(VR)、混合现实(MR),利用硬件设备结合多种技术手段,将虚拟的内容和真实场景融合。Extended Reality (XR) refers to the combination of reality and virtuality through computers to create a virtual environment for human-computer interaction. XR (extended reality) technology can further include augmented reality (AR), virtual reality (VR), and mixed reality (MR), using hardware devices combined with a variety of technical means to integrate virtual content and real scenes.
一般地,扩展现实系统为用户提供可穿戴设备来实现人机交互,该可穿戴设备可以是头戴式可穿戴设备。在一些场景下,可穿戴设备可以通过采集人眼图像来进行计算从而实现视线追踪或瞳距估计功能。Generally, the extended reality system provides a wearable device for the user to realize human-computer interaction, and the wearable device may be a head-mounted wearable device. In some scenarios, the wearable device may realize the line of sight tracking or pupil distance estimation function by collecting human eye images for calculation.
但是,在相关技术中,通常将采集图像的相机设置在镜筒之外,限制了瞳距估计和视线追踪算法的精度。However, in the related art, the camera for collecting images is usually set outside the lens barrel, which limits the accuracy of pupil distance estimation and gaze tracking algorithms.
发明内容Summary of the invention
本公开提出一种相机参数的标定方法及相关设备,以解决或部分解决上述问题。The present disclosure proposes a camera parameter calibration method and related devices to solve or partially solve the above problems.
本公开第一方面,提供了一种相机参数的标定方法,包括:In a first aspect, the present disclosure provides a method for calibrating camera parameters, comprising:
接收由至少两个相机拍摄标定参考对象而得到的多张图像,所述至少两个相机设置在可穿戴设备的镜筒内;所述可穿戴设备还包括双目显示模组,所述镜筒设置在所述双目显示模组的出光侧,所述镜筒内设置光学组件,所述至少两个相机位于所述双目显示模组和所述光学组件之间;Receiving a plurality of images obtained by photographing a calibration reference object by at least two cameras, wherein the at least two cameras are arranged in a lens barrel of a wearable device; the wearable device further comprises a binocular display module, the lens barrel is arranged on a light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component;
确定所述多张图像的像素点与所述标定参考对象的投射关系;Determining a projection relationship between pixel points of the plurality of images and the calibration reference object;
根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集;Determining a first parameter set and a second parameter set of the at least two cameras according to the projection relationship;
其中,所述第一参数集包括多个第一参数,所述第一参数用于表征所述待标定相机所拍摄的图像中的目标像素点和所述目标像素点对应的投射方向,所述第二参数集包括至少一组第二参数,所述至少一组第二参数用于指示所述至少两个相机之间的位姿关系。Among them, the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
本公开第二方面,提供了一种可穿戴设备,包括:In a second aspect of the present disclosure, a wearable device is provided, comprising:
双目显示模组;Binocular display module;
设置在所述双目显示模组的出光侧的两个镜筒,所述两个镜筒中的至少其一包括设置在所述镜筒内部的用于采集人眼图像的至少两个相机以及光学组件,所述至少两个相机位于所述双目显示模组和所述光学组件之间。Two lens barrels are arranged on the light-emitting side of the binocular display module, and at least one of the two lens barrels includes at least two cameras and an optical component arranged inside the lens barrel for collecting human eye images, and the at least two cameras are located between the binocular display module and the optical component.
本公开第三方面,提供了一种相机参数的标定装置,包括:In a third aspect of the present disclosure, a camera parameter calibration device is provided, comprising:
接收模块,被配置为:接收由至少两个相机拍摄标定参考对象而得到的多张图像;所述可穿戴设备还包括双目显示模组,所述镜筒设置在所述双目显示模组的出光侧,所述镜筒内设置光学组件,所述至少两个相机位于所述双目显示模组和所述光学组件之间;The receiving module is configured to: receive a plurality of images obtained by photographing a calibration reference object by at least two cameras; the wearable device further comprises a binocular display module, the lens barrel is arranged on a light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component;
第一确定模块,被配置为:确定所述多张图像的像素点与所述标定参考对象的投射关系;A first determination module is configured to: determine a projection relationship between pixel points of the plurality of images and the calibration reference object;
第二确定模块,被配置为:根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集;A second determination module is configured to: determine a first parameter set and a second parameter set of the at least two cameras according to the projection relationship;
其中,所述第一参数集包括多个第一参数,所述第一参数用于表征所述待标定相机所拍摄的图像中的目标像素点和所述目标像素点对应的投射方向,所述第二参数集包括至少一组第二参数,所述至少一组第二参数用于指示所述至少两个相机之间的位姿关系。Among them, the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
本公开第四方面,提供了一种可穿戴设备,包括:In a fourth aspect of the present disclosure, a wearable device is provided, comprising:
镜筒,所述镜筒内设置显示模组、光学组件、至少两个相机,所述至少两个相机和所述光学组件位于所述显示模组的出光侧,所述至少两个相机位于所述光学组件和所述显示模组之间;A lens barrel, wherein a display module, an optical component, and at least two cameras are arranged in the lens barrel, wherein the at least two cameras and the optical component are located at a light-emitting side of the display module, and the at least two cameras are located between the optical component and the display module;
处理模块,与所述至少两个相机电耦接,且被配置为:获取采用第一方面的方法得到的第一参数集和第二参数集以及所述相机所拍摄的图像;利用所述第一参数集和所述第二参数集以及所述图像,求解所述图像中的目标区域在空间中的位置。A processing module is electrically coupled to the at least two cameras and is configured to: obtain a first parameter set and a second parameter set obtained by the method of the first aspect and an image taken by the camera; and use the first parameter set, the second parameter set and the image to solve the position of the target area in the image in space.
本公开第五方面,提供了一种计算机设备,包括一个或者多个处理器、存储器;和一个或多个程序,其中所述一个或多个程序被存储在所述存储器中,并且被所述一个或多个处理器执行,所述程序包括用于执行根据第一方面所述的方法的指令。In a fifth aspect of the present disclosure, a computer device is provided, comprising one or more processors, a memory; and one or more programs, wherein the one or more programs are stored in the memory and executed by the one or more processors, and the program includes instructions for executing the method described in the first aspect.
本公开第六方面,提供了一种包含计算机程序的非易失性计算机可读存储介质,当所述计算机程序被一个或多个处理器执行时,使得所述处理器执行第一方面所述的方法。In a sixth aspect of the present disclosure, a non-volatile computer-readable storage medium containing a computer program is provided. When the computer program is executed by one or more processors, the processors execute the method described in the first aspect.
本公开第七方面,提供了一种计算机程序产品,包括计算机程序指令,当所述计算机程序指令在计算机上运行时,使得计算机执行第一方面所述的方法。In a seventh aspect of the present disclosure, a computer program product is provided, comprising computer program instructions, which, when executed on a computer, cause the computer to execute the method described in the first aspect.
为了更清楚地说明本公开或相关技术中的技术方案,下面将对实施例或相关技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本公开的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。In order to more clearly illustrate the technical solutions in the present disclosure or related technologies, the drawings required for use in the embodiments or related technical descriptions are briefly introduced below. Obviously, the drawings described below are only embodiments of the present disclosure. For ordinary technicians in this field, other drawings can be obtained based on these drawings without creative work.
图1A示出了本公开实施例所提供的示例性系统的示意图。FIG. 1A shows a schematic diagram of an exemplary system provided by an embodiment of the present disclosure.
图1B示出了一种示例性头戴式可穿戴设备的示意图。FIG. 1B shows a schematic diagram of an exemplary head-mounted wearable device.
图1C和图1D示出了示例性的人眼图像的示意图。1C and 1D are schematic diagrams showing exemplary human eye images.
图2A示出了本公开实施例所提供的一种示例性可穿戴设备的示意图。FIG. 2A shows a schematic diagram of an exemplary wearable device provided by an embodiment of the present disclosure.
图2B示出了本公开实施例所提供的另一种示例性可穿戴设备的示意图。FIG. 2B shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
图2C示出了本公开实施例所提供的又一种示例性可穿戴设备的示意图。FIG. 2C shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
图2D示出了本公开实施例所提供的又一种示例性可穿戴设备的示意图。FIG. 2D shows a schematic diagram of another exemplary wearable device provided by an embodiment of the present disclosure.
图3A示出了本公开实施例所提供的示例性方法的流程示意图。FIG. 3A shows a schematic flow chart of an exemplary method provided in an embodiment of the present disclosure.
图3B示出了根据本公开实施例的确定投射关系的示例性方法的流程示意图。FIG. 3B shows a flowchart of an exemplary method for determining a projection relationship according to an embodiment of the present disclosure.
图4A示出了根据本公开实施例的一种示例性采图场景的示意图。FIG4A shows a schematic diagram of an exemplary image collection scenario according to an embodiment of the present disclosure.
图4B示出了根据本公开实施例的另一种示例性采图场景的示意图。FIG4B shows a schematic diagram of another exemplary image collection scenario according to an embodiment of the present disclosure.
图4C示出了被选出的3张目标图像的示意图。FIG. 4C is a schematic diagram showing three selected target images.
图4D示出了根据本公开实施例的像素点与标定参考对象的投射关系的示意图。FIG4D is a schematic diagram showing a projection relationship between a pixel point and a calibration reference object according to an embodiment of the present disclosure.
图4E示出了根据本公开实施例的示例性第一参数的示意图。FIG. 4E shows a schematic diagram of an exemplary first parameter according to an embodiment of the present disclosure.
图4F示出了根据本公开实施例的示例性的异侧镜筒内相机的第二参数的标定原理图。FIG. 4F shows a schematic diagram of a calibration principle of a second parameter of an exemplary camera in a different-side lens barrel according to an embodiment of the present disclosure.
图4G示出了本公开实施例中采集得到的图像经过二值化处理后得到的图像的示意图。FIG. 4G is a schematic diagram showing an image obtained after the captured image is binarized in an embodiment of the present disclosure.
图5示出了本公开实施例所提供的示例性可穿戴设备的示意图。FIG5 shows a schematic diagram of an exemplary wearable device provided by an embodiment of the present disclosure.
图6示出了本公开实施例所提供的示例性计算机设备的硬件结构示意图。FIG. 6 shows a schematic diagram of the hardware structure of an exemplary computer device provided in an embodiment of the present disclosure.
图7示出了本公开实施例所提供的示例性装置的示意图。FIG. 7 shows a schematic diagram of an exemplary device provided by an embodiment of the present disclosure.
为使本公开的目的、技术方案和优点更加清楚明白,以下结合具体实施例,并参照附图,对本公开进一步详细说明。In order to make the objectives, technical solutions and advantages of the present disclosure more clearly understood, the present disclosure is further described in detail below in combination with specific embodiments and with reference to the accompanying drawings.
需要说明的是,除非另外定义,本公开实施例使用的技术术语或者科学术语应当为本公开所属领域内具有一般技能的人士所理解的通常意义。本公开实施例中使用的“第一”、“第二”以及类似的词语并不表示任何顺序、数量或者重要性,而只是用来区分不同的组成部分。“包括”或者“包含”等类似的词语意指出现该词前面的元件或者物件涵盖出现在该词后面列举的元件或者物件及其等同,而不排除其他元件或者物件。“连接”或者“相连”等类似的词语并非限定于物理的或者机械的连接,而是可以包括电性的连接,不管是直接的还是间接的。“上”、“下”、“左”、“右”等仅用于表示相对位置关系,当被描述对象的绝对位置改变后,则该相对位置关系也可能相应地改变。It should be noted that, unless otherwise defined, the technical terms or scientific terms used in the embodiments of the present disclosure should be understood by people with ordinary skills in the field to which the present disclosure belongs. The "first", "second" and similar words used in the embodiments of the present disclosure do not indicate any order, quantity or importance, but are only used to distinguish different components. "Including" or "comprising" and similar words mean that the elements or objects appearing before the word cover the elements or objects listed after the word and their equivalents, without excluding other elements or objects. "Connect" or "connected" and similar words are not limited to physical or mechanical connections, but can include electrical connections, whether direct or indirect. "Up", "down", "left", "right" and the like are only used to indicate relative positional relationships. When the absolute position of the described object changes, the relative positional relationship may also change accordingly.
可以理解的是,在使用本公开各实施例公开的技术方案之前,均应当依据相关法律法规通过恰当的方式对本公开所涉及个人信息的类型、使用范围、使用场景等告知用户并获得用户的授权。It is understandable that before using the technical solutions disclosed in the embodiments of the present disclosure, the types, scope of use, usage scenarios, etc. of the personal information involved in the present disclosure should be informed to the user and the user's authorization should be obtained in an appropriate manner in accordance with relevant laws and regulations.
例如,在响应于接收到用户的主动请求时,向用户发送提示信息,以明确地提示用户,其请求执行的操作将需要获取和使用到用户的个人信息。从而,使得用户可以根据提示信息来自主地选择是否向执行本公开技术方案的操作的电子设备、应用程序、服务器或存储介质等软件或硬件提供个人信息。For example, in response to receiving an active request from a user, a prompt message is sent to the user to clearly prompt the user that the operation requested to be performed will require obtaining and using the user's personal information. Thus, the user can autonomously choose whether to provide personal information to software or hardware such as an electronic device, application, server, or storage medium that performs the operation of the technical solution of the present disclosure according to the prompt message.
作为一种可选的但非限定性的实现方式,响应于接收到用户的主动请求,向用户发送提示信息的方式例如可以是弹窗的方式,弹窗中可以以文字的方式呈现提示信息。此外,弹窗中还可以承载供用户选择“同意”或者“不同意”向电子设备提供个人信息的选择控件。As an optional but non-limiting implementation, in response to receiving an active request from the user, the prompt information may be sent to the user in the form of a pop-up window, in which the prompt information may be presented in text form. In addition, the pop-up window may also carry a selection control for the user to choose "agree" or "disagree" to provide personal information to the electronic device.
可以理解的是,上述通知和获取用户授权过程仅是示意性的,不对本公开的实现方式构成限定,其它满足相关法律法规的方式也可应用于本公开的实现方式中。It is understandable that the above notification and the process of obtaining user authorization are merely illustrative and do not constitute a limitation on the implementation of the present disclosure. Other methods that meet the relevant laws and regulations may also be applied to the implementation of the present disclosure.
可以理解的是,本技术方案所涉及的数据(包括但不限于数据本身、数据的获取或使用)应当遵循相应法律法规及相关规定的要求。It is understandable that the data involved in this technical solution (including but not limited to the data itself, the acquisition or use of the data) shall comply with the requirements of relevant laws, regulations and relevant provisions.
图1A示出了本公开实施例所提供的示例性扩展现实系统100的示意图。FIG. 1A shows a schematic diagram of an exemplary extended reality system 100 provided by an embodiment of the present disclosure.
扩展现实(Extended Reality,简称XR),是指通过计算机将真实与虚拟相结合,打造一个可人机交互的虚拟环境。XR(扩展现实)技术可以进一步包含增强现实(AR)、虚拟现实(VR)、混合现实(MR),利用硬件设备结合多种技术手段,将虚拟的内容和真实场景融合。Extended Reality (XR) refers to the combination of reality and virtuality through computers to create a virtual environment for human-computer interaction. XR (extended reality) technology can further include augmented reality (AR), virtual reality (VR), and mixed reality (MR), using hardware devices combined with a variety of technical means to integrate virtual content and real scenes.
如图1A所示,该系统100可以包括各种类型的可穿戴设备,例如,头戴式可穿戴设备(例如,VR/AR眼镜或头戴式显示器(HMD))104、操作手柄108等。在一些场景中,还可以设置有用于拍摄操作者(用户)102的照片的相机/摄像头110。在一些实施例中,当前述的设备不具有处理功能时,系统100还可以包括外接的控制设备112,用于提供处理功能。该控制设备112,例如,可以是手机、电脑等计算机设备。在一些实施例中,前述的各设备中的任一设备作为控制设备或主控制设备时,可以与系统100中的其他设备通过有线或无线的通信方式来实现信息交互。As shown in FIG1A , the system 100 may include various types of wearable devices, such as a head-mounted wearable device (e.g., VR/AR glasses or a head-mounted display (HMD)) 104, an operating handle 108, etc. In some scenarios, a camera/camera 110 for taking photos of the operator (user) 102 may also be provided. In some embodiments, when the aforementioned device does not have a processing function, the system 100 may also include an external control device 112 for providing a processing function. The control device 112, for example, may be a computer device such as a mobile phone or a computer. In some embodiments, when any of the aforementioned devices is used as a control device or a main control device, it may communicate with other devices in the system 100 through wired or wireless communication methods to achieve information interaction.
在系统100中,用户102可以利用头戴式可穿戴设备104、操作手柄108来实现与扩展现实系统100的交互。在一些场景中,系统100可以利用相机/摄像头110所采集的画面来识别用户102的姿态、手势等,进而基于识别出的姿态、手势来完成与用户102的交互。在一些实施例中,用户130还可以通过裸手来实现手势输入,头戴式可穿戴设备104可以通过设置在头戴式可穿戴设备104的前方的摄像头或相机等实时采集前方图像,并通过对该图像进行识别来对用户130的手势进行识别。In the system 100, the user 102 can use the head-mounted wearable device 104 and the operating handle 108 to interact with the extended reality system 100. In some scenarios, the system 100 can use the images captured by the camera/camera 110 to recognize the posture and gesture of the user 102, and then complete the interaction with the user 102 based on the recognized posture and gesture. In some embodiments, the user 130 can also implement gesture input through bare hands, and the head-mounted wearable device 104 can collect the front image in real time through the camera or camera set in front of the head-mounted wearable device 104, and recognize the gesture of the user 130 by recognizing the image.
在一些实施例中,如图1A所示,系统100还可以与服务器114进行通信,并可以从服务器114中获取数据,例如,图片、音频、视频等,并可以将这些数据通过头戴式可穿戴设备104进行输出,例如,在头戴式可穿戴设备104的显示屏上显示图片或视频,利用头戴式可穿戴设备104的扬声器播放音频和视频所携带的音频,等等。在一些实施例中,如图1A所示,服务器114可以从用于存储数据的数据库服务器116中调取所需的数据,例如,图片、音频、视频等。In some embodiments, as shown in FIG1A , the system 100 may also communicate with a server 114 and may obtain data, such as pictures, audio, video, etc., from the server 114 and may output these data through the head-mounted wearable device 104, such as displaying pictures or videos on the display screen of the head-mounted wearable device 104, playing audio and video carried by audio using the speakers of the head-mounted wearable device 104, etc. In some embodiments, as shown in FIG1A , the server 114 may retrieve required data, such as pictures, audio, video, etc., from a database server 116 for storing data.
在一些实施例中,头戴式可穿戴设备104上可以设置用于采集信息的采集单元。采集单元的类型可以是各种各样的。In some embodiments, a collection unit for collecting information may be provided on the head-mounted wearable device 104. The types of the collection unit may be various.
在一些实施例中,所述采集单元可以进一步包括环境获取单元和定位追踪单元,所述环境获取单元可以用于采集可穿戴设备104周围(例如,前方)的环境信息,所述定位追踪单元可以用于对可穿戴设备104进行定位追踪。可选地,所述环境获取单元可以包括但不限于三色摄像头(例如RGB摄像头)、深度相机、双目相机、激光等感光元件,所述定位追踪单元可以包括但不限于视觉即时定位与地图构建(视觉SLAM)、惯性测量单元(IMU)、全球定位系统(GPS)、超宽带无线通信技术(UWB)、激光等模组。In some embodiments, the acquisition unit may further include an environment acquisition unit and a positioning tracking unit, wherein the environment acquisition unit may be used to acquire environment information around (for example, in front of) the wearable device 104, and the positioning tracking unit may be used to position and track the wearable device 104. Optionally, the environment acquisition unit may include but is not limited to a three-color camera (for example, an RGB camera), a depth camera, a binocular camera, a laser and other photosensitive elements, and the positioning tracking unit may include but is not limited to visual simultaneous positioning and mapping (visual SLAM), an inertial measurement unit (IMU), a global positioning system (GPS), an ultra-wideband wireless communication technology (UWB), a laser and other modules.
在一些实施例中,头戴式可穿戴设备104还可以设置用于采集头戴式可穿戴设备104的速度信息或加速度信息的速度传感器、加速度传感器、角速度传感器(例如,陀螺仪),等等。又例如,操作手柄108也可以设置用于操作手柄108的速度信息或加速度信息的速度传感器、加速度传感器、角速度传感器(例如,陀螺仪),等等。需要说明的是,前述的采集单元除了可以设置在头戴式可穿戴设备104和操作手柄108上之外,还可以不依托硬件设备而是直接通过贴附的方式设置在交互用户102的身体部位上,从而采集该身体部位的相关信息,例如,速度或加速度或角速度信息,或者其他传感器或采集单元所采集的信息。In some embodiments, the head-mounted wearable device 104 may also be provided with a speed sensor, an acceleration sensor, an angular velocity sensor (e.g., a gyroscope), etc., for collecting speed information or acceleration information of the head-mounted wearable device 104. For another example, the operating handle 108 may also be provided with a speed sensor, an acceleration sensor, an angular velocity sensor (e.g., a gyroscope), etc., for collecting speed information or acceleration information of the operating handle 108. It should be noted that, in addition to being provided on the head-mounted wearable device 104 and the operating handle 108, the aforementioned collection unit may also be provided on a body part of the interactive user 102 directly by attachment without relying on a hardware device, so as to collect relevant information of the body part, such as speed, acceleration, or angular velocity information, or information collected by other sensors or collection units.
在一些实施例中,头戴式可穿戴设备104也可以设置用于拍摄操作者(用户)102的照片(例如,手部或脚部的照片)以及环境图像的相机或者摄像头。In some embodiments, the head-mounted wearable device 104 may also be provided with a camera or a camera for taking photos of the operator (user) 102 (eg, photos of the hands or feet) and images of the environment.
在一些实施例中,系统100通过采集的信息可以对用户102的姿态、手势等进行识别,进而可以根据识别到的用户姿态、手势来进行相应的交互。In some embodiments, the system 100 can identify the posture and gestures of the user 102 through the collected information, and then can perform corresponding interactions based on the identified user posture and gestures.
图1B示出了一种示例性头戴式可穿戴设备104的示意图。FIG. 1B shows a schematic diagram of an exemplary head-mounted wearable device 104 .
如图1B所示,头戴式可穿戴设备104可以包括镜筒1042,该镜筒1042内部可以设置用于显示图像的显示屏1044、用于对光路进行处理的光学组件1046。可选地,该光学组件1046可以进一步包括多个透镜(例如,透镜1046A和1046B),多个透镜的组合可以将显示屏1044发出的光线投射到人眼1022中,使得人眼1022可以观看到显示屏1044显示的画面。可以理解,图1B中仅示例性地示出了头戴式可穿戴设备104的单侧结构,为了实现双目显示,头戴式可穿戴设备104中可以包括并列排布的两个镜筒结构。As shown in FIG. 1B , the head-mounted wearable device 104 may include a lens barrel 1042, and a display screen 1044 for displaying an image and an optical component 1046 for processing an optical path may be provided inside the lens barrel 1042. Optionally, the optical component 1046 may further include a plurality of lenses (e.g., lenses 1046A and 1046B), and the combination of the plurality of lenses may project light emitted by the display screen 1044 into the human eye 1022, so that the human eye 1022 may see the image displayed on the display screen 1044. It is understood that FIG. 1B only exemplarily shows a single-sided structure of the head-mounted wearable device 104. In order to realize binocular display, the head-mounted wearable device 104 may include two lens barrel structures arranged in parallel.
在一些实施例中,如图1B所示,头戴式可穿戴设备104还可以设置用于采集人眼图像的相机1048,该相机1048可以是电荷耦合器件(CCD)图像传感器、互补金属氧化物半导体(CMOS)图像传感器,等等。In some embodiments, as shown in FIG. 1B , the head-mounted wearable device 104 may also be provided with a camera 1048 for capturing images of the human eye, and the camera 1048 may be a charge coupled device (CCD) image sensor, a complementary metal oxide semiconductor (CMOS) image sensor, or the like.
可选地,该相机1048可以是视线追踪(Eye Tracking,简称ET)相机,其所采集的人眼图像可以用于实现瞳距估计和视线追踪等功能。Optionally, the camera 1048 can be an eye tracking (ET) camera, and the human eye images collected by the camera can be used to realize functions such as pupil distance estimation and eye tracking.
如图1B所示,在相关技术中,相机1048通常设置在镜筒外部并且通常每个镜筒仅对应设置一个相机。并且,为了能够更好地采集到完整的人眼图像,且不影响人眼观察显示屏1044的画面,相机1048的常见部署位置一般是在外眼角位置或鼻翼位置。参考图1B所示,若相机1048靠近设备外侧,则图1B示出的相机部署位置是外眼角位置,若相机1048靠近设备内侧,则图1B示出的相机部署位置是鼻翼位置。As shown in FIG1B , in the related art, the camera 1048 is usually arranged outside the lens barrel and usually only one camera is arranged corresponding to each lens barrel. In addition, in order to better capture the complete human eye image and not affect the human eye's observation of the screen 1044, the common deployment position of the camera 1048 is generally at the outer corner of the eye or the nose wing. Referring to FIG1B , if the camera 1048 is close to the outside of the device, the camera deployment position shown in FIG1B is the outer corner of the eye, and if the camera 1048 is close to the inside of the device, the camera deployment position shown in FIG1B is the nose wing.
但是,相关技术中安装相机的方式容易使相机1048相对人眼1022的安装倾角较大,导致相机1048的朝向与人眼1022的正视方向之间的夹角α较大,使得采集到的人眼图像难以反映人眼正视角度的图像,如图1C和图1D所示。However, the camera installation method in the related art easily causes the camera 1048 to be installed at a large inclination angle relative to the human eye 1022, resulting in a large angle α between the direction of the camera 1048 and the normal viewing direction of the human eye 1022, making it difficult for the collected human eye image to reflect the image at the normal viewing angle of the human eye, as shown in Figures 1C and 1D.
在一些情况下,用户可能需要先佩戴眼镜,然后再使用头戴式可穿戴设备104。然而,由于相机1048设置在镜筒1042外,使得相机1048高出镜筒1042,进而容易挤压眼镜,影响头戴式可穿戴设备104的佩戴舒适性。同时,相机1048的成像容易受眼镜边缘的影响,光线经过眼镜边缘产生折射使相机1048的图像清晰度降低同时在图像中形成许多折射光斑,进而影响后续算法的精度。特别是,当镜筒对应的相机仅有一个时,这样的问题会进一步加重。In some cases, the user may need to wear glasses before using the head-mounted wearable device 104. However, since the camera 1048 is arranged outside the lens barrel 1042, the camera 1048 is higher than the lens barrel 1042, which is easy to squeeze the glasses, affecting the wearing comfort of the head-mounted wearable device 104. At the same time, the imaging of the camera 1048 is easily affected by the edge of the glasses. The light refracts through the edge of the glasses, which reduces the image clarity of the camera 1048 and forms many refracted light spots in the image, thereby affecting the accuracy of the subsequent algorithm. In particular, when there is only one camera corresponding to the lens barrel, such a problem will be further aggravated.
有鉴于此,本公开实施例提供了一种可穿戴设备,将在镜筒内部设置至少两个相机,能够在一定程度上解决或部分解决上述问题。In view of this, an embodiment of the present disclosure provides a wearable device, in which at least two cameras are arranged inside the lens barrel, which can solve or partially solve the above-mentioned problems to a certain extent.
图2A示出了本公开实施例所提供的一种示例性可穿戴设备200的示意图。FIG. 2A shows a schematic diagram of an exemplary wearable device 200 provided in an embodiment of the present disclosure.
如图2A所示,类似地,可穿戴设备200也可以包括镜筒202,以及,设置在镜筒202内部的显示模组204、光学组件206。该光学组件206可以进一步包括多个透镜(例如,透镜2062和2064),多个透镜的组合可以将显示模组204发出的光线投射到人眼1022中,使得人眼1022可以观看到显示模组204显示的画面。As shown in FIG2A , similarly, the wearable device 200 may also include a lens barrel 202, and a display module 204 and an optical component 206 disposed inside the lens barrel 202. The optical component 206 may further include a plurality of lenses (e.g., lenses 2062 and 2064), and the combination of the plurality of lenses may project the light emitted by the display module 204 into the human eye 1022, so that the human eye 1022 may view the image displayed by the display module 204.
与图1B所示的可穿戴设备104不同的是,可穿戴设备200包括两个相机208A和208B,并且,两个相机208A和208B均设置在镜筒202内部,由于相机208A和208B被放置在镜筒202内部,也就不会影响眼镜的佩戴,从而可以提高可穿戴设备200的舒适度。同时,如图2A所示,由于相机208A和208B设置在镜筒202内部,延长了相机208A和208B到人眼1022的距离,使得相机208A和208B相对人眼1022的安装倾角变小,进而相机208A和208B的朝向与人眼1022的正视方向之间的夹角β相较于夹角α更小,从而相机208A和208B具有更好的观测角度,采集到的人眼图像更能反映人眼正视角度的图像,成像质量更好。并且,由于相机208A和208B被放置在镜筒202内部,使得眼镜也不会影响对相机208A和208B的成像产生干扰,进一步提高了成像质量。成像质量的提高,使得瞳距估计或视线追踪等算法的精度也得到了提高。此外,由于采用了两个相机208A和208B,可以提供更多的观测信息(采集的图像更多),也能使得瞳距估计或视线追踪等算法的精度得到提高。Unlike the wearable device 104 shown in FIG. 1B , the wearable device 200 includes two cameras 208A and 208B, and both cameras 208A and 208B are arranged inside the lens barrel 202. Since the cameras 208A and 208B are placed inside the lens barrel 202, the wearing of the glasses will not be affected, thereby improving the comfort of the wearable device 200. At the same time, as shown in FIG. 2A , since the cameras 208A and 208B are arranged inside the lens barrel 202, the distance between the cameras 208A and 208B and the human eye 1022 is extended, so that the installation inclination angle of the cameras 208A and 208B relative to the human eye 1022 becomes smaller, and then the angle β between the orientation of the cameras 208A and 208B and the normal viewing direction of the human eye 1022 is smaller than the angle α, so that the cameras 208A and 208B have a better observation angle, and the collected human eye image can better reflect the image of the normal viewing angle of the human eye, and the imaging quality is better. Furthermore, since the cameras 208A and 208B are placed inside the lens barrel 202, the glasses will not interfere with the imaging of the cameras 208A and 208B, further improving the imaging quality. The improved imaging quality also improves the accuracy of algorithms such as pupil distance estimation or line of sight tracking. In addition, since two cameras 208A and 208B are used, more observation information can be provided (more images can be collected), which can also improve the accuracy of algorithms such as pupil distance estimation or line of sight tracking.
在一些实施例中,为了提高后续的瞳距估计或视线追踪等算法的精度,如图2A所示,两个相机208A和208B可以是相对于镜筒202的轴线(图2A的中心虚线)对称设置在镜筒202内的。这样,两个相机208A和208B所采集的图像可以是对称的,后续算法的处理效率可以进一步提高。In some embodiments, in order to improve the accuracy of subsequent algorithms such as pupil distance estimation or line of sight tracking, as shown in FIG2A , two cameras 208A and 208B may be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2A ) of the lens barrel 202. In this way, the images captured by the two cameras 208A and 208B may be symmetrical, and the processing efficiency of subsequent algorithms may be further improved.
在一些实施例中,如图2A所示,两个相机208A和208B均朝向镜筒202的出光侧且相机208A和208B的朝向与镜筒202的轴线(图2A的中心虚线)的夹角相等,例如,夹角都是β。这样,两个相机208A和208B所采集的图像可以是严格对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in FIG2A , both cameras 208A and 208B face the light-emitting side of the lens barrel 202 and the angles between the camera 208A and 208B and the axis of the lens barrel 202 (the central dotted line of FIG2A ) are equal, for example, the angles are both β. In this way, the images collected by the two cameras 208A and 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
图2B示出了本公开实施例所提供的另一种示例性可穿戴设备200的示意图。FIG. 2B shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
如图2B所示,在一些实施例中,可穿戴设备200还可以包括两个反射结构210A和210B,该反射结构210A和210B可以是反射膜或反射镜等具有反射面的结构。反射结构210A和210B可以分别对应相机208A和208B,两个相机208A和208B均朝向显示模组204且反射结构210A用于将来自所述人眼1022的光线反射到相机208A中,反射结构210B用于将来自所述人眼1022的光线反射到相机208B中。通过反射结构210A和210B对光线的反射,使得相机208A和208B仍能采集人眼图像。并且,因为光路中增加了一次反射过程,使得观测角度γ被进一步缩小,相机208A和208B能够更好地进行成像。As shown in FIG. 2B , in some embodiments, the wearable device 200 may further include two reflective structures 210A and 210B, which may be structures having a reflective surface such as a reflective film or a reflective mirror. The reflective structures 210A and 210B may correspond to cameras 208A and 208B, respectively. Both cameras 208A and 208B face the display module 204, and the reflective structure 210A is used to reflect the light from the human eye 1022 into the camera 208A, and the reflective structure 210B is used to reflect the light from the human eye 1022 into the camera 208B. Through the reflection of the light by the reflective structures 210A and 210B, the cameras 208A and 208B can still collect images of the human eye. Moreover, because a reflection process is added to the optical path, the observation angle γ is further reduced, and the cameras 208A and 208B can better perform imaging.
在一些实施例中,为了提高后续的瞳距估计或视线追踪等算法的精度,如图2B所示,两个相机208A和208B可以是相对于镜筒202的轴线(图2B的中心虚线)对称设置在镜筒202内的,反射结构210A和210B也可以是相对于镜筒202的轴线(图2B的中心虚线)对称设置在镜筒202内的。这样,两个相机208A和208B所采集的图像可以是对称的,后续算法的处理效率可以进一步提高。In some embodiments, in order to improve the accuracy of subsequent algorithms such as pupil distance estimation or line of sight tracking, as shown in FIG2B , two cameras 208A and 208B may be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2B ) of the lens barrel 202, and the reflective structures 210A and 210B may also be symmetrically arranged in the lens barrel 202 relative to the axis (the central dotted line of FIG2B ) of the lens barrel 202. In this way, the images collected by the two cameras 208A and 208B may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
在一些实施例中,如图2B所示,两个相机208A和208B均朝向显示模组204且两个相机208A和208B的朝向与镜筒202的轴线(图2B的中心虚线)的夹角相等,例如,夹角都是γ;同时,反射结构210A和210B的朝向与镜筒202的轴线(图2B的中心虚线)的夹角也是相等的。这样,两个相机208A和208B所采集的图像可以是严格对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in FIG2B , both cameras 208A and 208B face the display module 204 and the angles between the two cameras 208A and 208B and the axis of the lens barrel 202 (the central dotted line of FIG2B ) are equal, for example, the angles are both γ; at the same time, the angles between the reflection structures 210A and 210B and the axis of the lens barrel 202 (the central dotted line of FIG2B ) are also equal. In this way, the images collected by the two cameras 208A and 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
前述实施例均以单个镜筒内设置两个相机进行说明,可以理解,当相机的数量进一步增多时,可以进一步增加观测数据,从而进一步提高算法精度,因此,在单个镜筒内设置两个或更多相机的实施例均应属于本公开的保护范围。The aforementioned embodiments are all described by arranging two cameras in a single lens barrel. It can be understood that when the number of cameras is further increased, the observation data can be further increased, thereby further improving the algorithm accuracy. Therefore, embodiments in which two or more cameras are arranged in a single lens barrel should all fall within the protection scope of the present disclosure.
图2A和图2B中仅示例性地示出了可穿戴设备200的单侧结构,可以理解,为了实现双目显示,可穿戴设备200中可以包括并列排布的两个镜筒结构。FIG. 2A and FIG. 2B only exemplarily show a single-side structure of the wearable device 200 . It can be understood that in order to achieve binocular display, the wearable device 200 may include two lens barrel structures arranged in parallel.
图2C示出了本公开实施例所提供的又一种示例性可穿戴设备200的示意图。FIG. 2C shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
如图2C所示,可穿戴设备200可以包括双目显示模组,该双目显示模组可以进一步包括第一显示模组204A和第二显示模组204B。第一显示模组204A和第二显示模组204B可以分别显示供第一眼1022A(例如,右眼)观察的图像和供第二眼1022B(例如,左眼)观察的图像。As shown in Fig. 2C, the wearable device 200 may include a binocular display module, which may further include a first display module 204A and a second display module 204B. The first display module 204A and the second display module 204B may respectively display an image for observation by the first eye 1022A (e.g., right eye) and an image for observation by the second eye 1022B (e.g., left eye).
在一些实施例中,如图2C所示,可穿戴设备200还可以包括设置在所述双目显示模组的出光侧的两个镜筒,例如,第一镜筒202A和第二镜筒202B。所述两个镜筒中的至少其一可以进一步包括设置在所述镜筒内部的用于采集人眼图像的至少两个相机,从而提供更多的观测数据,提升后续算法的精度。In some embodiments, as shown in FIG2C , the wearable device 200 may further include two lens barrels disposed on the light-emitting side of the binocular display module, for example, a first lens barrel 202A and a second lens barrel 202B. At least one of the two lens barrels may further include at least two cameras disposed inside the lens barrel for collecting human eye images, thereby providing more observation data and improving the accuracy of subsequent algorithms.
作为一个可选实施例,如图2C所示,第一镜筒202A可以设置在第一显示模组204A的出光侧,第一镜筒202A内可以进一步设置用于采集第一眼1022A的人眼图像的第一相机208A和第二相机208B。这样,针对于第一眼1022A(例如,右眼)所采集的人眼图像,在后续用于进行瞳距计算或视线追踪时能够得到更精确的计算结果。在一些实施例中,如图2C所示,第一镜筒202A内还可以设置第一光学组件206A,用于将第一显示模组204A显示的图像通过光学原理投射到第一眼1022A中。可选地,第一光学组件206A可以进一步包括第一透镜2062A和第二透镜2064A,第一透镜2062A和第二透镜2064A的参数可以是不同的或相同的,并可以根据实际需求进行设计。可以理解,第一光学组件206A中的透镜类型和数量是可变的,具体所采用的透镜类型和数量可以根据实际需求进行设计。As an optional embodiment, as shown in FIG2C , the first lens barrel 202A can be arranged on the light-emitting side of the first display module 204A, and the first lens barrel 202A can further be provided with a first camera 208A and a second camera 208B for collecting the human eye image of the first eye 1022A. In this way, for the human eye image collected by the first eye 1022A (for example, the right eye), a more accurate calculation result can be obtained when it is subsequently used for pupil distance calculation or line of sight tracking. In some embodiments, as shown in FIG2C , the first optical assembly 206A can also be arranged in the first lens barrel 202A, for projecting the image displayed by the first display module 204A into the first eye 1022A through optical principles. Optionally, the first optical assembly 206A can further include a first lens 2062A and a second lens 2064A, and the parameters of the first lens 2062A and the second lens 2064A can be different or the same, and can be designed according to actual needs. It can be understood that the type and number of lenses in the first optical component 206A are variable, and the specific type and number of lenses used can be designed according to actual needs.
类似地,作为另一个可选实施例,如图2C所示,第二镜筒202B可以设置在第二显示模组204A的出光侧,第二镜筒202A内可以进一步设置用于采集第二眼1022B的人眼图像的第三相机208C和第四相机208D。这样,针对于第二眼1022B(例如,左眼)所采集的人眼图像,在后续用于进行瞳距计算或视线追踪时能够得到更精确的计算结果。类似地,在一些实施例中,如图2C所示,第二镜筒202B内还可以设置第二光学组件206B,用于将第二显示模组204B显示的图像通过光学原理投射到第二眼1022B中。可选地,第二光学组件206B可以进一步包括第三透镜2062B和第四透镜2064B,第三透镜2062B和第四透镜2064B的参数可以是不同的或相同的,并可以根据实际需求进行设计。可以理解,第二光学组件206B中的透镜类型和数量是可变的,具体所采用的透镜类型和数量可以根据实际需求进行设计。Similarly, as another optional embodiment, as shown in FIG2C, the second lens barrel 202B can be arranged on the light-emitting side of the second display module 204A, and the third camera 208C and the fourth camera 208D for collecting the human eye image of the second eye 1022B can be further arranged in the second lens barrel 202A. In this way, for the human eye image collected by the second eye 1022B (for example, the left eye), a more accurate calculation result can be obtained when it is used for pupil distance calculation or line of sight tracking in the future. Similarly, in some embodiments, as shown in FIG2C, the second optical component 206B can also be arranged in the second lens barrel 202B, for projecting the image displayed by the second display module 204B into the second eye 1022B through optical principles. Optionally, the second optical component 206B can further include a third lens 2062B and a fourth lens 2064B, and the parameters of the third lens 2062B and the fourth lens 2064B can be different or the same, and can be designed according to actual needs. It can be understood that the type and number of lenses in the second optical component 206B are variable, and the specific type and number of lenses used can be designed according to actual needs.
在一些实施例中,如图2C所示,第一相机208A和第二相机208B可以相对于第一镜筒202A的轴线对称设置在所述第一镜筒202A内。这样,第一相机208A和第二相机208B所采集的第一眼1022A的人眼图像可以是对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in Fig. 2C, the first camera 208A and the second camera 208B may be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A. In this way, the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B may be symmetrical, and the processing efficiency of the subsequent algorithm may be further improved.
类似地,在一些实施例中,如图2C所示,第三相机208C和第四相机208D相对于第二镜筒202B的轴线对称设置在所述第二镜筒202B内。这样,第三相机208C和第四相机208D所采集的第二眼1022B的人眼图像也可以是对称的,后续算法的处理效率可以进一步提高。Similarly, in some embodiments, as shown in FIG2C , the third camera 208C and the fourth camera 208D are symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B. In this way, the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can also be symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
在一些实施例中,如图2C所示,第一相机208A和第二相机208B均朝向所述第一显示模组204A的出光侧且第一相机208A和第二相机208B的朝向与第一镜筒202A的轴线的夹角相等,例如,夹角都是β。这样,第一相机208A和第二相机208B所采集的第一眼1022A的人眼图像可以是严格对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in FIG2C , the first camera 208A and the second camera 208B are both oriented toward the light-emitting side of the first display module 204A, and the first camera 208A and the second camera 208B are oriented at the same angle to the axis of the first lens barrel 202A, for example, the angles are both β. In this way, the human eye images of the first eye 1022A captured by the first camera 208A and the second camera 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
类似地,在一些实施例中,如图2C所示,第三相机208C和第四相机208D均朝向所述第二显示模组204B的出光侧且第三相机208C和第四相机208D的朝向与第二镜筒202B的轴线的夹角相等,例如,夹角都是β。这样,第三相机208C和第四相机208D所采集的第二眼1022B的人眼图像可以是严格对称的,后续算法的处理效率可以进一步提高。Similarly, in some embodiments, as shown in FIG2C , the third camera 208C and the fourth camera 208D are both facing the light-emitting side of the second display module 204B, and the angles between the third camera 208C and the fourth camera 208D and the axis of the second lens barrel 202B are equal, for example, the angles are both β. In this way, the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
图2D示出了本公开实施例所提供的又一种示例性可穿戴设备200的示意图。FIG. 2D shows a schematic diagram of another exemplary wearable device 200 provided by an embodiment of the present disclosure.
在一些实施例中,如图2D所示,第一相机208A和第二相机208B均朝向所述第一显示模组204A,所述第一镜筒202A内设置与所述第一相机208A对应的第一反射结构210A和与所述第二相机208B对应的第二反射结构210B,第一反射结构210A和第二反射结构210B可以是反射膜或反射镜等具有反射面的结构。第一反射结构210A和第二反射结构210B可以分别对应第一相机208A和第二相机208B,所述第一反射结构210A用于将来自所述第一眼1022A的光线反射到所述第一相机中208A,所述第二反射结构210B用于将来自所述第一眼1022A的光线反射到所述第二相机208B中。通过第一反射结构210A和第二反射结构210B对光线的反射,使得第一相机208A和第二相机208B仍能采集第一眼1022A的人眼图像。并且,因为光路中增加了一次反射过程,使得观测角度γ被进一步缩小,第一相机208A和第二相机208B能够更好地对第一眼1022A进行成像。In some embodiments, as shown in FIG2D , the first camera 208A and the second camera 208B are both facing the first display module 204A, and the first reflection structure 210A corresponding to the first camera 208A and the second reflection structure 210B corresponding to the second camera 208B are provided in the first lens barrel 202A. The first reflection structure 210A and the second reflection structure 210B may be structures with reflection surfaces such as reflection films or reflection mirrors. The first reflection structure 210A and the second reflection structure 210B may correspond to the first camera 208A and the second camera 208B respectively. The first reflection structure 210A is used to reflect the light from the first eye 1022A to the first camera 208A, and the second reflection structure 210B is used to reflect the light from the first eye 1022A to the second camera 208B. Through the reflection of the light by the first reflection structure 210A and the second reflection structure 210B, the first camera 208A and the second camera 208B can still collect the human eye image of the first eye 1022A. Furthermore, since a reflection process is added to the optical path, the observation angle γ is further reduced, and the first camera 208A and the second camera 208B can better image the first eye 1022A.
类似地,在一些实施例中,如图2D所示,第三相机208C和第四相机208D均朝向所述第二显示模组204B,所述第二镜筒202B内设置与所述第三相机208C对应的第三反射结构210C和与所述第四相机208D对应的第四反射结构210D,第三反射结构210C和第四反射结构210D可以是反射膜或反射镜等具有反射面的结构。第三反射结构210C和第四反射结构210D可以分别对应第三相机208C和第四相机208D,所述第三反射结构210C用于将来自所述第二眼1022B的光线反射到所述第三相机208C中,所述第四反射结构210D用于将来自所述第二眼1022B的光线反射到所述第四相机208D中。第三反射结构210C和第四反射结构210D对光线的反射,使得第三相机208C和第四相机208D仍能采集第二眼1022B的人眼图像。并且,因为光路中增加了一次反射过程,使得观测角度γ被进一步缩小,第三相机208C和第四相机208D能够更好地对第二眼1022B进行成像。Similarly, in some embodiments, as shown in FIG. 2D , the third camera 208C and the fourth camera 208D are both facing the second display module 204B, and the second lens barrel 202B is provided with a third reflective structure 210C corresponding to the third camera 208C and a fourth reflective structure 210D corresponding to the fourth camera 208D, and the third reflective structure 210C and the fourth reflective structure 210D can be structures with a reflective surface such as a reflective film or a reflective mirror. The third reflective structure 210C and the fourth reflective structure 210D can correspond to the third camera 208C and the fourth camera 208D respectively, and the third reflective structure 210C is used to reflect the light from the second eye 1022B to the third camera 208C, and the fourth reflective structure 210D is used to reflect the light from the second eye 1022B to the fourth camera 208D. The reflection of the light by the third reflective structure 210C and the fourth reflective structure 210D enables the third camera 208C and the fourth camera 208D to still capture the human eye image of the second eye 1022B. In addition, because a reflection process is added to the optical path, the observation angle γ is further reduced, and the third camera 208C and the fourth camera 208D can better image the second eye 1022B.
在一些实施例中,为了提高后续的瞳距估计或视线追踪等算法的精度,如图2D所示,第一相机208A和第二相机208B可以是相对于第一镜筒202A的轴线对称设置在第一镜筒202A内的,第一反射结构210A和第二反射结构210B也可以是相对于第一镜筒202A的轴线对称设置在第一镜筒202A内的。这样,第一相机208A和第二相机208B所采集的第一眼1022A的人眼图像可以是对称的,后续算法的处理效率可以进一步提高。In some embodiments, in order to improve the accuracy of subsequent algorithms such as pupil distance estimation or line of sight tracking, as shown in FIG2D , the first camera 208A and the second camera 208B may be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A, and the first reflection structure 210A and the second reflection structure 210B may also be symmetrically arranged in the first lens barrel 202A relative to the axis of the first lens barrel 202A. In this way, the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
类似地,在一些实施例中,为了提高后续的瞳距估计或视线追踪等算法的精度,如图2D所示,第三相机208C和第四相机208D可以是相对于第二镜筒202B的轴线对称设置在第二镜筒202B内的,第三反射结构210C和第四反射结构210D也可以是相对于第二镜筒202B的轴线对称设置在第二镜筒202B内的。这样,第三相机208C和第四相机208D所采集的第二眼1022B的人眼图像可以是对称的,后续算法的处理效率可以进一步提高。Similarly, in some embodiments, in order to improve the accuracy of subsequent algorithms such as pupil distance estimation or line of sight tracking, as shown in FIG2D , the third camera 208C and the fourth camera 208D may be symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B, and the third reflection structure 210C and the fourth reflection structure 210D may also be symmetrically arranged in the second lens barrel 202B relative to the axis of the second lens barrel 202B. In this way, the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D may be symmetrical, and the processing efficiency of the subsequent algorithms may be further improved.
在一些实施例中,如图2D所示,第一相机208A和第二相机208B均朝向第一显示模组204A且第一相机208A和第二相机208B的朝向与第一镜筒202A的轴线的夹角相等,例如,夹角都是γ;同时,第一反射结构210A和第二反射结构210B的朝向与第一镜筒202A的轴线的夹角也是相等的。这样,第一相机208A和第二相机208B所采集的第一眼1022A的人眼图像可以是严格对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in FIG2D , the first camera 208A and the second camera 208B are both facing the first display module 204A, and the angles between the first camera 208A and the second camera 208B and the axis of the first lens barrel 202A are equal, for example, the angles are both γ; at the same time, the angles between the first reflective structure 210A and the second reflective structure 210B and the axis of the first lens barrel 202A are also equal. In this way, the human eye images of the first eye 1022A collected by the first camera 208A and the second camera 208B can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
在一些实施例中,如图2D所示,第三相机208C和第四相机208D均朝向第二显示模组204B且第三相机208C和第四相机208D的朝向与第二镜筒202B的轴线的夹角相等,例如,夹角都是γ;同时,第三反射结构210C和第四反射结构210D的朝向与第二镜筒202B的轴线的夹角也是相等的。这样,第三相机208C和第四相机208D所采集的第二眼1022B的人眼图像可以是严格对称的,后续算法的处理效率可以进一步提高。In some embodiments, as shown in FIG2D , the third camera 208C and the fourth camera 208D are both facing the second display module 204B and the angles between the third camera 208C and the fourth camera 208D and the axis of the second lens barrel 202B are equal, for example, the angles are both γ; at the same time, the angles between the third reflective structure 210C and the fourth reflective structure 210D and the axis of the second lens barrel 202B are also equal. In this way, the human eye images of the second eye 1022B collected by the third camera 208C and the fourth camera 208D can be strictly symmetrical, and the processing efficiency of the subsequent algorithm can be further improved.
前述实施例均以两个镜筒内分别设置两个相机进行说明,可以理解,可以运行两个镜筒中的其中之一设置两个或多个相机,而另一镜筒设置一个相机,这种情况下,仍然能够在一定程度上提升算法精度,也应属于本公开的保护范围。此外,图2A和图2B所示的结构也可以在同一可穿戴设备200中进行搭配,例如,对应于第一眼的镜筒具有图2A所示的结构,对应于第二眼的镜筒具有图2B所示的结构。当然,也可以调换第一眼和第二眼对应的结构。The above embodiments are all described by setting two cameras in two lens barrels respectively. It can be understood that one of the two lens barrels can be set with two or more cameras, and the other lens barrel can be set with one camera. In this case, the algorithm accuracy can still be improved to a certain extent, which should also fall within the protection scope of the present disclosure. In addition, the structures shown in Figures 2A and 2B can also be matched in the same wearable device 200. For example, the lens barrel corresponding to the first eye has the structure shown in Figure 2A, and the lens barrel corresponding to the second eye has the structure shown in Figure 2B. Of course, the structures corresponding to the first eye and the second eye can also be swapped.
从上述实施例可以看出,本公开实施例通过把至少两个相机208设置在可穿戴设备200的至少一个镜筒的内部,能够得到更好的成像质量并可以提高可穿戴设备200的舒适度。在一些实施例中,通过在可穿戴设备200的左右两个镜筒内各放置2个相机,不仅可以避免镜筒外相机带来的前述不利影响,2个相机可以提供更多的观测信息,且受遮挡影响更小,因此可以提高瞳距估计或视线追踪等算法的精度。It can be seen from the above embodiments that the embodiments of the present disclosure can obtain better imaging quality and improve the comfort of the wearable device 200 by setting at least two cameras 208 inside at least one lens barrel of the wearable device 200. In some embodiments, by placing two cameras in each of the left and right lens barrels of the wearable device 200, not only can the aforementioned adverse effects caused by the cameras outside the lens barrel be avoided, but the two cameras can provide more observation information and are less affected by occlusion, so the accuracy of algorithms such as pupil distance estimation or line of sight tracking can be improved.
进一步地,与将相机放置在镜筒外相比,将相机放置在镜筒内,在相机进行成像时,除了受相机模组本身的影响,还会受到镜筒中光学组件的影响。Furthermore, compared with placing the camera outside the lens barrel, placing the camera inside the lens barrel will affect the camera imaging not only by the camera module itself but also by the optical components in the lens barrel.
如图2A和图2B所示,相对于图1B的相机1048,在相机208A、208B到达人眼1022的光路中,还存在镜筒内的光学组件206或光学组件206的一部分(根据相机208A、208B与光学组件206中各透镜的相对位置关系的不同而不同,例如,除了图2A和图2B所示的位置,相机208A、208B还可以设置在透镜2062和透镜2064之间),使得相机208A、208B在对人眼1022进行成像时,光学组件206或光学组件206的一部分影响或改变了相机208A、208B到达人眼1022的光路,导致相机208A、208B成像所得的图像的畸变可能不再对称。As shown in FIGS. 2A and 2B , relative to the camera 1048 of FIG. 1B , in the optical path from the cameras 208A and 208B to the human eye 1022, there is an optical component 206 or a portion of the optical component 206 in the lens barrel (depending on the relative position relationship between the cameras 208A and 208B and the lenses in the optical component 206, for example, in addition to the positions shown in FIGS. 2A and 2B , the cameras 208A and 208B may also be arranged between the lens 2062 and the lens 2064), so that when the cameras 208A and 208B image the human eye 1022, the optical component 206 or a portion of the optical component 206 affects or changes the optical path from the cameras 208A and 208B to the human eye 1022, resulting in that the distortion of the images obtained by the imaging of the cameras 208A and 208B may no longer be symmetrical.
并且,由于相机208A、208B到达人眼1022的光路被改变,使得相机208A、208B没有一个统一的投影中心,导致无法使用参数化相机模型去拟合镜筒内相机的投影过程来对畸变进行处理,使得相机参数难以标定。Furthermore, since the optical paths from the cameras 208A and 208B to the human eye 1022 are changed, the cameras 208A and 208B do not have a unified projection center, which makes it impossible to use a parameterized camera model to fit the projection process of the cameras in the lens barrel to process the distortion, making it difficult to calibrate the camera parameters.
此外,在一些情况下,如图2C和图2D所示,由于不同相机在可穿戴设备200中设置的位置不同,因此,还需要对不同相机之间的位姿关系进行标定,这样的标定也比较难以实现。特别是,当不同相机位于不同镜筒内时(例如,第一相机208A与第三相机208C或第四相机208D之间,或者,第二相机208B与第三相机208C或第四相机208D之间),这样的标定将会变得更加复杂。In addition, in some cases, as shown in FIG. 2C and FIG. 2D , since different cameras are set at different positions in the wearable device 200, it is also necessary to calibrate the posture relationship between different cameras, and such calibration is also difficult to implement. In particular, when different cameras are located in different lens barrels (for example, between the first camera 208A and the third camera 208C or the fourth camera 208D, or between the second camera 208B and the third camera 208C or the fourth camera 208D), such calibration will become more complicated.
有鉴于此,本公开实施例还提供了一种相机参数的标定方法,可以使用非参数化相机模型来实现对镜筒内相机的参数标定。在一些实施例中,对于处于不同镜筒内的相机,提出了一种基于非共视点的相机外参标定算法,从而解决了不同镜筒内相机标定难的问题。In view of this, the embodiments of the present disclosure also provide a camera parameter calibration method, which can use a non-parametric camera model to implement parameter calibration of the camera in the lens barrel. In some embodiments, for cameras in different lens barrels, a camera extrinsic calibration algorithm based on non-common viewpoints is proposed, thereby solving the problem of difficult calibration of cameras in different lens barrels.
图3A示出了本公开实施例所提供的示例性方法300的流程示意图。FIG. 3A is a schematic flow chart of an exemplary method 300 provided in an embodiment of the present disclosure.
该方法300,可以应用于任意的具有数据处理能力的计算机设备,并可以用于对图2A、图2B、图2C或图2D所示的可穿戴设备200中的相机208A、208B、208C或208D的参数进行标定。如图3A所示,该方法300可以进一步包括以下步骤。The method 300 can be applied to any computer device with data processing capabilities, and can be used to calibrate the parameters of the camera 208A, 208B, 208C or 208D in the wearable device 200 shown in Figure 2A, Figure 2B, Figure 2C or Figure 2D. As shown in Figure 3A, the method 300 can further include the following steps.
在步骤302,可以接收由至少两个相机拍摄标定参考对象而得到的多张图像。在本步骤中,所述多张图像可以是由不同相机采集的图像。作为一个可选实施例,为了实现对相机的第一参数(例如,内参)和第二参数(例如,外参)的标定,可以分别针对内参标定采集多张图像以及针对外参标定采集多张图像。In step 302, multiple images obtained by photographing a calibration reference object by at least two cameras may be received. In this step, the multiple images may be images captured by different cameras. As an optional embodiment, in order to implement calibration of a first parameter (e.g., an intrinsic parameter) and a second parameter (e.g., an extrinsic parameter) of a camera, multiple images may be captured for intrinsic parameter calibration and multiple images may be captured for extrinsic parameter calibration, respectively.
图4A示出了根据本公开实施例的一种示例性采图场景400的示意图。FIG4A shows a schematic diagram of an exemplary image collection scenario 400 according to an embodiment of the present disclosure.
如图4A所示,在采图场景400中,可以在某一固定位置设置一个标定参考对象402。在一些实施例中,该标定参考对象402可以是标定板或图卡,在该标定参考对象402上设置了一些黑白的棋盘格。当相机拍摄该标定参考对象402得到相应的图像后,根据图像中的棋盘格与标定参考对象402的棋盘格的对应关系,则可以计算出相应的相机参数。As shown in FIG4A , in the image acquisition scene 400, a calibration reference object 402 may be set at a fixed position. In some embodiments, the calibration reference object 402 may be a calibration plate or a chart, and some black and white checkerboards are set on the calibration reference object 402. When the camera captures the calibration reference object 402 to obtain a corresponding image, the corresponding camera parameters may be calculated according to the correspondence between the checkerboards in the image and the checkerboards of the calibration reference object 402.
在对标定参考对象402进行采图(或图像采集)时,可以控制待标定相机(例如,相机208A)沿某一运动轨迹进行运动,并在相机的运动过程中连续拍摄该标定参考对象402,从而得到相机在不同姿态下采集得到的多张图像,作为后续计算相机参数的数据。由于采集的图像是相机在多姿态下进行图像采集而得到的,后续计算得到的相机参数的鲁棒性更好,能够适用更广的范围。When capturing images (or capturing images) of the calibration reference object 402, the camera to be calibrated (e.g., camera 208A) can be controlled to move along a certain motion trajectory, and the calibration reference object 402 can be continuously captured during the movement of the camera, thereby obtaining multiple images captured by the camera in different postures, which are used as data for subsequent calculation of camera parameters. Since the captured images are obtained by the camera performing image capture in multiple postures, the camera parameters calculated subsequently are more robust and can be applied to a wider range.
在一些实施例中,若设置在可穿戴设备中的所述至少两个相机是不同类型的相机或具有不同相机参数的相机,则针对每种相机或每种参数的相机需要分别按照上述方式来采集多张图像,用于对相应相机的第一参数(例如,内参)进行标定。若设置在可穿戴设备中的所述至少两个相机的相机参数是一致的,则可以选择一个相机来实施前述的采图操作,从而用于对相机的内参进行标定,标定得到的内参可以适用于可穿戴设备中的其他相机。In some embodiments, if the at least two cameras provided in the wearable device are cameras of different types or cameras with different camera parameters, multiple images need to be collected in the above manner for each camera or camera with each parameter, respectively, for calibrating the first parameter (e.g., internal parameter) of the corresponding camera. If the camera parameters of the at least two cameras provided in the wearable device are consistent, a camera can be selected to implement the aforementioned image acquisition operation, so as to calibrate the internal parameters of the camera, and the calibrated internal parameters can be applied to other cameras in the wearable device.
图4B示出了根据本公开实施例的另一种示例性采图场景410的示意图。FIG4B shows a schematic diagram of another exemplary image collection scenario 410 according to an embodiment of the present disclosure.
如图4B所示,与采图场景400类似的,在采图场景410中,也可以在某一固定位置设置一个标定参考对象402。As shown in FIG. 4B , similar to the image collection scene 400 , in the image collection scene 410 , a calibration reference object 402 may also be set at a fixed position.
与采图场景400不同的是,为了计算可穿戴设备200中至少两个相机之间的位姿关系(第二参数),可以将所述至少两个相机安装在可穿戴设备200的镜筒中并完成可穿戴设备200的完整装配后,利用装配完成的可穿戴设备200或其样机来进行外参采图,进而基于采集的图像来计算得到相应的外参(例如,不同相机之间的位姿关系)。Different from the image collection scene 400, in order to calculate the pose relationship (second parameter) between at least two cameras in the wearable device 200, the at least two cameras can be installed in the lens barrel of the wearable device 200 and the wearable device 200 can be fully assembled. The assembled wearable device 200 or its prototype can be used to perform external parameter image collection, and then the corresponding external parameters (for example, the pose relationship between different cameras) can be calculated based on the collected images.
如图4B所示,以可穿戴设备200的两个镜筒中分别设置两个相机为例,在对标定参考对象402进行采图(或图像采集)时,可以控制可穿戴设备200沿某一运动轨迹进行运动,并在可穿戴设备200的运动过程中利用四个相机连续拍摄该标定参考对象402,从而在四个相机的相对位姿关系不变的情况下,得到四个相机在不同姿态下采集得到的多张图像,作为后续计算第二参数的图像数据。由于采集的图像是相机在多姿态下进行图像采集而得到的,后续计算得到的相机参数的鲁棒性更好,能够适用更广的范围。As shown in FIG. 4B , taking the case where two cameras are respectively arranged in the two lens barrels of the wearable device 200, when capturing images (or capturing images) of the calibration reference object 402, the wearable device 200 can be controlled to move along a certain motion trajectory, and the calibration reference object 402 can be continuously captured by four cameras during the movement of the wearable device 200, so that when the relative posture relationship of the four cameras remains unchanged, multiple images captured by the four cameras in different postures are obtained as image data for subsequent calculation of the second parameter. Since the captured images are obtained by the camera capturing images in multiple postures, the camera parameters calculated subsequently have better robustness and can be applied to a wider range.
根据上述内容,可以理解的是,在采图场景410中,每个相机均可以采集到多张图像,这些图像除了可以用于计算四个相机的相对位姿关系外,每个相机所采集的图像也能用于对其自身进行内参计算,因此,在一些实施例中,可以仅使用采图场景410来采集图像,并可以基于这些图像来对相机的第一参数和第二参数进行标定。Based on the above content, it can be understood that in the image acquisition scene 410, each camera can capture multiple images. In addition to being used to calculate the relative posture relationship of the four cameras, the images captured by each camera can also be used to calculate its own internal parameters. Therefore, in some embodiments, only the image acquisition scene 410 can be used to capture images, and the first parameter and the second parameter of the camera can be calibrated based on these images.
在步骤302中,计算机设备可以接收在前述场景400和/或场景410中采集得到的多张图像以用于后续处理。In step 302, the computer device may receive a plurality of images captured in the aforementioned scene 400 and/or scene 410 for subsequent processing.
在采集得到所需的多张图像之后,可以基于所述多张图像来对相机进行参数标定。参数标定可以包括对第一参数和第二参数进行标定,其中,第一参数可以是相机的内参,第二参数可以是相机的外参。After acquiring the required multiple images, the camera can be calibrated based on the multiple images. The parameter calibration can include calibrating a first parameter and a second parameter, wherein the first parameter can be an intrinsic parameter of the camera and the second parameter can be an extrinsic parameter of the camera.
在一些实施例中,可以先对相机的内参进行标定。但是,如前所述,由于相机被设置在镜筒中,相机成像不仅受本身模组的影响,也会受到镜筒的透镜作用,使得成像畸变不再关于图像主点对称,并且,相机没有一个统一的投影中心,导致无法使用参数化相机模型去拟合镜筒内相机的投影过程来对不对称的畸变进行处理。因此,在一些实施例中,提供了一种非参数化相机模型来对相机的内参进行标定。In some embodiments, the intrinsic parameters of the camera may be calibrated first. However, as mentioned above, since the camera is set in the lens barrel, the camera imaging is not only affected by the module itself, but also by the lens effect of the lens barrel, so that the imaging distortion is no longer symmetrical about the image principal point, and the camera does not have a unified projection center, which makes it impossible to use a parametric camera model to fit the projection process of the camera in the lens barrel to handle the asymmetric distortion. Therefore, in some embodiments, a non-parametric camera model is provided to calibrate the intrinsic parameters of the camera.
于是,在步骤304,可以先确定所述多张图像的像素点与所述标定参考对象的投射关系。本步骤中,可以根据采集得到的图像来建立图像与标定参考对象402之间的投射关系,从而建立像素点与投射方向(pixel-ray)的对应关系。Therefore, the projection relationship between the pixels of the multiple images and the calibration reference object may be determined in step 304. In this step, the projection relationship between the image and the calibration reference object 402 may be established based on the acquired image, thereby establishing a corresponding relationship between the pixel and the projection direction (pixel-ray).
可以理解,图像的像素点与标定参考对象的投射关系表示了相机的固有参数(即内参),因此,在对特定相机(例如,相机208A)进行投射关系计算时,需要使用该特定相机采集的图像来建立所述投射关系。因此,以图4B所示的采图场景为例,所述多张图像中需要区分第一相机208A采集的多张第一图像、第二相机208B采集的多张第二图像、第三相机208C采集的多张第三图像、第四相机208D采集的多张第四图像。It can be understood that the projection relationship between the pixel points of the image and the calibration reference object represents the intrinsic parameters (i.e., the internal parameters) of the camera. Therefore, when calculating the projection relationship for a specific camera (e.g., camera 208A), it is necessary to use the images captured by the specific camera to establish the projection relationship. Therefore, taking the image capture scene shown in FIG. 4B as an example, it is necessary to distinguish among the multiple images the multiple first images captured by the first camera 208A, the multiple second images captured by the second camera 208B, the multiple third images captured by the third camera 208C, and the multiple fourth images captured by the fourth camera 208D.
下面以第一相机208A为例进行投射关系计算。The following takes the first camera 208A as an example to perform projection relationship calculation.
在一些实施例中,如图3B所示,确定所述多张图像的像素点与所述标定参考对象的投射关系的步骤304,可以进一步包括以下步骤:In some embodiments, as shown in FIG. 3B , step 304 of determining the projection relationship between the pixels of the plurality of images and the calibration reference object may further include the following steps:
在步骤3042,从所述多张图像中选取第一数量的目标图像。In step 3042, a first number of target images are selected from the plurality of images.
本步骤中,可以通过从所述多张图像中的多张第一图像(第一相机208A采集的图像)中选取一定数量的目标图像先建立投射关系,剩余第一图像可以用于补足未建立投射关系的部分。In this step, a projection relationship can be first established by selecting a certain number of target images from the multiple first images (images captured by the first camera 208A) in the multiple images, and the remaining first images can be used to supplement the parts for which the projection relationship is not established.
可以理解,该第一数量只要足以继续进行后续步骤即可,其数量不做具体限制。作为一个可选实施例,第一数量可以是3。图4C示出了被选出的3张目标图像412、414、416。如图4C所示,目标图像412、414、416分别示出了第一相机208A在不同姿态下拍摄标定参考对象402而得的图像,可以理解,通过对不同相机姿态下采集的图像进行相机参数的标定,可以使计算得到的相机参数的鲁棒性更好。It can be understood that the first number is not specifically limited as long as it is sufficient to proceed with the subsequent steps. As an optional embodiment, the first number can be 3. FIG. 4C shows three selected target images 412, 414, and 416. As shown in FIG. 4C, the target images 412, 414, and 416 respectively show images obtained by the first camera 208A taking the calibration reference object 402 in different postures. It can be understood that by calibrating the camera parameters of the images captured in different camera postures, the calculated camera parameters can be made more robust.
在步骤3044,在每张所述目标图像中选取至少一个局部区域(例如,棋盘格中的一个格子对应的区域可以被认为是一个局部区域),并建立每个所述局部区域与所述标定参考对象的单应性变换矩阵。In step 3044, at least one local area is selected in each of the target images (for example, the area corresponding to a grid in the chessboard can be considered as a local area), and a homography transformation matrix between each of the local areas and the calibration reference object is established.
在选取出每张目标图像的至少一个局部区域之后,针对每张目标图像的每个局部区域,可以分别构建与标定参考对象402的单应性变换矩阵,以建立局部区域所包含的每个像素点与标定参考对象402的坐标系之间(pixel-coordinate)的对应关系。After selecting at least one local area of each target image, a homography transformation matrix with the calibration reference object 402 can be constructed for each local area of each target image to establish a corresponding relationship (pixel-coordinate) between each pixel point contained in the local area and the coordinate system of the calibration reference object 402.
可以知道,标定参考对象402所在位置对的坐标系是已知的,并且,标定参考对象402上的棋盘格与目标图像中的棋盘格图像具有对应关系,同时局部区域对应的四个顶点在该目标图像所在的相机坐标系中的坐标也是已知的,根据这些已知信息,则可以局部区域与标定参考对象402的单应性变换矩阵H。It can be known that the coordinate system of the position of the calibration reference object 402 is known, and the checkerboard on the calibration reference object 402 corresponds to the checkerboard image in the target image. At the same time, the coordinates of the four vertices corresponding to the local area in the camera coordinate system of the target image are also known. Based on these known information, the homography transformation matrix H between the local area and the calibration reference object 402 can be obtained.
这样,就将局部区域所包含的像素点与标定参考对象402或标定参考对象402所在坐标系建立了对应关系。In this way, a corresponding relationship is established between the pixel points included in the local area and the calibration reference object 402 or the coordinate system where the calibration reference object 402 is located.
在步骤3046,根据所述单应性变换矩阵确定所述局部区域包含的像素点与所述标定参考对象的投射关系。In step 3046, the projection relationship between the pixel points included in the local area and the calibration reference object is determined according to the homography transformation matrix.
图4D示出了根据本公开实施例的像素点与标定参考对象的投射关系的示意图。FIG4D is a schematic diagram showing a projection relationship between a pixel point and a calibration reference object according to an embodiment of the present disclosure.
如图4D所示,以图中4条虚线连接的4个顶点对应的区域为所述局部区域,通过构建所述单应性变换矩阵,可以利用该单应性变换矩阵来计算目标图像中所述局部区域所包含的每个像素点与标定参考对象上对应的点之间的投影关系。As shown in Figure 4D, the area corresponding to the four vertices connected by the four dotted lines in the figure is the local area. By constructing the homography transformation matrix, the homography transformation matrix can be used to calculate the projection relationship between each pixel point contained in the local area in the target image and the corresponding point on the calibration reference object.
在步骤3048,将所述第一数量的目标图像转换到参考坐标系下,并根据所述局部区域对应的已建立投射关系的区域,确定所述多张图像的像素点与所述标定参考对象的投射关系。In step 3048, the first number of target images are converted into a reference coordinate system, and the projection relationship between the pixel points of the multiple images and the calibration reference object is determined based on the area with the established projection relationship corresponding to the local area.
如图4C所示,3张目标图像412、414、416是在不同相机姿态下采集得到的,前面在构建局部区域的单应性变换矩阵时,是以目标图像各自对应的相机坐标系来进行构建的,可以理解,为了将各目标图像得到的投影关系在同一参考坐标系下进行统一,本步骤中,可以把3张目标图像412、414、416转换到参考坐标系下。As shown in FIG4C , the three target images 412 , 414 , and 416 are acquired under different camera postures. When constructing the homography transformation matrix of the local area, it is constructed based on the camera coordinate system corresponding to each target image. It can be understood that in order to unify the projection relationship obtained by each target image under the same reference coordinate system, in this step, the three target images 412 , 414 , and 416 can be converted to the reference coordinate system.
在按照前述步骤完成3张目标图像412、414、416的处理之后,可以采用与前面类似的方法,在剩余的第一图像中继续选取新的目标图像的局部区域来构建单应性变换矩阵,直至完成整张图像中的每个像素点与所述标定参考对象的投射关系。After completing the processing of the three target images 412, 414, and 416 according to the above steps, a method similar to the previous one can be used to continue selecting local areas of new target images in the remaining first image to construct a homography transformation matrix until the projection relationship between each pixel point in the entire image and the calibration reference object is completed.
根据前述实施例可以知道,在一些实施例中,若将全部图像处理完成且整张图像也完成标定之后,可能存在有些像素点对应了多个投射方向数据,因此,可以将这些数据进行均值处理之后得到一个投射方向来作为该像素点的投射方向来得到更好的投射方向数据。此外,后续存储时也可以仅存储这一个均值处理后的投射方向,从而节省存储空间。According to the above embodiments, it can be known that in some embodiments, after all image processing is completed and the entire image is calibrated, some pixels may correspond to multiple projection direction data. Therefore, these data can be averaged to obtain a projection direction as the projection direction of the pixel to obtain better projection direction data. In addition, only the projection direction after the average processing can be stored in subsequent storage, thereby saving storage space.
可以理解,针对第二相机208B、第三相机208C、第四相机208D以及可能的更多个相机(视可穿戴设备中相机设置数量而定),均可以采用前述的方法来建立图像像素点与标定参考对象402的投射关系,在此不再赘述。It can be understood that the aforementioned method can be used to establish the projection relationship between the image pixels and the calibration reference object 402 for the second camera 208B, the third camera 208C, the fourth camera 208D and possibly more cameras (depending on the number of cameras set in the wearable device), and will not be repeated here.
在建立好投射关系之后,在步骤306,可以根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集。其中,所述第一参数集包括多个第一参数,所述第一参数用于表征所述待标定相机所拍摄的图像中的目标像素点和所述目标像素点对应的投射方向,所述第二参数集包括至少一组第二参数,所述至少一组第二参数用于指示所述至少两个相机之间的位姿关系。After the projection relationship is established, the first parameter set and the second parameter set of the at least two cameras may be determined according to the projection relationship in step 306. The first parameter set includes a plurality of first parameters, which are used to characterize the target pixel points in the image captured by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one set of second parameters, which are used to indicate the posture relationship between the at least two cameras.
在一些实施例中,可以先根据所述投射关系,确定所述至少两个相机的第一参数集,然后在确定第二参数集。In some embodiments, a first parameter set of the at least two cameras may be determined first according to the projection relationship, and then a second parameter set may be determined.
图4E示出了根据本公开实施例的示例性第一参数的示意图。FIG. 4E shows a schematic diagram of an exemplary first parameter according to an embodiment of the present disclosure.
如图4E所示,在构建前述的投影关系时,在参考坐标系下的图像中的至少部分像素点与标定参考对象上的一些点一一对应,从而可以得到一个关于像素点的直线方程,该直线方程所表示的直线穿过对应的像素点并具有一个如图所示的射线(ray)来表示其投射方向。这样的像素点与投射方向的组合可以作为第一参数(该相机参数可以被认为是相机的内参),利用该相机参数实现相机采集的图像的坐标系到相机坐标系的转换,从而可以用于计算相机采集的图像中某些特征(例如,瞳孔)在三维空间(相机坐标系)中的位置信息。并且,由于该相机参数的计算过程中利用第一图像的多个局部区域来分别建立与标定参考对象的单应性变换矩阵,从而利用该相机参数进行坐标系转换计算时,获取的像素点对应的投射方向已包含了对该位置图像畸变的校正,而无需再额外进行畸变校正。As shown in FIG. 4E , when constructing the aforementioned projection relationship, at least some of the pixels in the image under the reference coordinate system correspond to some points on the calibration reference object one by one, so that a straight line equation about the pixel can be obtained, and the straight line represented by the straight line equation passes through the corresponding pixel and has a ray (ray) as shown in the figure to represent its projection direction. Such a combination of pixel points and projection directions can be used as a first parameter (the camera parameter can be considered as an internal parameter of the camera), and the camera parameter is used to realize the conversion from the coordinate system of the image captured by the camera to the camera coordinate system, so that it can be used to calculate the position information of certain features (for example, pupil) in the image captured by the camera in three-dimensional space (camera coordinate system). In addition, since multiple local areas of the first image are used to respectively establish homography transformation matrices with the calibration reference object during the calculation of the camera parameters, when the camera parameters are used to perform coordinate system conversion calculations, the projection direction corresponding to the pixel point obtained already includes the correction of the image distortion at the position, and no additional distortion correction is required.
在一些实施例中,可以对前面得到的数据进行精简后作为相机的第一参数,从而可以节省存储第一参数的空间。例如,可以利用样条曲面对前述标定得到的第一参数的初始集合进行拟合,然后利用光束法平差(BA)算法对拟合后的样条曲面的控制点进行优化,将优化得到的样条曲面的所有控制点对应的第一参数座位所述至少两个相机的第一参数集。In some embodiments, the previously obtained data may be simplified and used as the first parameters of the camera, thereby saving space for storing the first parameters. For example, a spline surface may be used to fit the initial set of first parameters obtained by the aforementioned calibration, and then the control points of the fitted spline surface may be optimized using a bundle adjustment (BA) algorithm, and the first parameters corresponding to all the control points of the optimized spline surface may be used as the first parameter set of the at least two cameras.
当需要使用第一参数时,根据该拟合样条曲面和所述优化后的控制点,通过插值算法就可以得到完整的第一参数的集合。When the first parameters need to be used, a complete set of first parameters can be obtained through an interpolation algorithm based on the fitting spline surface and the optimized control points.
可以看出,前述方法仅给出了针对单个相机的第一参数的计算方式,针对可穿戴设备200中的每个相机,均可以采用上述方法计算得到相应的第一参数,在此不再赘述。It can be seen that the above method only provides a calculation method for the first parameter of a single camera. For each camera in the wearable device 200, the above method can be used to calculate the corresponding first parameter, which will not be repeated here.
在得到第一参数集之后,可以进一步标定第二参数集。After obtaining the first parameter set, the second parameter set may be further calibrated.
回到图2C,可穿戴设备200中包括设置在第一镜筒202A中的用于采集第一眼1022A的人眼图像的第一相机208A和第二相机208B以及设置在第二镜头202B中的用于采集第二眼1022B的人眼图像的第三相机208C和第四相机208D。由于四个相机在可穿戴设备200中的所处位置是不同的,为了确定四个相机所采集的图像的相对关系,就需要知道四个相机之间的位姿关系来用作第二参数,从而完成相机外参的标定。Returning to FIG. 2C , the wearable device 200 includes a first camera 208A and a second camera 208B disposed in the first lens barrel 202A for collecting human eye images of the first eye 1022A, and a third camera 208C and a fourth camera 208D disposed in the second lens 202B for collecting human eye images of the second eye 1022B. Since the four cameras are located at different positions in the wearable device 200, in order to determine the relative relationship of the images collected by the four cameras, it is necessary to know the position relationship between the four cameras as the second parameter, thereby completing the calibration of the camera extrinsic parameters.
下面以对可穿戴设备200中的四个相机进行标定为例,示例性地说明如何计算第二参数。The following takes the calibration of four cameras in the wearable device 200 as an example to illustrate how to calculate the second parameter.
如图2C所示,由于设置在可穿戴设备200的同一镜筒内的两个相机需要对同一人眼进行图像采集,因此,一般地,这两个相机具有共视点(或光路交汇点),因此,可以采用双目标定算法来求得处于同一镜筒内的两个相机之间的位姿关系。As shown in FIG2C , since the two cameras arranged in the same lens barrel of the wearable device 200 need to capture images of the same human eye, generally, the two cameras have a common viewpoint (or an intersection of optical paths). Therefore, a binocular positioning algorithm can be used to obtain the posture relationship between the two cameras in the same lens barrel.
因此,在一些实施例中,所述第二参数可以包括所述第一相机208A和所述第二相机208B之间的位姿关系以及所述第三相机208C和所述第四相机208D之间的位姿关系,所述第一相机和所述第二相机之间的位姿关系以及所述第三相机和所述第四相机之间的位姿关系是基于双目标定算法根据所述投射关系计算得到的。Therefore, in some embodiments, the second parameter may include the pose relationship between the first camera 208A and the second camera 208B and the pose relationship between the third camera 208C and the fourth camera 208D, and the pose relationship between the first camera and the second camera and the pose relationship between the third camera and the fourth camera are calculated according to the projection relationship based on a binocular positioning algorithm.
可选地,根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集的步骤306,可以进一步包括:根据所述投射关系,利用双目标定算法,确定第一相机208A和第二相机208B之间的位姿关系和第三相机208C和第四相机208D之间的位姿关系。Optionally, step 306 of determining the first parameter set and the second parameter set of the at least two cameras according to the projection relationship may further include: determining the posture relationship between the first camera 208A and the second camera 208B and the posture relationship between the third camera 208C and the fourth camera 208D according to the projection relationship using a binocular positioning algorithm.
继续参照图2C,可以理解,由于不同镜筒内的相机采集的是不同人眼的图像,因此,不同镜筒内的相机可能并不存在共视点,不能采用双目标定算法来计算位姿关系,或者采用双目标定算法计算得到的位姿关系可能是不准确的。因此,本公开的实施例提供了一种计算不同镜筒内相机之间位姿关系的方法。2C, it can be understood that, because the cameras in different lens barrels collect images of different human eyes, the cameras in different lens barrels may not have a common viewpoint, and the binocular positioning algorithm cannot be used to calculate the pose relationship, or the pose relationship calculated by the binocular positioning algorithm may be inaccurate. Therefore, an embodiment of the present disclosure provides a method for calculating the pose relationship between cameras in different lens barrels.
图4F示出了根据本公开实施例的示例性的异侧镜筒内相机的第二参数的标定原理图。FIG. 4F shows a schematic diagram of a calibration principle of a second parameter of an exemplary camera in a different-side lens barrel according to an embodiment of the present disclosure.
[根据细则91更正 15.01.2025]
如图4F所示,以计算第一相机208A和第四相机208D之间的位姿关系为例,两个相机没有共视点,在二者运动到某个位置下,两个相机与标定参考对象402的相对位置如图4F所示,第一相机208A相对标定参考对象402的位姿关系为T1,i,第四相机208D相对标定参考对象402的位姿关系为T4,i,第一相机208A和第四相机208D之间的位姿关系为T14,则利用第一相机208A和第四相机208D的观测数据(采图步骤得到的第一图像集合和第四图像集合),可构建BA优化问题,并通过求解该问题来得到位姿关系T14。[Corrected 15.01.2025 in accordance with Article 91]
As shown in FIG4F , taking the calculation of the pose relationship between the first camera 208A and the fourth camera 208D as an example, the two cameras have no common viewpoint. When the two cameras move to a certain position, the relative positions of the two cameras and the calibration reference object 402 are shown in FIG4F . The pose relationship of the first camera 208A relative to the calibration reference object 402 is T 1,i , the pose relationship of the fourth camera 208D relative to the calibration reference object 402 is T 4,i , and the pose relationship between the first camera 208A and the fourth camera 208D is T 14 . Then, using the observation data of the first camera 208A and the fourth camera 208D (the first image set and the fourth image set obtained in the mapping step), a BA optimization problem can be constructed, and the pose relationship T 14 can be obtained by solving the problem.
因此,在一些实施例中,所述第二参数还可以包括所述第一相机和所述第三相机之间的位姿关系、所述第一相机和所述第四相机之间的位姿关系、所述第二相机和所述第三相机之间的位姿关系、所述第二相机和所述第四相机之间的位姿关系。Therefore, in some embodiments, the second parameter may also include a pose relationship between the first camera and the third camera, a pose relationship between the first camera and the fourth camera, a pose relationship between the second camera and the third camera, and a pose relationship between the second camera and the fourth camera.
可选地,根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集的步骤306,可以进一步包括:根据所述多张图像和所述投射关系,确定所述第一相机和所述第三相机之间的位姿关系、所述第一相机和所述第四相机之间的位姿关系、所述第二相机和所述第三相机之间的位姿关系、所述第二相机和所述第四相机之间的位姿关系。Optionally, step 306 of determining the first parameter set and the second parameter set of the at least two cameras according to the projection relationship may further include: determining the posture relationship between the first camera and the third camera, the posture relationship between the first camera and the fourth camera, the posture relationship between the second camera and the third camera, and the posture relationship between the second camera and the fourth camera according to the multiple images and the projection relationship.
具体地,可以先构建优化函数。Specifically, an optimization function may be constructed first.
可选地,所述优化函数可以包括用于表示第一相机208A采集的第一图像中的检测点与该检测点对应的空间位置(标定参考对象402上的对应位置)利用该检测点对应的第一相机208A的第一参数投影到图像坐标系中的二维点之间的误差的第一公式。可选地,该第一公式表示为:Optionally, the optimization function may include a first formula for representing the error between a detection point in the first image captured by the first camera 208A and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using a first parameter of the first camera 208A corresponding to the detection point. Optionally, the first formula is expressed as:
fcam1,1=π1(T1,i,Pcam1)-dcam1 f cam1,1 =π1(T 1,i ,P cam1 )-d cam1
其中,π1为第一相机208A对应的投影函数(即前面得到的像素点与标定参考对象402的投射关系),T1,i为第i个第一图像对应的第一相机208A与标定参考对象402的外参(可以根据第一图像以及标定参考对象402的空间坐标来计算得到),Pcam1为第一相机208A的第i个第一图像在标定参考对象402上对应的3D点(亦即第一图像的各像素点在标定参考对象402上对应的点的空间坐标),dcam1为第一相机208A对应的所有检测点。Wherein, π1 is the projection function corresponding to the first camera 208A (i.e., the projection relationship between the pixel points obtained previously and the calibration reference object 402), T 1,i is the external parameter of the first camera 208A and the calibration reference object 402 corresponding to the i-th first image (which can be calculated based on the spatial coordinates of the first image and the calibration reference object 402), P cam1 is the 3D point corresponding to the i-th first image of the first camera 208A on the calibration reference object 402 (i.e., the spatial coordinates of the points corresponding to each pixel point of the first image on the calibration reference object 402), and d cam1 is all the detection points corresponding to the first camera 208A.
可选地,所述优化函数还可以包括用于表示第四相机208D采集的第四图像中的检测点与该检测点对应的空间位置(标定参考对象402上的对应位置)利用该检测点对应的第四相机208A的第一参数投影到图像坐标系中的二维点之间的误差的第二公式。可选地,该第二公式表示为:Optionally, the optimization function may further include a second formula for expressing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the fourth camera 208A corresponding to the detection point. Optionally, the second formula is expressed as:
fcam4,2=π4(T4,i,Pcam4)-dcam4 f cam4,2 =π4(T 4,i ,P cam4 )-d cam4
其中,π4为第四相机208D对应的投影函数(即前面得到的像素点与标定参考对象402的投射关系),T4,i为第i个第四图像对应的第四相机208D与标定参考对象402的外参(可以根据第四图像以及标定参考对象402的坐标信息来计算得到),Pcam4为第四相机208D的第i个第四图像在标定参考对象402上对应的3D点(亦即第四图像的各像素点在标定参考对象402上对应的点的空间坐标),dcam4为第四相机208D对应的所有检测点。Wherein, π4 is the projection function corresponding to the fourth camera 208D (that is, the projection relationship between the pixel points obtained previously and the calibration reference object 402), T 4,i is the external parameter of the fourth camera 208D and the calibration reference object 402 corresponding to the i-th fourth image (which can be calculated according to the coordinate information of the fourth image and the calibration reference object 402), P cam4 is the 3D point corresponding to the i-th fourth image of the fourth camera 208D on the calibration reference object 402 (that is, the spatial coordinates of the points corresponding to each pixel point of the fourth image on the calibration reference object 402), and d cam4 is all the detection points corresponding to the fourth camera 208D.
可选地,所述优化函数还可以包括用于表示第四相机208D采集的第四图像中的检测点与该检测点对应的空间位置(标定参考对象402上的对应位置)利用该检测点对应的第一相机208A的第一参数以及所述第一相机和所述第四相机之间的位姿关系投影到图像坐标系中的二维点之间的误差的第三公式。可选地,该第三公式表示为:Optionally, the optimization function may further include a third formula for representing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the first camera 208A corresponding to the detection point and the posture relationship between the first camera and the fourth camera. Optionally, the third formula is expressed as:
fcam4,1=π4(T14*T1,i,Pcam4)-dcam4 f cam4,1 =π4(T 14 *T 1,i ,P cam4 )-d cam4
其中,T14为所述第一相机208A和所述第四相机208D之间的位姿关系,即本公开实施例所需要求得的所述第一相机和所述第四相机的外参。Wherein, T14 is the posture relationship between the first camera 208A and the fourth camera 208D, that is, the external parameters of the first camera and the fourth camera required to be obtained in the embodiment of the present disclosure.
可选地,所述优化函数还可以包括用于表示第四相机208D采集的第四图像中的检测点与该检测点对应的空间位置(标定参考对象402上的对应位置)利用该检测点对应的所述第四相机的第一参数以及所述第一相机和所述第四相机之间的位姿关系投影到图像坐标系中的二维点之间的误差的第四公式。可选地,该第四公式表示为: Optionally, the optimization function may further include a fourth formula for representing the error between a detection point in the fourth image captured by the fourth camera 208D and a spatial position corresponding to the detection point (a corresponding position on the calibration reference object 402) projected to a two-dimensional point in the image coordinate system using the first parameter of the fourth camera corresponding to the detection point and the posture relationship between the first camera and the fourth camera. Optionally, the fourth formula is expressed as:
其中,为所述第一相机208A和所述第四相机208D之间的位姿关系的转置矩阵,通过对T14求转置而得到。in, is the transposed matrix of the position and posture relationship between the first camera 208A and the fourth camera 208D, which is obtained by transposing T14 .
在一些实施例中,所述检测点dcam1和dcam4可以采用以下方式进行计算。In some embodiments, the detection points d cam1 and d cam4 may be calculated in the following manner.
图4G示出了本公开实施例中采集得到的图像经过二值化处理后得到的图像的示意图。FIG. 4G is a schematic diagram showing an image obtained after the captured image is binarized in an embodiment of the present disclosure.
如图4G所示,经过二值化处理之后,图像上的像素点要么是黑色要么是白色,通过特征检测,可以识别棋盘格中格子的顶点,这些能够识别出的顶点则可以作为检测点d,在第i张图像上检测到的检测点则可以表示为di。As shown in FIG4G , after binarization, the pixels on the image are either black or white. Through feature detection, the vertices of the checkerboard grid can be identified. These identifiable vertices can be used as detection points d. The detection points detected on the i-th image can be represented as d i .
可选地,所述优化函数还可以进一步基于上述的第一公式、第二公式、第三公式、第四公式共同确定一个误差函数,可选地,该误差函数表示为:Optionally, the optimization function may further determine an error function based on the first formula, the second formula, the third formula, and the fourth formula. Optionally, the error function is expressed as:
f=fcam1,1+fcam4,1+fcam1,2+fcam4,2 f= fcam1,1 + fcam4,1 + fcam1,2 + fcam4,2
最后,可选地,可以基于所述误差函数构建优化函数,确定最终的第一相机208A与第四相机208D的外参。可选地,该优化函数表示如下: Finally, optionally, an optimization function may be constructed based on the error function to determine the final external parameters of the first camera 208A and the fourth camera 208D. Optionally, the optimization function is expressed as follows:
其中,ρ为损失函数,f为所述误差函数Oi为第i个图像所包含的所有信息,I为输入图像的张数。Among them, ρ is the loss function, f is the error function O i is all the information contained in the i-th image, and I is the number of input images.
可以看出,损失函数res(π,T)是一个关于位姿的函数,通过利用列文伯格-马夸尔特法(Levenberg-Marquardt,简称LM)对损失函数进行求解,得到的最优解就是最后的位姿关系T14。It can be seen that the loss function res(π, T) is a function of posture. By solving the loss function using the Levenberg-Marquardt method (LM for short), the optimal solution obtained is the final posture relationship T 14 .
可以理解,对于异侧镜筒内的两个相机,均可以采用上述方法来得到位姿关系,在此不再赘述。It can be understood that the above method can be used to obtain the posture relationship of the two cameras in the different side lens barrels, and will not be repeated here.
考虑到异侧相机的排列组合方式较多,如果每对相机都采用上述方法来计算位姿关系,会增加较多的计算量。因此,在一些实施例中,根据所述多张图像和所述投射关系,所述第一相机和所述第三相机之间的位姿关系、所述第一相机和所述第四相机之间的位姿关系、所述第二相机和所述第三相机之间的位姿关系、所述第二相机和所述第四相机之间的位姿关系,包括:根据所述多张图像和所述投射关系,确定所述第一相机与所述第四相机之间的位姿关系;根据所述第三相机和所述第四相机之间的位姿关系以及所述第一相机与所述第四相机之间的位姿关系,确定所述第一相机与所述第三相机之间的位姿关系;根据所述第一相机和所述第二相机之间的位姿关系以及所述第一相机与所述第三相机之间的位姿关系,确定所述第二相机与所述第三相机之间的位姿关系;根据所述第三相机和所述第四相机之间的位姿关系以及所述第二相机与所述第三相机之间的位姿关系,确定所述第二相机与所述第四相机之间的位姿关系。Considering that there are many permutations and combinations of cameras on opposite sides, if each pair of cameras uses the above method to calculate the pose relationship, the amount of calculation will increase. Therefore, in some embodiments, according to the multiple images and the projection relationship, the pose relationship between the first camera and the third camera, the pose relationship between the first camera and the fourth camera, the pose relationship between the second camera and the third camera, and the pose relationship between the second camera and the fourth camera include: determining the pose relationship between the first camera and the fourth camera according to the multiple images and the projection relationship; determining the pose relationship between the first camera and the third camera according to the pose relationship between the third camera and the fourth camera and the pose relationship between the first camera and the fourth camera; determining the pose relationship between the second camera and the third camera according to the pose relationship between the first camera and the second camera and the pose relationship between the first camera and the third camera; determining the pose relationship between the second camera and the fourth camera according to the pose relationship between the third camera and the fourth camera and the pose relationship between the second camera and the third camera.
这样,在计算得到第一相机208A和第四相机208D之间的位姿关系之后,基于已经计算得到的第一相机208A和第二相机208B之间的位姿关系以及第三相机208C和第四相机208D之间的位姿关系,可以通过数据转换的方式求得其他排列组合的位姿关系,从而节省了计算量。In this way, after calculating the posture relationship between the first camera 208A and the fourth camera 208D, based on the calculated posture relationship between the first camera 208A and the second camera 208B and the posture relationship between the third camera 208C and the fourth camera 208D, the posture relationships of other arrangements and combinations can be obtained through data conversion, thereby saving the amount of calculation.
可以理解,上述方法提供了异侧镜筒内的2个相机不存在共视点时可使用的计算方法,但是,当异侧相机存在共视点时,仍可以使用双目标定算法来计算位姿关系。It can be understood that the above method provides a calculation method that can be used when there is no common viewpoint between the two cameras in the different-side barrels. However, when there is a common viewpoint between the two cameras in the different-side barrels, the binocular positioning algorithm can still be used to calculate the pose relationship.
需要说明的是,上述示例仅以可穿戴设备200的两个镜筒分别设置两个相机为例进行说明,可以理解的是,根据实际需要的不同,镜筒内的相机数量可以更少或更多,但无论如何,基于本公开所提供的实施例的发明构思,仍然能够计算出相应的相机参数,在此不再赘述。It should be noted that the above example only takes the example of setting two cameras in the two lens barrels of the wearable device 200 respectively. It can be understood that the number of cameras in the lens barrel can be less or more according to actual needs. However, in any case, based on the inventive concept of the embodiment provided by the present disclosure, the corresponding camera parameters can still be calculated, which will not be repeated here.
在一个更为具体的实施例中,本公开实施例提供的相机参数的标定方法,可以包括采图、内参标定、外参标定、优化并保存参数的步骤,能够得到较好的相机参数,来供后续算法进行计算。In a more specific embodiment, the camera parameter calibration method provided by the embodiment of the present disclosure may include the steps of image acquisition, internal parameter calibration, external parameter calibration, and optimizing and saving parameters, so as to obtain better camera parameters for subsequent algorithm calculation.
从上述实施例可以看出,本公开实施例提供的相机参数的标定方法,对于镜筒内相机成像畸变不对称、无统一的投影中心的场景,提供一种可行的相机的内参和外参的标定方案。It can be seen from the above embodiments that the camera parameter calibration method provided by the embodiments of the present disclosure provides a feasible calibration scheme for the intrinsic and extrinsic parameters of the camera for scenes where the imaging distortion of the camera in the lens barrel is asymmetric and there is no unified projection center.
需要说明的是,本公开实施例的方法可以由单个设备执行,例如一台计算机或服务器等。本实施例的方法也可以应用于分布式场景下,由多台设备相互配合来完成。在这种分布式场景的情况下,这多台设备中的一台设备可以只执行本公开实施例的方法中的某一个或多个步骤,这多台设备相互之间会进行交互以完成所述的方法。It should be noted that the method of the embodiment of the present disclosure can be performed by a single device, such as a computer or a server. The method of the present embodiment can also be applied in a distributed scenario and completed by multiple devices cooperating with each other. In the case of such a distributed scenario, one of the multiple devices can only perform one or more steps in the method of the embodiment of the present disclosure, and the multiple devices will interact with each other to complete the described method.
需要说明的是,上述对本公开的一些实施例进行了描述。其它实施例在所附权利要求书的范围内。在一些情况下,在权利要求书中记载的动作或步骤可以按照不同于上述实施例中的顺序来执行并且仍然可以实现期望的结果。另外,在附图中描绘的过程不一定要求示出的特定顺序或者连续顺序才能实现期望的结果。在某些实施方式中,多任务处理和并行处理也是可以的或者可能是有利的。It should be noted that the above describes some embodiments of the present disclosure. Other embodiments are within the scope of the appended claims. In some cases, the actions or steps recorded in the claims can be performed in an order different from that in the above embodiments and still achieve the desired results. In addition, the processes depicted in the accompanying drawings do not necessarily require the specific order or continuous order shown to achieve the desired results. In some embodiments, multitasking and parallel processing are also possible or may be advantageous.
本公开实施例还提供了一种可穿戴设备。The embodiment of the present disclosure also provides a wearable device.
图5示出了本公开实施例所提供的示例性可穿戴设备500的示意图。FIG. 5 shows a schematic diagram of an exemplary wearable device 500 provided by an embodiment of the present disclosure.
如图5所示,与图2A和图2B类似地,可穿戴设备500包括镜筒202,所述镜筒202内设置显示模组204、相机208A、208B、光学组件206,所述相机208A、208B和所述光学组件206位于所述显示模组204的出光侧,所述相机208A、208B位于所述光学组件206和所述显示模组204之间。As shown in FIG. 5 , similar to FIG. 2A and FIG. 2B , the wearable device 500 includes a lens barrel 202 , in which a display module 204 , cameras 208A, 208B, and an optical component 206 are disposed. The cameras 208A, 208B and the optical component 206 are located on the light emitting side of the display module 204 , and the cameras 208A, 208B are located between the optical component 206 and the display module 204 .
在一些实施例中,所述相机208A、208B的拍摄方向朝向所述光学组件206,参考图2A所示。在另一些实施例中,参考图2B所示,所述显示模组204与所述相机208A、208B之间设置反射结构210A、210B,所述反射结构210A、210B的反射面朝向所述显示模组204的出光侧,所述相机208A、208B的拍摄方向朝向所述反射结构210A、210B。In some embodiments, the shooting direction of the cameras 208A and 208B is toward the optical assembly 206, as shown in FIG2A. In other embodiments, as shown in FIG2B, a reflective structure 210A and 210B is disposed between the display module 204 and the cameras 208A and 208B, the reflective surface of the reflective structure 210A and 210B is toward the light emitting side of the display module 204, and the shooting direction of the cameras 208A and 208B is toward the reflective structure 210A and 210B.
进一步地,如图5所示,所述可穿戴设备500还包括处理模块502,分别与所述相机208和所述显示模组204电耦接,且被配置为:获取采用前述方法300得到的第一参数集和第二参数集以及所述相机208A、208B所拍摄的图像;利用所述第一参数集和所述第二参数集以及所述图像,求解所述图像中的目标区域在空间中的位置。Further, as shown in Figure 5, the wearable device 500 also includes a processing module 502, which is electrically coupled to the camera 208 and the display module 204, respectively, and is configured to: obtain the first parameter set and the second parameter set obtained by the aforementioned method 300 and the images taken by the cameras 208A and 208B; use the first parameter set and the second parameter set and the image to solve the position of the target area in the image in space.
如前所述,第一参数集中包括以像素点与投射方向为组合的相机参数(该相机参数可以被认为是相机的内参),利用该相机参数实现相机采集的图像的坐标系到相机坐标系的转换,从而可以用于计算相机采集的图像中某些特征(例如,瞳孔)在三维空间(相机坐标系)中的位置信息。并且,由于该相机参数的计算过程中包含了对图像畸变的校正,从而利用该相机参数进行坐标系转换计算时,可以直接校正畸变,而无需再额外进行畸变校正。As mentioned above, the first parameter set includes camera parameters in combination of pixel points and projection directions (the camera parameters can be considered as the internal parameters of the camera), and the camera parameters are used to realize the conversion from the coordinate system of the image captured by the camera to the camera coordinate system, so that it can be used to calculate the position information of certain features (for example, pupil) in the image captured by the camera in the three-dimensional space (camera coordinate system). In addition, since the calculation process of the camera parameters includes the correction of image distortion, when the camera parameters are used for coordinate system conversion calculation, the distortion can be directly corrected without the need for additional distortion correction.
第二参数集中包括了表示不同相机之间的位姿关系的第二参数,可以根据该第二参数来将不同相机采集的图像信息进行统一(例如,统一到同一相机坐标系下)。The second parameter set includes a second parameter representing the position and posture relationship between different cameras, and the image information collected by different cameras can be unified (for example, unified into the same camera coordinate system) according to the second parameter.
本公开实施例提供的可穿戴设备,将相机设置在镜筒内部,受遮挡影响更小,可以提高瞳距估计或视线追踪算法的精度,进一步地,通过将相机设置为至少两个,可以提供更多的观测信息,进一步提高算法精度。本公开实施例提供的相机参数的标定方法及相关设备,对于镜筒内相机成像畸变不对称、无统一的投影中心的场景,提供一种可行的相机参数的标定方案。The wearable device provided by the embodiment of the present disclosure sets the camera inside the lens barrel, which is less affected by occlusion, and can improve the accuracy of pupil distance estimation or line of sight tracking algorithm. Furthermore, by setting the camera to at least two, more observation information can be provided, further improving the accuracy of the algorithm. The camera parameter calibration method and related equipment provided by the embodiment of the present disclosure provide a feasible camera parameter calibration solution for scenes where the camera imaging distortion in the lens barrel is asymmetric and there is no unified projection center.
本公开实施例还提供了一种计算机设备,用于实现上述的方法300。图6示出了本公开实施例所提供的示例性计算机设备600的硬件结构示意图。计算机设备600可以用于实现图1A的头戴式可穿戴设备104、图2A至图2D的可穿戴设备200,也可以用于实现图1A的外接设备112,也可以用于实现图1A的服务器114。在一些场景中,该计算机设备600也可以用于实现图1A的数据库服务器116。The embodiment of the present disclosure also provides a computer device for implementing the above-mentioned method 300. FIG6 shows a schematic diagram of the hardware structure of an exemplary computer device 600 provided in the embodiment of the present disclosure. The computer device 600 can be used to implement the head-mounted wearable device 104 of FIG1A, the wearable device 200 of FIG2A to FIG2D, and can also be used to implement the external device 112 of FIG1A, and can also be used to implement the server 114 of FIG1A. In some scenarios, the computer device 600 can also be used to implement the database server 116 of FIG1A.
如图6所示,计算机设备600可以包括:处理器602、存储器604、网络模块606、外围接口608和总线610。其中,处理器602、存储器604、网络模块606和外围接口608通过总线610实现彼此之间在计算机设备600的内部的通信连接。As shown in Fig. 6, computer device 600 may include: processor 602, memory 604, network module 606, peripheral interface 608 and bus 610. Processor 602, memory 604, network module 606 and peripheral interface 608 are connected to each other in communication within computer device 600 via bus 610.
处理器602可以是中央处理器(Central Processing Unit,CPU)、图像处理器、神经网络处理器(NPU)、微控制器(MCU)、可编程逻辑器件、数字信号处理器(DSP)、应用专用集成电路(Application Specific Integrated Circuit,ASIC)、或者一个或多个集成电路。处理器602可以用于执行与本公开描述的技术相关的功能。在一些实施例中,处理器602还可以包括集成为单一逻辑组件的多个处理器。例如,如图6所示,处理器602可以包括多个处理器602a、602b和602c。Processor 602 may be a central processing unit (CPU), an image processor, a neural network processor (NPU), a microcontroller (MCU), a programmable logic device, a digital signal processor (DSP), an application specific integrated circuit (ASIC), or one or more integrated circuits. Processor 602 may be used to perform functions related to the technology described in the present disclosure. In some embodiments, processor 602 may also include multiple processors integrated into a single logical component. For example, as shown in FIG. 6 , processor 602 may include multiple processors 602a, 602b, and 602c.
存储器604可以配置为存储数据(例如,指令、计算机代码等)。如图6所示,存储器604存储的数据可以包括程序指令(例如,用于实现本公开实施例的方法300或500的程序指令)以及要处理的数据(例如,存储器可以存储其他模块的配置文件等)。处理器602也可以访问存储器604存储的程序指令和数据,并且执行程序指令以对要处理的数据进行操作。存储器604可以包括易失性存储装置或非易失性存储装置。在一些实施例中,存储器604可以包括随机访问存储器(RAM)、只读存储器(ROM)、光盘、磁盘、硬盘、固态硬盘(SSD)、闪存、存储棒等。The memory 604 may be configured to store data (e.g., instructions, computer codes, etc.). As shown in Figure 6, the data stored in the memory 604 may include program instructions (e.g., program instructions for implementing the method 300 or 500 of the embodiment of the present disclosure) and data to be processed (e.g., the memory may store configuration files of other modules, etc.). The processor 602 may also access the program instructions and data stored in the memory 604, and execute the program instructions to operate on the data to be processed. The memory 604 may include a volatile storage device or a non-volatile storage device. In some embodiments, the memory 604 may include a random access memory (RAM), a read-only memory (ROM), an optical disk, a magnetic disk, a hard disk, a solid-state drive (SSD), a flash memory, a memory stick, etc.
网络接口606可以配置为经由网络向计算机设备600提供与其他外部设备的通信。该网络可以是能够传输和接收数据的任何有线或无线的网络。例如,该网络可以是有线网络、本地无线网络(例如,蓝牙、WiFi、近场通信(NFC)等)、蜂窝网络、因特网、或上述的组合。可以理解的是,网络的类型不限于上述具体示例。The network interface 606 can be configured to provide the computer device 600 with communication with other external devices via a network. The network can be any wired or wireless network capable of transmitting and receiving data. For example, the network can be a wired network, a local wireless network (e.g., Bluetooth, WiFi, near field communication (NFC)), a cellular network, the Internet, or a combination thereof. It is understood that the type of network is not limited to the above specific examples.
外围接口608可以配置为将计算机设备600与一个或多个外围装置连接,以实现信息输入及输出。例如,外围装置可以包括键盘、鼠标、触摸板、触摸屏、麦克风、各类传感器等输入设备以及显示器、扬声器、振动器、指示灯等输出设备。The peripheral interface 608 can be configured to connect the computer device 600 to one or more peripheral devices to achieve information input and output. For example, the peripheral devices can include input devices such as a keyboard, a mouse, a touch pad, a touch screen, a microphone, and various sensors, and output devices such as a display, a speaker, a vibrator, and an indicator light.
总线610可以被配置为在计算机设备600的各个组件(例如处理器602、存储器604、网络接口606和外围接口608)之间传输信息,诸如内部总线(例如,处理器-存储器总线)、外部总线(USB端口、PCI-E总线)等。The bus 610 may be configured to transmit information between various components of the computer device 600 (e.g., the processor 602, the memory 604, the network interface 606, and the peripheral interface 608), such as an internal bus (e.g., a processor-memory bus), an external bus (USB port, PCI-E bus), etc.
需要说明的是,尽管上述计算机设备600的架构仅示出了处理器602、存储器604、网络接口606、外围接口608和总线610,但是在具体实施过程中,该计算机设备600的架构还可以包括实现正常运行所必需的其他组件。此外,本领域的技术人员可以理解的是,上述计算机设备600的架构中也可以仅包含实现本公开实施例方案所必需的组件,而不必包含图中所示的全部组件。It should be noted that, although the architecture of the above-mentioned computer device 600 only shows the processor 602, the memory 604, the network interface 606, the peripheral interface 608 and the bus 610, in the specific implementation process, the architecture of the computer device 600 may also include other components necessary for normal operation. In addition, it can be understood by those skilled in the art that the architecture of the above-mentioned computer device 600 may also only include the components necessary for implementing the embodiments of the present disclosure, and does not necessarily include all the components shown in the figure.
本公开实施例还提供了一种相机参数的标定装置。图7示出了本公开实施例所提供的示例性装置700的示意图。如图7所示,该装置700可以用于实现方法300,并可以进一步包括以下模块。The embodiment of the present disclosure also provides a camera parameter calibration device. Figure 7 shows a schematic diagram of an exemplary device 700 provided by the embodiment of the present disclosure. As shown in Figure 7, the device 700 can be used to implement the method 300, and can further include the following modules.
接收模块702,被配置为:接收由至少两个相机拍摄标定参考对象而得到的多张图像;所述可穿戴设备还包括双目显示模组,所述镜筒设置在所述双目显示模组的出光侧,所述镜筒内设置光学组件,所述至少两个相机位于所述双目显示模组和所述光学组件之间;第一确定模块704,被配置为:确定所述多张图像的像素点与所述标定参考对象的投射关系;第二确定模块706,被配置为:根据所述投射关系,确定所述至少两个相机的第一参数集和第二参数集。The receiving module 702 is configured to: receive multiple images obtained by photographing a calibrated reference object by at least two cameras; the wearable device also includes a binocular display module, the lens barrel is arranged on the light-emitting side of the binocular display module, an optical component is arranged in the lens barrel, and the at least two cameras are located between the binocular display module and the optical component; the first determination module 704 is configured to: determine the projection relationship between the pixel points of the multiple images and the calibrated reference object; the second determination module 706 is configured to: determine the first parameter set and the second parameter set of the at least two cameras according to the projection relationship.
其中,所述第一参数集包括多个第一参数,所述第一参数用于表征所述待标定相机所拍摄的图像中的目标像素点和所述目标像素点对应的投射方向,所述第二参数集包括至少一组第二参数,所述至少一组第二参数用于指示所述至少两个相机之间的位姿关系。Among them, the first parameter set includes multiple first parameters, the first parameters are used to characterize the target pixel points in the image taken by the camera to be calibrated and the projection direction corresponding to the target pixel points, and the second parameter set includes at least one group of second parameters, and the at least one group of second parameters is used to indicate the posture relationship between the at least two cameras.
在一些实施例中,所述至少两个相机包括用于采集第一眼的人眼图像的第一相机和第二相机以及用于采集第二眼的人眼图像的第三相机和第四相机,所述第一相机和所述第二相机设置在所述可穿戴设备的第一镜筒中,所述第三相机和所述第四相机设置在所述可穿戴设备的第二镜筒中;所述第二参数包括所述第一相机和所述第二相机之间的位姿关系以及所述第三相机和所述第四相机之间的位姿关系,所述第一相机和所述第二相机之间的位姿关系以及所述第三相机和所述第四相机之间的位姿关系是基于双目标定算法根据所述投射关系计算得到的。In some embodiments, the at least two cameras include a first camera and a second camera for capturing a first eye image of a human eye and a third camera and a fourth camera for capturing a second eye image of a human eye, the first camera and the second camera are arranged in a first lens barrel of the wearable device, and the third camera and the fourth camera are arranged in a second lens barrel of the wearable device; the second parameter includes a posture relationship between the first camera and the second camera and a posture relationship between the third camera and the fourth camera, and the posture relationship between the first camera and the second camera and the posture relationship between the third camera and the fourth camera are calculated based on a binocular positioning algorithm according to the projection relationship.
在一些实施例中,所述第二参数还包括所述第一相机和所述第三相机之间的位姿关系、所述第一相机和所述第四相机之间的位姿关系、所述第二相机和所述第三相机之间的位姿关系、所述第二相机和所述第四相机之间的位姿关系。In some embodiments, the second parameters also include a posture relationship between the first camera and the third camera, a posture relationship between the first camera and the fourth camera, a posture relationship between the second camera and the third camera, and a posture relationship between the second camera and the fourth camera.
在一些实施例中,第二确定模块706,被配置为:根据所述多张图像和所述投射关系,确定所述第一相机与所述第四相机之间的位姿关系;根据所述第三相机和所述第四相机之间的位姿关系以及所述第一相机与所述第四相机之间的位姿关系,确定所述第一相机与所述第三相机之间的位姿关系;根据所述第一相机和所述第二相机之间的位姿关系以及所述第一相机与所述第三相机之间的位姿关系,确定所述第二相机与所述第三相机之间的位姿关系;根据所述第三相机和所述第四相机之间的位姿关系以及所述第二相机与所述第三相机之间的位姿关系,确定所述第二相机与所述第四相机之间的位姿关系。In some embodiments, the second determination module 706 is configured to: determine the posture relationship between the first camera and the fourth camera based on the multiple images and the projection relationship; determine the posture relationship between the first camera and the third camera based on the posture relationship between the third camera and the fourth camera and the posture relationship between the first camera and the fourth camera; determine the posture relationship between the second camera and the third camera based on the posture relationship between the first camera and the second camera and the posture relationship between the first camera and the third camera; determine the posture relationship between the second camera and the fourth camera based on the posture relationship between the third camera and the fourth camera and the posture relationship between the second camera and the third camera.
在一些实施例中,第一确定模块704,被配置为:从所述多张图像中选取第一数量的目标图像;在每张所述目标图像中选取至少一个局部区域,并建立每个所述局部区域与所述标定参考对象的单应性变换矩阵;根据所述单应性变换矩阵确定所述局部区域包含的像素点与所述标定参考对象的投射关系;将所述第一数量的目标图像转换到参考坐标系下,并根据所述局部区域对应的已建立投射关系的区域,确定所述多张图像的像素点与所述标定参考对象的投射关系。In some embodiments, the first determination module 704 is configured to: select a first number of target images from the multiple images; select at least one local area in each of the target images, and establish a homography transformation matrix between each of the local areas and the calibrated reference object; determine the projection relationship between the pixel points contained in the local area and the calibrated reference object according to the homography transformation matrix; convert the first number of target images to a reference coordinate system, and determine the projection relationship between the pixel points of the multiple images and the calibrated reference object according to the area with the established projection relationship corresponding to the local area.
在一些实施例中,第二确定模块706,被配置为:根据所述投射关系,确定所述至少两个相机的初始第一参数集,所述初始第一参数集包括与所述图像的多个像素点一一对应的多个第一参数;采用样条曲面对所述初始第一参数集进行拟合,得到拟合样条曲面;基于光束法平差算法对所述拟合样条曲面的控制点进行优化,得到所述至少两个相机的第一参数集。In some embodiments, the second determination module 706 is configured to: determine an initial first parameter set of the at least two cameras according to the projection relationship, the initial first parameter set including multiple first parameters corresponding one-to-one to multiple pixel points of the image; fit the initial first parameter set using a spline surface to obtain a fitted spline surface; and optimize the control points of the fitted spline surface based on a bundle adjustment algorithm to obtain the first parameter set of the at least two cameras.
为了描述的方便,描述以上装置时以功能分为各种模块分别描述。当然,在实施本公开时可以把各模块的功能在同一个或多个软件和/或硬件中实现。For the convenience of description, the above device is described by dividing it into various modules according to its functions. Of course, when implementing the present disclosure, the functions of each module can be implemented in the same or multiple software and/or hardware.
上述实施例的装置用于实现前述任一实施例中相应的方法300,并且具有相应的方法实施例的有益效果,在此不再赘述。The device of the above embodiment is used to implement the corresponding method 300 in any of the above embodiments, and has the beneficial effects of the corresponding method embodiment, which will not be described in detail here.
基于同一发明构思,与上述任意实施例方法相对应的,本公开还提供了一种非暂态计算机可读存储介质,所述非暂态计算机可读存储介质存储计算机指令,所述计算机指令用于使所述计算机执行如上任一实施例所述的方法300。Based on the same inventive concept, corresponding to any of the above-mentioned embodiments, the present disclosure also provides a non-transitory computer-readable storage medium, wherein the non-transitory computer-readable storage medium stores computer instructions, and the computer instructions are used to enable the computer to execute method 300 described in any of the above embodiments.
本实施例的计算机可读介质包括永久性和非永久性、可移动和非可移动媒体可以由任何方法或技术来实现信息存储。信息可以是计算机可读指令、数据结构、程序的模块或其他数据。计算机的存储介质的例子包括,但不限于相变内存(PRAM)、静态随机存取存储器(SRAM)、动态随机存取存储器(DRAM)、其他类型的随机存取存储器(RAM)、只读存储器(ROM)、电可擦除可编程只读存储器(EEPROM)、快闪记忆体或其他内存技术、只读光盘只读存储器(CD-ROM)、数字多功能光盘(DVD)或其他光学存储、磁盒式磁带,磁带磁磁盘存储或其他磁性存储设备或任何其他非传输介质,可用于存储可以被计算设备访问的信息。The computer-readable medium of this embodiment includes permanent and non-permanent, removable and non-removable media, and information storage can be implemented by any method or technology. Information can be computer-readable instructions, data structures, modules of programs, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), static random access memory (SRAM), dynamic random access memory (DRAM), other types of random access memory (RAM), read-only memory (ROM), electrically erasable programmable read-only memory (EEPROM), flash memory or other memory technology, read-only compact disk read-only memory (CD-ROM), digital versatile disk (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices or any other non-transmission media that can be used to store information that can be accessed by a computing device.
上述实施例的存储介质存储的计算机指令用于使所述计算机执行如上任一实施例所述的方法300,并且具有相应的方法实施例的有益效果,在此不再赘述。The computer instructions stored in the storage medium of the above embodiment are used to enable the computer to execute the method 300 described in any of the above embodiments, and have the beneficial effects of the corresponding method embodiments, which will not be repeated here.
基于同一发明构思,与上述任意实施例方法300相对应的,本公开还提供了一种计算机程序产品,其包括计算机程序指令。在一些实施例中,所述计算机程序指令可以由计算机的一个或多个处理器执行以使得所述计算机和/或所述处理器执行所述的方法300。对应于方法300各实施例中各步骤对应的执行主体,执行相应步骤的处理器可以是属于相应执行主体的。Based on the same inventive concept, corresponding to the method 300 of any of the above embodiments, the present disclosure further provides a computer program product, which includes computer program instructions. In some embodiments, the computer program instructions can be executed by one or more processors of a computer so that the computer and/or the processor execute the method 300. Corresponding to the execution subject corresponding to each step in each embodiment of the method 300, the processor that executes the corresponding step can belong to the corresponding execution subject.
上述实施例的计算机程序产品用于使所述计算机和/或所述处理器执行如上任一实施例所述的方法300,并且具有相应的方法实施例的有益效果,在此不再赘述。The computer program product of the above embodiment is used to enable the computer and/or the processor to execute the method 300 described in any of the above embodiments, and has the beneficial effects of the corresponding method embodiments, which will not be described in detail here.
所属领域的普通技术人员应当理解:以上任何实施例的讨论仅为示例性的,并非旨在暗示本公开的范围(包括权利要求)被限于这些例子;在本公开的思路下,以上实施例或者不同实施例中的技术特征之间也可以进行组合,步骤可以以任意顺序实现,并存在如上所述的本公开实施例的不同方面的许多其它变化,为了简明它们没有在细节中提供。Those skilled in the art should understand that the discussion of any of the above embodiments is merely illustrative and is not intended to imply that the scope of the present disclosure (including the claims) is limited to these examples. Based on the concept of the present disclosure, the technical features in the above embodiments or different embodiments may be combined, the steps may be implemented in any order, and there are many other variations of the different aspects of the embodiments of the present disclosure as described above, which are not provided in detail for the sake of simplicity.
另外,为简化说明和讨论,并且为了不会使本公开实施例难以理解,在所提供的附图中可以示出或可以不示出与集成电路(IC)芯片和其它部件的公知的电源/接地连接。此外,可以以框图的形式示出装置,以便避免使本公开实施例难以理解,并且这也考虑了以下事实,即关于这些框图装置的实施方式的细节是高度取决于将要实施本公开实施例的平台的(即,这些细节应当完全处于本领域技术人员的理解范围内)。在阐述了具体细节(例如,电路)以描述本公开的示例性实施例的情况下,对本领域技术人员来说显而易见的是,可以在没有这些具体细节的情况下或者这些具体细节有变化的情况下实施本公开实施例。因此,这些描述应被认为是说明性的而不是限制性的。In addition, to simplify the description and discussion, and in order not to make the embodiments of the present disclosure difficult to understand, the known power/ground connections to the integrated circuit (IC) chips and other components may or may not be shown in the provided figures. In addition, the device can be shown in the form of a block diagram to avoid making the embodiments of the present disclosure difficult to understand, and this also takes into account the fact that the details of the implementation of these block diagram devices are highly dependent on the platform on which the embodiments of the present disclosure will be implemented (that is, these details should be fully within the scope of understanding of those skilled in the art). Where specific details (e.g., circuits) are set forth to describe exemplary embodiments of the present disclosure, it is apparent to those skilled in the art that the embodiments of the present disclosure can be implemented without these specific details or with changes in these specific details. Therefore, these descriptions should be considered illustrative rather than restrictive.
尽管已经结合了本公开的具体实施例对本公开进行了描述,但是根据前面的描述,这些实施例的很多替换、修改和变型对本领域普通技术人员来说将是显而易见的。例如,其它存储器架构(例如,动态RAM(DRAM))可以使用所讨论的实施例。Although the present disclosure has been described in conjunction with specific embodiments of the present disclosure, many replacements, modifications and variations of these embodiments will be apparent to those skilled in the art from the foregoing description. For example, other memory architectures (e.g., dynamic RAM (DRAM)) may use the embodiments discussed.
本公开实施例旨在涵盖落入所附权利要求的宽泛范围之内的所有这样的替换、修改和变型。因此,凡在本公开实施例的精神和原则之内,所做的任何省略、修改、等同替换、改进等,均应包含在本公开的保护范围之内。The embodiments of the present disclosure are intended to cover all such substitutions, modifications and variations that fall within the broad scope of the appended claims. Therefore, any omissions, modifications, equivalent substitutions, improvements, etc. made within the spirit and principles of the embodiments of the present disclosure should be included in the scope of protection of the present disclosure.
Claims (15)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| CN202311484874.4A CN119963652B (en) | 2023-11-08 | 2023-11-08 | Calibration method of camera parameters and related equipment |
| CN202311484874.4 | 2023-11-08 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025098400A1 true WO2025098400A1 (en) | 2025-05-15 |
Family
ID=95586553
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/CN2024/130269 Pending WO2025098400A1 (en) | 2023-11-08 | 2024-11-06 | Camera parameter calibration method and related device |
Country Status (2)
| Country | Link |
|---|---|
| CN (1) | CN119963652B (en) |
| WO (1) | WO2025098400A1 (en) |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150124151A1 (en) * | 2013-11-07 | 2015-05-07 | Pelican Imaging Corporation | Array Camera Modules Incorporating Independently Aligned Lens Stacks |
| CN109118545A (en) * | 2018-07-26 | 2019-01-01 | 深圳市易尚展示股份有限公司 | 3-D imaging system scaling method and system based on rotary shaft and binocular camera |
| CN112308925A (en) * | 2019-08-02 | 2021-02-02 | 上海肇观电子科技有限公司 | Binocular calibration method and device of wearable device and storage medium |
| CN113902796A (en) * | 2021-08-31 | 2022-01-07 | 杭州易现先进科技有限公司 | A method and system for calibrating an image-based head-mounted display device |
| CN115187658A (en) * | 2022-08-29 | 2022-10-14 | 合肥埃科光电科技股份有限公司 | Multi-camera visual large target positioning method, system and equipment |
Family Cites Families (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN105872527A (en) * | 2015-01-21 | 2016-08-17 | 成都理想境界科技有限公司 | Binocular AR (Augmented Reality) head-mounted display device and information display method thereof |
| SG11202112758VA (en) * | 2019-07-31 | 2021-12-30 | Xenon Vr Inc | Ophthalmologic testing systems and methods |
| CN112346558B (en) * | 2019-08-06 | 2024-08-02 | 苹果公司 | Eye tracking system |
| CN116338941B (en) * | 2021-12-22 | 2025-12-05 | 华为技术有限公司 | Eye-tracking device, display device and storage medium |
-
2023
- 2023-11-08 CN CN202311484874.4A patent/CN119963652B/en active Active
-
2024
- 2024-11-06 WO PCT/CN2024/130269 patent/WO2025098400A1/en active Pending
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150124151A1 (en) * | 2013-11-07 | 2015-05-07 | Pelican Imaging Corporation | Array Camera Modules Incorporating Independently Aligned Lens Stacks |
| CN109118545A (en) * | 2018-07-26 | 2019-01-01 | 深圳市易尚展示股份有限公司 | 3-D imaging system scaling method and system based on rotary shaft and binocular camera |
| CN112308925A (en) * | 2019-08-02 | 2021-02-02 | 上海肇观电子科技有限公司 | Binocular calibration method and device of wearable device and storage medium |
| CN113902796A (en) * | 2021-08-31 | 2022-01-07 | 杭州易现先进科技有限公司 | A method and system for calibrating an image-based head-mounted display device |
| CN115187658A (en) * | 2022-08-29 | 2022-10-14 | 合肥埃科光电科技股份有限公司 | Multi-camera visual large target positioning method, system and equipment |
Also Published As
| Publication number | Publication date |
|---|---|
| CN119963652B (en) | 2025-11-04 |
| CN119963652A (en) | 2025-05-09 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12348730B2 (en) | Reprojecting holographic video to enhance streaming bandwidth/quality | |
| CN111857329B (en) | Gaze calculation method, device and equipment | |
| US11080874B1 (en) | Apparatuses, systems, and methods for high-sensitivity active illumination imaging | |
| US20170134713A1 (en) | Image calibrating, stitching and depth rebuilding method of a panoramic fish-eye camera and a system thereof | |
| JP2018511098A (en) | Mixed reality system | |
| CN108307675A (en) | More baseline camera array system architectures of depth enhancing in being applied for VR/AR | |
| WO2017020150A1 (en) | Image processing method, device and camera | |
| WO2023072030A1 (en) | Automatic focusing method and apparatus for lens, and electronic device and computer-readable storage medium | |
| WO2019229887A1 (en) | Camera apparatus | |
| CN108401551B (en) | Twin-lens low-light stereoscopic full views imaging device and its ultra-large vision field distance measuring method | |
| CN107560637A (en) | Wear display device calibration result verification method and wear display device | |
| TWM594322U (en) | Camera configuration system with omnidirectional stereo vision | |
| JP5805013B2 (en) | Captured image display device, captured image display method, and program | |
| WO2021093804A1 (en) | Omnidirectional stereo vision camera configuration system and camera configuration method | |
| WO2025098400A1 (en) | Camera parameter calibration method and related device | |
| WO2025113508A1 (en) | Method for determining position of target object, and related device | |
| CN119234196A (en) | Gesture detection method and system with hand shape calibration | |
| US12423916B1 (en) | Eye-tracking assisted passthrough rendering | |
| CN119963653A (en) | Camera parameter calibration method and related equipment | |
| CN114020150A (en) | Image display method, image display device, electronic apparatus, and medium | |
| CN120167884B (en) | Methods and related equipment for determining corneal center and pupillary distance | |
| TWI885862B (en) | See-through display method and see-through display system | |
| CN114093020A (en) | Motion capture method, motion capture device, electronic device and storage medium | |
| US20250322615A1 (en) | See-through display method and see-through display system | |
| CN114792332B (en) | Image registration method and device |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24888011 Country of ref document: EP Kind code of ref document: A1 |