[go: up one dir, main page]

US20240265579A1 - Electronic device, parameter calibration method, and non-transitory computer readable storage medium - Google Patents

Electronic device, parameter calibration method, and non-transitory computer readable storage medium Download PDF

Info

Publication number
US20240265579A1
US20240265579A1 US18/432,065 US202418432065A US2024265579A1 US 20240265579 A1 US20240265579 A1 US 20240265579A1 US 202418432065 A US202418432065 A US 202418432065A US 2024265579 A1 US2024265579 A1 US 2024265579A1
Authority
US
United States
Prior art keywords
camera
pose
cameras
image
electronic device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
US18/432,065
Inventor
Chunkai HUANG
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
HTC Corp
Original Assignee
HTC Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Assigned to HTC CORPORATION reassignment HTC CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: HUANG, CHUNKAI
Application filed by HTC Corp filed Critical HTC Corp
Priority to US18/432,065 priority Critical patent/US20240265579A1/en
Publication of US20240265579A1 publication Critical patent/US20240265579A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N17/00Diagnosis, testing or measuring for television systems or their details
    • H04N17/002Diagnosis, testing or measuring for television systems or their details for television cameras
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • G06T7/85Stereo camera calibration
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/521Depth or shape recovery from laser ranging, e.g. using interferometry; from the projection of structured light
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/70Determining position or orientation of objects or cameras
    • G06T7/73Determining position or orientation of objects or cameras using feature-based methods
    • G06T7/74Determining position or orientation of objects or cameras using feature-based methods involving reference images or patches
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/80Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/243Image signal generators using stereoscopic image cameras using three or more 2D image sensors
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/246Calibration of cameras
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/204Image signal generators using stereoscopic image cameras
    • H04N13/254Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/261Image signal generators with monoscopic-to-stereoscopic image conversion
    • H04N13/268Image signal generators with monoscopic-to-stereoscopic image conversion based on depth image-based rendering [DIBR]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/20Image signal generators
    • H04N13/271Image signal generators wherein the generated image signals comprise depth maps or disparity maps
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30244Camera pose

Definitions

  • the present application relates to an electronic device, a parameter calibration method, and a non-transitory computer readable storage medium. More particularly, the present application relates to an electronic device, a parameter calibration method, and a non-transitory computer readable storage medium with a SLAM module.
  • Self-tracking devices such as VR headsets or trackers, access a SLAM module to determine their positions in the real space with the images captured by the cameras within the self-tracking devices.
  • changes in the self-tracking devices such as damage or breakage during deliver or usage, can affect the relative position and the relation rotation between the cameras of the self-tracking devices, and the pre-set extrinsic parameters, including the pre-set relative position parameter and the pre-set relative rotation parameter, between the cameras of the self-tracking devices may no longer be used, in which the performance of the SLAM module may be decreased.
  • the self-tracking devices may become unable to track themselves with the SLAM module even if the cameras themselves are functioning properly.
  • Several methods are proposed to recalculate the extrinsic parameters of the cameras of the self-tracking devices, such as recalculating the extrinsic parameters with a checkerboard or a Deltille grid.
  • the disclosure provides an electronic device.
  • the electronic device includes a memory, several cameras, and a processor.
  • the memory is configured to store a SLAM module.
  • the several cameras are configured to capture several images of a real space.
  • the processor is coupled to the camera and the memory.
  • the processor is configured to: process the SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to several images; and perform a calibration process.
  • the operation of performing the calibration process includes: calculating several poses of several cameras within the environment coordinate system according to several light spots within each of several images, in which several light spots are generated by a structured light generation device; and calibrating several extrinsic parameters between several cameras according to several poses.
  • the disclosure provides a parameter calibration method suitable for an electronic device.
  • the parameter calibration method includes the following operations: capturing several images of a real space by several cameras; processing a SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to several images by a processor; and performing a calibration process by the processor.
  • the operation of performing the calibration process includes the following operations: calculating several poses of several cameras within the environment coordinate system according to several light spots within each of several images, wherein several light spots are generated by a structured light generation device; and calibrating several extrinsic parameters between several cameras according to several poses.
  • the disclosure provides a non-transitory computer readable storage medium with a computer program to execute aforesaid parameter calibration method.
  • FIG. 1 is a schematic block diagram illustrating an electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 2 is a schematic block diagram illustrating another electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 3 is a schematic diagram illustrating a user operating the electronic device as illustrated in FIG. 1 in accordance with some embodiments of the present disclosure.
  • FIG. 4 is a schematic diagram illustrating an electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 5 is a flowchart illustrating a parameter calibration method in accordance with some embodiments of the present disclosure.
  • FIG. 6 is a flow chart illustrating an operation of FIG. 5 in accordance with some embodiments of the present disclosure.
  • FIG. 7 is a flow chart illustrating an operation of FIG. 6 in accordance with some embodiments of the present disclosure.
  • FIG. 8 A is a schematic diagram illustrating an image captured by an camera as illustrated in FIG. 1 and FIG. 4 .
  • FIG. 8 B is a schematic diagram illustrating an image captured by another camera as illustrated in FIG. 1 and FIG. 4 .
  • FIG. 9 is a flow chart illustrating an operation of FIG. 6 in accordance with some embodiments of the present disclosure.
  • FIG. 10 is a flow chart illustrating an operation of FIG. 5 in accordance with some embodiments of the present disclosure.
  • FIG. 1 is a schematic block diagram illustrating an electronic device 100 in accordance with some embodiments of the present disclosure.
  • the electronic device 100 includes a memory 110 , a processor 130 , several cameras 150 A to 150 C, and a structured light generation device 170 .
  • the memory 110 , the cameras 150 A to 150 C, and the structured light generation device 170 couple to the processor 130 .
  • FIG. 2 is a schematic block diagram illustrating another electronic device 200 in accordance with some embodiments of the present disclosure.
  • the electronic device 200 includes a memory 210 , a processor 230 , and several cameras 250 A to 250 C.
  • the memory 210 and the cameras 250 A to 250 C couple to the processor 230 .
  • the electronic device 200 couple to a structured light generation device 900 . That is, in some embodiments, the electronic device 200 and the structured light generation device 900 are separate devices.
  • FIG. 1 and FIG. 2 three cameras are illustrated.
  • the electronic device 100 and the electronic device 200 are for illustrative purposes only, and the embodiments of the present disclosure are not limited thereto.
  • the electronic device 100 and the electronic device 200 may include two cameras, or more than three cameras. It is noted that, the embodiments shown in FIG. 1 and FIG. 2 are merely an example and not meant to limit the present disclosure.
  • One or more programs are stored in the memory 110 and the memory 210 and are configured to be executed by the processor 130 or the processor 230 , in order to perform a parameter calibration method.
  • the electronic device 100 and the electronic device 200 may be an HMD (head-mounted display) device, a tracking device, or any other device with self-tracking function.
  • the HMD device may be wear on the head of a user.
  • the memory 110 and the memory 210 store a SLAM (Simultaneous localization and mapping) module.
  • the electronic device 100 and the electronic device 200 may be configured to process the SLAM module.
  • the SLAM module includes functions such as image capturing, features extracting from the image, and localizing according to the extracted features.
  • the SLAM module include a SLAM algorithm, in which the processor 130 access and process the SLAM module so as to localize the electronic device 100 according to the images captured by the cameras 150 A to 150 C.
  • the processor 230 access and process the SLAM module so as to localize the electronic device 200 according to the images captured by the cameras 250 A to 250 C. The details of the SLAM system will not be described herein.
  • the electronic device 100 may be applied in a virtual reality (VR)/mixed reality (MR)/augmented reality (AR) system.
  • the electronic device 100 may be realized by, a standalone head mounted display device (HMD) or VIVE HMD.
  • the processor 130 and 230 can be realized by, for example, one or more processing circuits, such as central processing circuits and/or micro processing circuits, but are not limited in this regard.
  • the memory 110 and 210 include one or more memory devices, each of which includes, or a plurality of which collectively include a computer readable storage medium.
  • the non-transitory computer readable storage medium may include a read-only memory (ROM), a flash memory, a floppy disk, a hard disk, an optical disc, a flash disk, a flash drive, a tape, a database accessible from a network, and/or any storage medium with the same functionality that can be contemplated by persons of ordinary skill in the art to which this disclosure pertains.
  • the cameras 150 A to 150 C and the cameras 250 A to 250 C are configured to capture one or more images of the real space that the electronic device 100 and 200 are operated.
  • the cameras 150 A to 150 C and the cameras 250 A to 250 C may be realized by camera circuit devices or any other camera circuits with image capture functions.
  • the electronic device 100 and 200 include other circuits such as a display circuit and an I/O circuit.
  • the display circuit covers a field of view of the user and shows a virtual image at the field of view of the user.
  • the following take the electronic device 100 as illustrated in FIG. 1 for illustrative purpose. It should be noted that the operation of the electronic device 200 as illustrated in FIG. 2 is similar to the electronic device 100 as illustrated in FIG. 1 .
  • FIG. 3 is a schematic diagram illustrating a user U operating the electronic device 100 as illustrated in FIG. 1 in accordance with some embodiments of the present disclosure.
  • the user U is wearing the electronic device 100 as illustrated in FIG. 1 on the head of the user U.
  • the cameras 150 A to 150 C capture several frames of images of the real space R.
  • the processor 130 process the SLAM module to establish a mixed reality environment coordinate system M in correspondence to the real space R according to several space feature points of the images captured by the cameras 150 A to 150 C.
  • the processor 130 obtains a device pose of the electronic device 100 within the mixed reality environment coordinate system M according to the feature points within the images.
  • the processor 130 tracks the device pose of the electronic device 100 within the mixed reality environment coordinate system M.
  • the mixed reality environment coordinate system M could be an augmented reality environment coordinate system or an extended reality environment coordinate system.
  • the following takes the mixed reality environment coordinate system M for examples for illustrative purposes; however, the embodiments of the present disclosure are not limited thereto.
  • the device pose of the electronic device 100 includes a position and a rotation angle.
  • the intrinsic parameter and the extrinsic parameter of each of the cameras 150 A to 150 C are considered.
  • the extrinsic parameters represent a rigid transformation from 3D world coordinate system to the 3D camera's coordinate system.
  • the intrinsic parameters represent a projective transformation from the 3D camera's coordinates into the 2D image coordinates.
  • the extrinsic parameters of the cameras 150 A to 150 C include the difference between the poses of the cameras.
  • FIG. 4 is a schematic diagram illustrating an electronic device 100 in accordance with some embodiments of the present disclosure.
  • camera 150 A and camera 150 B of the electronic device 100 are taken as an example for illustration.
  • the positions of the camera 150 A and the camera 150 B and the rotation angles of the camera 150 A and the camera 150 B relative to the electronic device 100 is preset, and an extrinsic parameter between the camera 150 A and the camera 150 B is preset within the SLAM module.
  • the extrinsic parameters between each two of the cameras are preset within the SLAM module.
  • the extrinsic parameters between each two of the cameras are preset within the SLAM module are considered.
  • the positions of the camera 150 A and the camera 150 B and the rotation angles of the camera 150 A and the camera 150 B relative to the electronic device 100 may be changed, and the SLAM module may no longer be working properly with the images captured by the cameras 150 A and 150 B and the preset extrinsic parameter between the cameras 150 A and 150 B. Therefore, a method for calibrating the extrinsic parameters between the cameras of the electronic device 100 is in need.
  • the extrinsic parameters are stored in the memory 110 for the processor 130 to access and operate with the SLAM module.
  • FIG. 5 is a flowchart illustrating a parameter calibration method 500 in accordance with some embodiments of the present disclosure.
  • the parameter calibration method 500 can be applied to a device having a structured that is the same as or similar to the structured of the electronic device 100 shown in FIG. 1 or the electronic device 200 shown in FIG. 2 .
  • the embodiments shown in FIG. 1 will be used as an example to describe the parameter calibration method 500 in accordance with some embodiments of the present disclosure.
  • the present disclosure is not limited to application to the embodiments shown in FIG. 1 .
  • the parameter calibration method 500 includes operations S 510 to S 540 .
  • the SLAM module is processed to track the device pose of the electronic device within the mixed reality environment coordinate system according to several images.
  • the processor 130 tracks the device pose of the electronic device 100 within the mixed reality environment coordinate system M according to several space feature points within the images captured by the cameras 130 A to 130 C.
  • operation S 520 it is determined whether the SLAM module is working properly with the extrinsic parameters. In some embodiments, when the SLAM module is working properly with the extrinsic parameters stored in the memory 110 , operation S 530 is performed. On the other hand, when the SLAM module is not working properly with the extrinsic parameters stored in the memory 110 , operation S 540 is performed.
  • the processor 130 of the electronic device 100 determines the pose of the electronic device 100 every period of time. When determining the pose of the electronic device 100 , the processor 130 refers to the previous pose of the electronic device 100 determined at a previous period of time. In some embodiments, the processor 130 further refers to the positions of the space feature points determined previously when determining the pose of the electronic device 100 .
  • the processor 130 When the processor 130 is unable to calculate the pose of the electronic device 100 in reference to the space feature points determined previously and/or the pose of the electronic device 100 determined at a previous period of time, it is determined that the SLAM module is not working properly with the extrinsic parameters. On the other hand, when the processor 130 is able to calculate the pose of the electronic device 100 in reference to the space feature points determined previously and/or the pose of the electronic device 100 determined at a previous period of time, it is determined that the SLAM module is working properly with the extrinsic parameters.
  • FIG. 6 is a flow chart illustrating operation S 530 of FIG. 5 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 6 , the operation S 530 includes operations S 532 to S 534 .
  • the light spots are generated by the structured light generation device 170 as illustrated in FIG. 1 or the structured light generation device 900 as illustrated in FIG. 2 . Take the structured light generation device 170 as illustrated in FIG. 1 as an example. In some embodiments, the structured light generation device 170 generates and emits several light spots every period time.
  • the structured light generation device 170 generates and emits several light spots with a fixed frequency.
  • the processor 130 adjusts the exposure of each of the cameras 150 A to 150 C, so that the cameras 150 A to 150 C are able to capture the images with the light spots.
  • FIG. 7 is a flow chart illustrating operation S 532 of FIG. 6 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 7 , operation S 532 includes operations S 532 A to S 532 C.
  • FIG. 8 A is a schematic diagram illustrating an image 800 A captured by the camera 150 A as illustrated in FIG. 1 and FIG. 4 .
  • FIG. 8 B is a schematic diagram illustrating an image 800 B captured by the camera 150 B as illustrated in FIG. 1 and FIG. 4 . It should be noted that the image 800 A and the image 800 B are captured with the electronic device 100 being at the same position of the mixed reality environment coordinate system M.
  • the processor 130 as illustrated in FIG. 1 obtains several space feature points FP 1 to FP 4 from the image 800 A.
  • the space feature points FP 1 to FP 4 are feature points of the lamp in the real space R as illustrated in FIG. 3 . It should be noted that the processor 130 does not only obtain the space feature points FP 1 to FP 4 , more space feature points may be obtained from FIG. 8 A .
  • the processor 130 as illustrated in FIG. 1 obtains the space feature points FP 1 to FP 4 from the image 800 B.
  • the space feature points FP 1 to FP 4 obtained from the image 800 A and the space feature points FP 1 to FP 4 obtained from the image 800 B are the same space feature points within the mixed reality environment coordinate system M. That is, the positions of the space feature points FP 1 to FP 4 of the image 800 A within the mixed reality environment coordinate system M and the positions of the space feature points FP 1 to FP 4 of the image 800 B within the mixed reality environment coordinate system M are the same.
  • a first light spot is selected from an area circled by the several feature points.
  • the mixed reality environment coordinate system M may include several areas circled by at least three of the space feature points.
  • the processor 130 selects the same area circled byte same space feature points of FIG. 8 A and FIG. 8 B .
  • the processor 130 selects the area FPA circled by the space feature points FP 1 to FP 4 in FIG. 8 A and FIG. 8 B . That is, the processor 130 selects the same area within the mixed reality environment coordinate system M from FIG. 8 A and FIG. 8 B .
  • the processor 130 selects one of the light spots from the area FPA. Reference is made to FIG. 8 A and FIG. 8 B together. As illustrated in FIG. 8 A and FIG. 8 B , the area FPA includes several light spots LP 1 to LP 3 . In some embodiments, in operation S 532 B, the processor 130 selects the light spot LP 1 in FIG. 8 A and FIG. 8 B . In some embodiments, the processor 130 calculates the position of the light spot LP 1 in the mixed reality environment coordinate system M according to the space feature points FP 1 to FP 4 and the images captured by the camera 150 A and 150 B.
  • a pose of the first camera is calculated according to the first image and a pose of the second camera is calculated according to the second image.
  • the processor 130 as illustrated in FIG. 1 calculates a pose of the camera 150 A according to the light spot LP 1 and the image 800 A.
  • the processor 130 as illustrated in FIG. 1 calculates a pose of the camera 150 B according to the light spot LP 1 and the image 800 B. That is, the processor 130 calculates the pose of the camera 150 A and the pose of the camera 150 B according to the position of the same light spot.
  • the pose of the camera 150 A and the pose of the camera 150 B may be calculated according to several light spots.
  • operation S 534 several extrinsic parameters between the several cameras are calibrated according to the several poses of the several cameras. Detail of operation S 534 will be described in the following in reference to FIG. 9 .
  • FIG. 9 is a flow chart illustrating operation S 534 of FIG. 6 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 9 , operation S 534 includes operations S 534 A and S 534 B.
  • a difference between the first pose of the first camera and the second pose of the second camera is calculated.
  • the pose of the electronic device 100 is P during operation S 530
  • the pose of the camera 150 A obtained in operation S 532 is PA
  • the pose of the camera 150 B obtained in operation S 532 is PB
  • the processor 130 as illustrated in FIG. 1 calculates the difference ⁇ P between the pose PA and the pose PB.
  • the difference is taken as the extrinsic parameter between the first camera and the second camera.
  • the processor 130 as illustrated in FIG. 1 takes the difference ⁇ P between the pose PA and the pose PB as the extrinsic parameter between the camera 150 A and the camera 150 B.
  • the processor 130 further updates the extrinsic parameter between the camera 150 A and the camera 150 B stored in the memory as illustrated in FIG. 1 to be the difference ⁇ P between the pose PA and the pose PB.
  • operation S 540 a reset process is performed to reset the extrinsic parameters. Detail of the operation S 540 will be described in reference to FIG. 10 in the following. In some embodiments, the operation S 540 is performed with the electronic device 100 being static.
  • FIG. 10 is a flow chart illustrating operation S 540 of FIG. 5 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 5 , the operation S 540 includes operations S 541 to S 547 .
  • the extrinsic parameter between the first camera and the second camera is reset.
  • the processor 130 resets the extrinsic parameter between the camera 150 A and the camera 150 B to be an initial value.
  • a first pose of the first camera is obtained according to the image captured by the first camera and a second pose of the second camera is obtained according to the image captured by the second camera.
  • the processor 130 as illustrated in FIG. 1 obtains the pose of the camera 150 A according to the space feature points of the image 800 A, and the processor 130 obtains the pose of the camera 150 B according to the space feature points of the image 800 B.
  • the pose of the camera 150 A and the pose of the camera 150 B are calculated with the extrinsic parameter between the camera 150 A and the camera 150 B.
  • a difference between the first pose and the second pose is calculated.
  • the processor 130 as illustrated in FIG. 1 calculates a difference between the pose of the camera 150 A and the pose of the camera 150 B obtained in operation S 543 .
  • operation S 546 the differences between the first pose and the second pose over a period of time is recorded when the first pose and the second pose are stably calculated.
  • the camera 150 A, the camera 150 B and the processor 130 as illustrated in FIG. 1 perform operations S 543 and S 545 over a period of time. For example, at a first time point, the camera 150 A captures a first image and the camera 150 B captures a second image, and the processor 130 calculates the first pose of the camera 150 A according to the first image captured at the first time point and the second pose of the camera 150 B according to the second image captured at the first time point. Then, the processor 130 calculates a difference between the first pose and the second pose corresponding to the first time point.
  • the processor 130 calculates the first pose of the camera 150 A according to the first image captured at the second time point and the second pose of the camera 150 B according to the second image captured at the second time point. Then, the processor 130 calculates a difference between the first pose and the second pose corresponding to the second time point. In this way, the processor 130 calculates several differences of several time points within the period of time.
  • the first pose and the second pose are stably calculated.
  • the processor 130 asks the user to change the pose of the electronic device 100 .
  • the processor 130 sends a signal to the display circuit (not shown) of the electronic device 100 so as to display the signal for asking the user to change the pose of the electronic device 100 .
  • the processor 130 resets or adjusts the extrinsic parameter between the first camera and the second camera again.
  • operation S 547 it is determined whether the differences within the period of time are smaller than a threshold value.
  • the threshold value is stored in the memory 110 as illustrated in FIG. 1 .
  • operation S 530 as illustrated in FIG. 5 is performed.
  • operation S 543 is performed.
  • the extrinsic parameter between the first camera and the second camera is adjusted by the processor 130 before performing operation S 543 .
  • the adjustment to the extrinsic parameter includes increasing/decreasing a distance value between the camera 150 A and the camera 150 B.
  • the adjustment to the extrinsic parameter includes increasing/decreasing a relative rotation value between the camera 150 A and the camera 150 B.
  • operation S 543 is performed so as to recalculate the pose of the camera 150 A and the pose of the camera 150 B with the adjusted extrinsic parameter between the camera 150 A and the camera 150 B.
  • operations S 540 is operated until all of the differences between the poses of the camera 150 A and the poses of the camera 150 B over the period of time are smaller than the threshold value.
  • the examples mentioned above takes the camera 150 A and the camera 150 B as illustrated in FIG. 1 and FIG. 4 for illustrative purposes so as to illustrate the detail of the operations.
  • the operations of other cameras are similar to the operations of the cameras 150 A and 150 B, and will not be described in detail here.
  • the pose and/or the positions of the devices and the feature points are obtained with the SLAM module.
  • the structured light generation devices 170 and 900 mentioned above are devices with the function of projecting a known pattern (often grids or horizontal bars) on to a scene.
  • the way that these deform when striking surfaces allows vision systems to calculate the depth and surface information of the objects in the scene, as used in structured light 3D scanners.
  • the embodiments of the present disclosure utilizes the function of the projecting a known pattern with light spots of the structured light generation device, so as to mimic the feature points of the chessboard or the Deltille grid, and to compensate for the problem of insufficient feature points in general environments, such as the real space R as mentioned above. By increasing the feature points, the accuracy of the calibration to the extrinsic parameters between the cameras is improved.
  • an electronic device a parameter calibration method, and a non-transitory computer readable storage medium are implemented.
  • the extrinsic parameters between the cameras of the self-tracking device can be calibrated with the structured light generation device, in which the deviations of the extrinsic parameters between the cameras can be corrected and the accuracy of the calibration of the extrinsic parameters between the cameras can be improved.
  • a chessboard or a Deltille grid is not necessary, and the users can operate the calibration process without a chessboard or a Deltille grid, which is more convenient.
  • the number of the feature points within the real space R is increased, which improves the accuracy of the calculation of the pose of the devices, and the accuracy of the calibration to the extrinsic parameters between the cameras is thereby improved.
  • the embodiments of the present disclosure can perform a reset process so as to recalculate the extrinsic parameters.
  • parameter calibration method 500 may be added to, replaced, and/or eliminated as appropriate, in accordance with various embodiments of the present disclosure.
  • circuits either dedicated circuits, or general purpose circuits, which operate under the control of one or more processing circuits and coded instructions
  • the functional blocks will typically include transistors or other circuit elements that are configured in such a way as to control the operation of the circuity in accordance with the functions and operations described herein.

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Electromagnetism (AREA)
  • Optics & Photonics (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

An electronic device is disclosed. The electronic device includes a memory, several cameras, and a processor. The memory is configured to store a SLAM module. The several cameras are configured to capture several images of a real space. The processor is configured to: process the SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to several images; and perform a calibration process. The operation of performing the calibration process includes: calculating several poses of several cameras within the environment coordinate system according to several light spots within each of several images, in which several light spots are generated by a structured light generation device; and calibrating several extrinsic parameters between several cameras according to several poses.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • This application claims priority to U.S. Provisional Application Ser. No. 63/483,760, filed Feb. 8, 2023, which is herein incorporated by reference.
  • BACKGROUND Field of Invention
  • The present application relates to an electronic device, a parameter calibration method, and a non-transitory computer readable storage medium. More particularly, the present application relates to an electronic device, a parameter calibration method, and a non-transitory computer readable storage medium with a SLAM module.
  • Description of Related Art
  • Self-tracking devices, such as VR headsets or trackers, access a SLAM module to determine their positions in the real space with the images captured by the cameras within the self-tracking devices. However, changes in the self-tracking devices, such as damage or breakage during deliver or usage, can affect the relative position and the relation rotation between the cameras of the self-tracking devices, and the pre-set extrinsic parameters, including the pre-set relative position parameter and the pre-set relative rotation parameter, between the cameras of the self-tracking devices may no longer be used, in which the performance of the SLAM module may be decreased.
  • When the changes to the cameras (such as the relative position and the relation rotation between the cameras) of the self-tracking devices become significant, the self-tracking devices may become unable to track themselves with the SLAM module even if the cameras themselves are functioning properly. Several methods are proposed to recalculate the extrinsic parameters of the cameras of the self-tracking devices, such as recalculating the extrinsic parameters with a checkerboard or a Deltille grid. However, it is impractical for the users to carry the checkerboard or the Deltille grid at any time.
  • Therefore, how to calibrate the extrinsic parameters between the cameras of the self-tracking device without the existence of the checkerboard or the Deltille grid is a problem to be solved.
  • SUMMARY
  • The disclosure provides an electronic device. The electronic device includes a memory, several cameras, and a processor. The memory is configured to store a SLAM module. The several cameras are configured to capture several images of a real space. The processor is coupled to the camera and the memory. The processor is configured to: process the SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to several images; and perform a calibration process. The operation of performing the calibration process includes: calculating several poses of several cameras within the environment coordinate system according to several light spots within each of several images, in which several light spots are generated by a structured light generation device; and calibrating several extrinsic parameters between several cameras according to several poses.
  • The disclosure provides a parameter calibration method suitable for an electronic device. The parameter calibration method includes the following operations: capturing several images of a real space by several cameras; processing a SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to several images by a processor; and performing a calibration process by the processor. The operation of performing the calibration process includes the following operations: calculating several poses of several cameras within the environment coordinate system according to several light spots within each of several images, wherein several light spots are generated by a structured light generation device; and calibrating several extrinsic parameters between several cameras according to several poses.
  • The disclosure provides a non-transitory computer readable storage medium with a computer program to execute aforesaid parameter calibration method.
  • It is to be understood that both the foregoing general description and the following detailed description are by examples and are intended to provide further explanation of the invention as claimed.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • Aspects of the present disclosure are best understood from the following detailed description when read with the accompanying figures. It is noted that, according to the standard practice in the industry, various features are not drawn to scale. In fact, the dimensions of the various features may be arbitrarily increased or reduced for clarity of discussion.
  • FIG. 1 is a schematic block diagram illustrating an electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 2 is a schematic block diagram illustrating another electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 3 is a schematic diagram illustrating a user operating the electronic device as illustrated in FIG. 1 in accordance with some embodiments of the present disclosure.
  • FIG. 4 is a schematic diagram illustrating an electronic device in accordance with some embodiments of the present disclosure.
  • FIG. 5 is a flowchart illustrating a parameter calibration method in accordance with some embodiments of the present disclosure.
  • FIG. 6 is a flow chart illustrating an operation of FIG. 5 in accordance with some embodiments of the present disclosure.
  • FIG. 7 is a flow chart illustrating an operation of FIG. 6 in accordance with some embodiments of the present disclosure.
  • FIG. 8A is a schematic diagram illustrating an image captured by an camera as illustrated in FIG. 1 and FIG. 4 .
  • FIG. 8B is a schematic diagram illustrating an image captured by another camera as illustrated in FIG. 1 and FIG. 4 .
  • FIG. 9 is a flow chart illustrating an operation of FIG. 6 in accordance with some embodiments of the present disclosure.
  • FIG. 10 is a flow chart illustrating an operation of FIG. 5 in accordance with some embodiments of the present disclosure.
  • DETAILED DESCRIPTION
  • Reference will now be made in detail to the present embodiments of the disclosure, examples of which are illustrated in the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the description to refer to the same or like parts.
  • It will be understood that, in the description herein and throughout the claims that follow, although the terms “first,” “second,” etc. may be used to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another. For example, a first element could be termed a second element, and, similarly, a second element could be termed a first element, without departing from the scope of the embodiments.
  • It will be understood that, in the description herein and throughout the claims that follow, the terms “comprise” or “comprising,” “include” or “including,” “have” or “having,” “contain” or “containing” and the like used herein are to be understood to be open-ended, i.e., to mean including but not limited to.
  • It will be understood that, in the description herein and throughout the claims that follow, the phrase “and/or” includes any and all combinations of one or more of the associated listed items.
  • Reference is made to FIG. 1 . FIG. 1 is a schematic block diagram illustrating an electronic device 100 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 1 , the electronic device 100 includes a memory 110, a processor 130, several cameras 150A to 150C, and a structured light generation device 170. The memory 110, the cameras 150A to 150C, and the structured light generation device 170 couple to the processor 130.
  • Reference is made to FIG. 2 . FIG. 2 is a schematic block diagram illustrating another electronic device 200 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 2 , the electronic device 200 includes a memory 210, a processor 230, and several cameras 250A to 250C. The memory 210 and the cameras 250A to 250C couple to the processor 230. In some embodiments, the electronic device 200 couple to a structured light generation device 900. That is, in some embodiments, the electronic device 200 and the structured light generation device 900 are separate devices.
  • It should be noted that, in FIG. 1 and FIG. 2 , three cameras are illustrated. However, the electronic device 100 and the electronic device 200 are for illustrative purposes only, and the embodiments of the present disclosure are not limited thereto. For example, in some embodiments, the electronic device 100 and the electronic device 200 may include two cameras, or more than three cameras. It is noted that, the embodiments shown in FIG. 1 and FIG. 2 are merely an example and not meant to limit the present disclosure.
  • One or more programs are stored in the memory 110 and the memory 210 and are configured to be executed by the processor 130 or the processor 230, in order to perform a parameter calibration method.
  • In some embodiments, the electronic device 100 and the electronic device 200 may be an HMD (head-mounted display) device, a tracking device, or any other device with self-tracking function. The HMD device may be wear on the head of a user.
  • In some embodiments, the memory 110 and the memory 210 store a SLAM (Simultaneous localization and mapping) module. The electronic device 100 and the electronic device 200 may be configured to process the SLAM module. The SLAM module includes functions such as image capturing, features extracting from the image, and localizing according to the extracted features. In some embodiments, the SLAM module include a SLAM algorithm, in which the processor 130 access and process the SLAM module so as to localize the electronic device 100 according to the images captured by the cameras 150A to 150C. Similarly, the processor 230 access and process the SLAM module so as to localize the electronic device 200 according to the images captured by the cameras 250A to 250C. The details of the SLAM system will not be described herein.
  • Specifically, in some embodiments, the electronic device 100 may be applied in a virtual reality (VR)/mixed reality (MR)/augmented reality (AR) system. For example, the electronic device 100 may be realized by, a standalone head mounted display device (HMD) or VIVE HMD.
  • In some embodiments, the processor 130 and 230 can be realized by, for example, one or more processing circuits, such as central processing circuits and/or micro processing circuits, but are not limited in this regard. In some embodiments, the memory 110 and 210 include one or more memory devices, each of which includes, or a plurality of which collectively include a computer readable storage medium. The non-transitory computer readable storage medium may include a read-only memory (ROM), a flash memory, a floppy disk, a hard disk, an optical disc, a flash disk, a flash drive, a tape, a database accessible from a network, and/or any storage medium with the same functionality that can be contemplated by persons of ordinary skill in the art to which this disclosure pertains.
  • The cameras 150A to 150C and the cameras 250A to 250C are configured to capture one or more images of the real space that the electronic device 100 and 200 are operated. In some embodiments, the cameras 150A to 150C and the cameras 250A to 250C may be realized by camera circuit devices or any other camera circuits with image capture functions.
  • In some embodiments, the electronic device 100 and 200 include other circuits such as a display circuit and an I/O circuit. In some embodiments, the display circuit covers a field of view of the user and shows a virtual image at the field of view of the user.
  • For ease of illustration, the following take the electronic device 100 as illustrated in FIG. 1 for illustrative purpose. It should be noted that the operation of the electronic device 200 as illustrated in FIG. 2 is similar to the electronic device 100 as illustrated in FIG. 1 .
  • Reference is made to FIG. 3 together. FIG. 3 is a schematic diagram illustrating a user U operating the electronic device 100 as illustrated in FIG. 1 in accordance with some embodiments of the present disclosure.
  • As illustrated in FIG. 3 , the user U is wearing the electronic device 100 as illustrated in FIG. 1 on the head of the user U. In some embodiments, the cameras 150A to 150C capture several frames of images of the real space R. The processor 130 process the SLAM module to establish a mixed reality environment coordinate system M in correspondence to the real space R according to several space feature points of the images captured by the cameras 150A to 150C. In some embodiments, the processor 130 obtains a device pose of the electronic device 100 within the mixed reality environment coordinate system M according to the feature points within the images. When the electronic device 100 moves in the real space R, the processor 130 tracks the device pose of the electronic device 100 within the mixed reality environment coordinate system M.
  • In other embodiments, the mixed reality environment coordinate system M could be an augmented reality environment coordinate system or an extended reality environment coordinate system. The following takes the mixed reality environment coordinate system M for examples for illustrative purposes; however, the embodiments of the present disclosure are not limited thereto.
  • In some embodiments, the device pose of the electronic device 100 includes a position and a rotation angle.
  • When calculating the device pose of the electronic device 100 according to the images captured by the cameras 150A to 150C, the intrinsic parameter and the extrinsic parameter of each of the cameras 150A to 150C are considered. In some embodiments, the extrinsic parameters represent a rigid transformation from 3D world coordinate system to the 3D camera's coordinate system. The intrinsic parameters represent a projective transformation from the 3D camera's coordinates into the 2D image coordinates. In some embodiments, the extrinsic parameters of the cameras 150A to 150C include the difference between the poses of the cameras.
  • Reference is made to FIG. 4 together. FIG. 4 is a schematic diagram illustrating an electronic device 100 in accordance with some embodiments of the present disclosure. In FIG. 4 , camera 150A and camera 150B of the electronic device 100 are taken as an example for illustration. In some embodiments, when the electronic device 100 is made, the positions of the camera 150A and the camera 150B and the rotation angles of the camera 150A and the camera 150B relative to the electronic device 100 is preset, and an extrinsic parameter between the camera 150A and the camera 150B is preset within the SLAM module. Similarly, the extrinsic parameters between each two of the cameras are preset within the SLAM module.
  • When the processor 130 tracks the device pose of the electronic device 100 within the mixed reality environment coordinate system M, the extrinsic parameters between each two of the cameras are preset within the SLAM module are considered. However, during the operation of the electronic device 100, the positions of the camera 150A and the camera 150B and the rotation angles of the camera 150A and the camera 150B relative to the electronic device 100 may be changed, and the SLAM module may no longer be working properly with the images captured by the cameras 150A and 150B and the preset extrinsic parameter between the cameras 150A and 150B. Therefore, a method for calibrating the extrinsic parameters between the cameras of the electronic device 100 is in need. In some embodiments, the extrinsic parameters are stored in the memory 110 for the processor 130 to access and operate with the SLAM module.
  • Reference is made to FIG. 5 . For better understanding of the present disclosure, the detailed operation of the electronic device 100 as illustrated in FIG. 1 will be discussed in accompanying with the embodiments shown in FIG. 5 . FIG. 5 is a flowchart illustrating a parameter calibration method 500 in accordance with some embodiments of the present disclosure. It should be noted that the parameter calibration method 500 can be applied to a device having a structured that is the same as or similar to the structured of the electronic device 100 shown in FIG. 1 or the electronic device 200 shown in FIG. 2 . To simplify the description below, the embodiments shown in FIG. 1 will be used as an example to describe the parameter calibration method 500 in accordance with some embodiments of the present disclosure. However, the present disclosure is not limited to application to the embodiments shown in FIG. 1 .
  • As shown in FIG. 5 , the parameter calibration method 500 includes operations S510 to S540.
  • In operation S510, the SLAM module is processed to track the device pose of the electronic device within the mixed reality environment coordinate system according to several images. In some embodiments, the processor 130 tracks the device pose of the electronic device 100 within the mixed reality environment coordinate system M according to several space feature points within the images captured by the cameras 130A to 130C.
  • In operation S520, it is determined whether the SLAM module is working properly with the extrinsic parameters. In some embodiments, when the SLAM module is working properly with the extrinsic parameters stored in the memory 110, operation S530 is performed. On the other hand, when the SLAM module is not working properly with the extrinsic parameters stored in the memory 110, operation S540 is performed.
  • In some embodiments, the processor 130 of the electronic device 100 determines the pose of the electronic device 100 every period of time. When determining the pose of the electronic device 100, the processor 130 refers to the previous pose of the electronic device 100 determined at a previous period of time. In some embodiments, the processor 130 further refers to the positions of the space feature points determined previously when determining the pose of the electronic device 100.
  • When the processor 130 is unable to calculate the pose of the electronic device 100 in reference to the space feature points determined previously and/or the pose of the electronic device 100 determined at a previous period of time, it is determined that the SLAM module is not working properly with the extrinsic parameters. On the other hand, when the processor 130 is able to calculate the pose of the electronic device 100 in reference to the space feature points determined previously and/or the pose of the electronic device 100 determined at a previous period of time, it is determined that the SLAM module is working properly with the extrinsic parameters.
  • In operation S530, a calibration process is performed. Detail of the calibration process will be described in reference to FIG. 6 in the following.
  • Reference is made to FIG. 6 together. FIG. 6 is a flow chart illustrating operation S530 of FIG. 5 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 6 , the operation S530 includes operations S532 to S534.
  • In operation S532, several poses of several cameras are calculated within the mixed reality environment coordinate system according to several light spots within each of the several images.
  • In some embodiments, the light spots are generated by the structured light generation device 170 as illustrated in FIG. 1 or the structured light generation device 900 as illustrated in FIG. 2 . Take the structured light generation device 170 as illustrated in FIG. 1 as an example. In some embodiments, the structured light generation device 170 generates and emits several light spots every period time.
  • In some embodiments, the structured light generation device 170 generates and emits several light spots with a fixed frequency. The processor 130 adjusts the exposure of each of the cameras 150A to 150C, so that the cameras 150A to 150C are able to capture the images with the light spots.
  • Detail of the operation S532 will be described in reference to FIG. 7 as following.
  • Reference is made to FIG. 7 together. FIG. 7 is a flow chart illustrating operation S532 of FIG. 6 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 7 , operation S532 includes operations S532A to S532C.
  • In operation S532A, several space feature points are detected from a first image captured by a first camera and a second image captured by a second camera.
  • Reference is made to FIG. 8A and FIG. 8B together. FIG. 8A is a schematic diagram illustrating an image 800A captured by the camera 150A as illustrated in FIG. 1 and FIG. 4 . FIG. 8B is a schematic diagram illustrating an image 800B captured by the camera 150B as illustrated in FIG. 1 and FIG. 4 . It should be noted that the image 800A and the image 800B are captured with the electronic device 100 being at the same position of the mixed reality environment coordinate system M.
  • The processor 130 as illustrated in FIG. 1 obtains several space feature points FP1 to FP4 from the image 800A. The space feature points FP1 to FP4 are feature points of the lamp in the real space R as illustrated in FIG. 3 . It should be noted that the processor 130 does not only obtain the space feature points FP1 to FP4, more space feature points may be obtained from FIG. 8A.
  • Similarly, the processor 130 as illustrated in FIG. 1 obtains the space feature points FP1 to FP4 from the image 800B. The space feature points FP1 to FP4 obtained from the image 800A and the space feature points FP1 to FP4 obtained from the image 800B are the same space feature points within the mixed reality environment coordinate system M. That is, the positions of the space feature points FP1 to FP4 of the image 800A within the mixed reality environment coordinate system M and the positions of the space feature points FP1 to FP4 of the image 800B within the mixed reality environment coordinate system M are the same.
  • Reference is made to FIG. 7 again. In operation S532B, a first light spot is selected from an area circled by the several feature points. The mixed reality environment coordinate system M may include several areas circled by at least three of the space feature points.
  • In some embodiments, the processor 130 selects the same area circled byte same space feature points of FIG. 8A and FIG. 8B. For example, the processor 130 selects the area FPA circled by the space feature points FP1 to FP4 in FIG. 8A and FIG. 8B. That is, the processor 130 selects the same area within the mixed reality environment coordinate system M from FIG. 8A and FIG. 8B.
  • After the processor 130 selects the area FPA from FIG. 8A and FIG. 8B, the processor 130 selects one of the light spots from the area FPA. Reference is made to FIG. 8A and FIG. 8B together. As illustrated in FIG. 8A and FIG. 8B, the area FPA includes several light spots LP1 to LP3. In some embodiments, in operation S532B, the processor 130 selects the light spot LP1 in FIG. 8A and FIG. 8B. In some embodiments, the processor 130 calculates the position of the light spot LP1 in the mixed reality environment coordinate system M according to the space feature points FP1 to FP4 and the images captured by the camera 150A and 150B.
  • In operation S532C, a pose of the first camera is calculated according to the first image and a pose of the second camera is calculated according to the second image. Reference is made to FIG. 1 and FIG. 4 together. In some embodiments, the processor 130 as illustrated in FIG. 1 calculates a pose of the camera 150A according to the light spot LP1 and the image 800A. Similarly, the processor 130 as illustrated in FIG. 1 calculates a pose of the camera 150B according to the light spot LP1 and the image 800B. That is, the processor 130 calculates the pose of the camera 150A and the pose of the camera 150B according to the position of the same light spot.
  • It should be noted that, in operation S532C, the pose of the camera 150A and the pose of the camera 150B may be calculated according to several light spots.
  • Reference is made to FIG. 6 again. In operation S534, several extrinsic parameters between the several cameras are calibrated according to the several poses of the several cameras. Detail of operation S534 will be described in the following in reference to FIG. 9 .
  • Reference is made to FIG. 9 together. FIG. 9 is a flow chart illustrating operation S534 of FIG. 6 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 9 , operation S534 includes operations S534A and S534B.
  • In operation S534A, a difference between the first pose of the first camera and the second pose of the second camera is calculated. Reference is made to FIG. 3 and FIG. 4 together. Assume that the pose of the electronic device 100 is P during operation S530, and the pose of the camera 150A obtained in operation S532 is PA, the pose of the camera 150B obtained in operation S532 is PB, the processor 130 as illustrated in FIG. 1 calculates the difference ΔP between the pose PA and the pose PB.
  • Reference is made to FIG. 9 again. In operation S534B, the difference is taken as the extrinsic parameter between the first camera and the second camera. For example, in some embodiments, the processor 130 as illustrated in FIG. 1 takes the difference ΔP between the pose PA and the pose PB as the extrinsic parameter between the camera 150A and the camera 150B. In some embodiments, the processor 130 further updates the extrinsic parameter between the camera 150A and the camera 150B stored in the memory as illustrated in FIG. 1 to be the difference ΔP between the pose PA and the pose PB.
  • Through the operations of S530, by calculating the poses of the cameras according to the same feature points within the mixed reality environment coordinate system M, the extrinsic parameters between the cameras can be calibrated.
  • Reference is made to FIG. 5 again. In operation S540, a reset process is performed to reset the extrinsic parameters. Detail of the operation S540 will be described in reference to FIG. 10 in the following. In some embodiments, the operation S540 is performed with the electronic device 100 being static.
  • Reference is made to FIG. 10 together. FIG. 10 is a flow chart illustrating operation S540 of FIG. 5 in accordance with some embodiments of the present disclosure. As illustrated in FIG. 5 , the operation S540 includes operations S541 to S547.
  • In operation S541, the extrinsic parameter between the first camera and the second camera is reset. Reference is made to FIG. 1 together. For example, the processor 130 resets the extrinsic parameter between the camera 150A and the camera 150B to be an initial value.
  • In operation S543, a first pose of the first camera is obtained according to the image captured by the first camera and a second pose of the second camera is obtained according to the image captured by the second camera. Reference is made to FIG. 8A and FIG. 8B together. In some embodiments, the processor 130 as illustrated in FIG. 1 obtains the pose of the camera 150A according to the space feature points of the image 800A, and the processor 130 obtains the pose of the camera 150B according to the space feature points of the image 800B. In some embodiments, the pose of the camera 150A and the pose of the camera 150B are calculated with the extrinsic parameter between the camera 150A and the camera 150B.
  • In operation S545, a difference between the first pose and the second pose is calculated. In some embodiments, the processor 130 as illustrated in FIG. 1 calculates a difference between the pose of the camera 150A and the pose of the camera 150B obtained in operation S543.
  • In operation S546, the differences between the first pose and the second pose over a period of time is recorded when the first pose and the second pose are stably calculated. In some embodiments, in operation S546, the camera 150A, the camera 150B and the processor 130 as illustrated in FIG. 1 perform operations S543 and S545 over a period of time. For example, at a first time point, the camera 150A captures a first image and the camera 150B captures a second image, and the processor 130 calculates the first pose of the camera 150A according to the first image captured at the first time point and the second pose of the camera 150B according to the second image captured at the first time point. Then, the processor 130 calculates a difference between the first pose and the second pose corresponding to the first time point. Similarly, at a second time point, the processor 130 calculates the first pose of the camera 150A according to the first image captured at the second time point and the second pose of the camera 150B according to the second image captured at the second time point. Then, the processor 130 calculates a difference between the first pose and the second pose corresponding to the second time point. In this way, the processor 130 calculates several differences of several time points within the period of time.
  • It should be noted that, in operation S546, the first pose and the second pose are stably calculated. In some embodiments, when the first pose and the second pose are not stably calculated, the processor 130 asks the user to change the pose of the electronic device 100. In some embodiments, the processor 130 sends a signal to the display circuit (not shown) of the electronic device 100 so as to display the signal for asking the user to change the pose of the electronic device 100. In some other embodiments, when the first pose and the second pose are not stably calculated, the processor 130 resets or adjusts the extrinsic parameter between the first camera and the second camera again.
  • In operation S547, it is determined whether the differences within the period of time are smaller than a threshold value. In some embodiments, the threshold value is stored in the memory 110 as illustrated in FIG. 1 . In some embodiments, when all of the differences between the poses of the camera 150A and the poses of the camera 150B recorded in operation S546 are smaller than the threshold value, operation S530 as illustrated in FIG. 5 is performed. On the other hand, when it is determined that not all of the differences between the poses of the camera 150A and the poses of the camera 150B recorded in operation S546 are smaller than the threshold value, operation S543 is performed.
  • In some embodiments, when not all of the differences between the poses of the camera 150A and the poses of the camera 150B recorded in operation S546 are smaller than the threshold value , the extrinsic parameter between the first camera and the second camera is adjusted by the processor 130 before performing operation S543. In some embodiments, the adjustment to the extrinsic parameter includes increasing/decreasing a distance value between the camera 150A and the camera 150B. In some other embodiments, the adjustment to the extrinsic parameter includes increasing/decreasing a relative rotation value between the camera 150A and the camera 150B.
  • In some embodiments, after the extrinsic parameter between the camera 150A and the camera 150B is adjusted, operation S543 is performed so as to recalculate the pose of the camera 150A and the pose of the camera 150B with the adjusted extrinsic parameter between the camera 150A and the camera 150B.
  • In some embodiments, operations S540 is operated until all of the differences between the poses of the camera 150A and the poses of the camera 150B over the period of time are smaller than the threshold value.
  • The examples mentioned above takes the camera 150A and the camera 150B as illustrated in FIG. 1 and FIG. 4 for illustrative purposes so as to illustrate the detail of the operations. The operations of other cameras are similar to the operations of the cameras 150A and 150B, and will not be described in detail here.
  • It should be noted that, in the embodiments of the present disclosure, the pose and/or the positions of the devices and the feature points are obtained with the SLAM module.
  • The structured light generation devices 170 and 900 mentioned above are devices with the function of projecting a known pattern (often grids or horizontal bars) on to a scene. The way that these deform when striking surfaces allows vision systems to calculate the depth and surface information of the objects in the scene, as used in structured light 3D scanners. The embodiments of the present disclosure utilizes the function of the projecting a known pattern with light spots of the structured light generation device, so as to mimic the feature points of the chessboard or the Deltille grid, and to compensate for the problem of insufficient feature points in general environments, such as the real space R as mentioned above. By increasing the feature points, the accuracy of the calibration to the extrinsic parameters between the cameras is improved.
  • Through the operations of various embodiments described above, an electronic device, a parameter calibration method, and a non-transitory computer readable storage medium are implemented. The extrinsic parameters between the cameras of the self-tracking device can be calibrated with the structured light generation device, in which the deviations of the extrinsic parameters between the cameras can be corrected and the accuracy of the calibration of the extrinsic parameters between the cameras can be improved.
  • Furthermore, in the embodiments of the present disclosure, a chessboard or a Deltille grid is not necessary, and the users can operate the calibration process without a chessboard or a Deltille grid, which is more convenient. Moreover, by generating the spot lights at the real space R, the number of the feature points within the real space R is increased, which improves the accuracy of the calculation of the pose of the devices, and the accuracy of the calibration to the extrinsic parameters between the cameras is thereby improved.
  • Additionally, when critical situations occur, for example, when the SLAM module is not working properly, the embodiments of the present disclosure can perform a reset process so as to recalculate the extrinsic parameters.
  • It should be noted that in the operations of the abovementioned parameter calibration method 500, no particular sequence is required unless otherwise specified. Moreover, the operations may also be performed simultaneously or the execution times thereof may at least partially overlap.
  • Furthermore, the operations of the parameter calibration method 500 may be added to, replaced, and/or eliminated as appropriate, in accordance with various embodiments of the present disclosure.
  • Various functional components or blocks have been described herein. As will be appreciated by persons skilled in the art, the functional blocks will preferably be implemented through circuits (either dedicated circuits, or general purpose circuits, which operate under the control of one or more processing circuits and coded instructions), which will typically include transistors or other circuit elements that are configured in such a way as to control the operation of the circuity in accordance with the functions and operations described herein.
  • Although the present disclosure has been described in considerable detail with reference to certain embodiments thereof, other embodiments are possible. Therefore, the scope of the appended claims should not be limited to the description of the embodiments contained herein.
  • It will be apparent to those skilled in the art that various modifications and variations can be made to the structured of the present invention without departing from the scope or spirit of the invention. In view of the foregoing, it is intended that the present invention cover modifications and variations of this invention provided they fall within the scope of the following claims.

Claims (20)

What is claimed is:
1. An electronic device, comprising:
a memory, configured to store a SLAM module;
a plurality of cameras, configured to capture a plurality of images of a real space; and
a processor, coupled to the plurality of cameras and the memory, configured to:
process the SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to the plurality of images; and
perform a calibration process, comprising:
calculate a plurality of poses of the plurality of cameras within the environment coordinate system according to a plurality of light spots within each of the plurality of images, wherein the plurality of light spots are generated by a structured light generation device; and
calibrate a plurality of extrinsic parameters between the plurality of cameras according to the plurality of poses.
2. The electronic device of claim 1, wherein a first camera of the plurality of cameras is configured to capture a first image of the plurality of images, and a second camera of the plurality of cameras is configured to capture a second image of the plurality of images, wherein the processor is further configured to:
calculate a first pose of the first camera according to the first image;
calculate a second pose of the second camera according to the second image;
calculate a difference between the first pose and the second pose; and
take the difference as a first extrinsic parameter between the first camera and the second camera.
3. The electronic device of claim 2, wherein the plurality of light spots comprise a first light spot, wherein both of the first image and the second image comprise the first light spot, wherein the first pose is calculated according to the first light spot within the first image, and the second pose is calculated according to the first light spot within the second image.
4. The electronic device of claim 3, wherein the processor is further configured to:
detect a plurality of space feature points from the first image and the second image; and
select the first light spot from an area circled by the plurality of space feature points.
5. The electronic device of claim 1, wherein the processor is further configured to:
determine whether the SLAM module is working properly with the plurality of extrinsic parameters;
perform the calibration process when the SLAM module is working properly; and
perform a reset process to reset the plurality of extrinsic parameters until the SLAM module is working properly with the plurality of extrinsic parameters.
6. The electronic device of claim 5, wherein the processor is further configured to:
obtain a first pose of a first camera according to a first image captured by the first camera of the plurality of cameras;
obtain a second pose of a second camera according to a second image captured by the second camera of the plurality of cameras; and
adjust a first extrinsic parameters between the first camera and the second camera until a difference between the first pose and the second pose is smaller than a threshold value.
7. The electronic device of claim 1, wherein the plurality of light spots are generated with a frequency, wherein the processor is further configured to:
adjust a plurality of exposures of the plurality of cameras so that the plurality of cameras are able to capture the plurality of images with the plurality of light spots.
8. A parameter calibration method, suitable for an electronic device, comprising:
capturing a plurality of images of a real space by a plurality of cameras;
processing a SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to the plurality of images by a processor; and
performing a calibration process by the processor, comprising:
calculating a plurality of poses of the plurality of cameras within the environment coordinate system according to a plurality of light spots within each of the plurality of images, wherein the plurality of light spots are generated by a structured light generation device; and
calibrating a plurality of extrinsic parameters between the plurality of cameras according to the plurality of poses.
9. The parameter calibration method of claim 8, further comprising:
capturing a first image of the plurality of images by a first camera of the plurality of cameras;
capturing a second image of the plurality of images by a second camera of the plurality of cameras;
calculating a first pose of the first camera according to the first image;
calculating a second pose of the second camera according to the second image;
calculating a difference between the first pose and the second pose; and
taking the difference as a first extrinsic parameter between the first camera and the second camera.
10. The parameter calibration method of claim 9, wherein the plurality of light spots comprise a first light spot, wherein both of the first image and the second image comprise the first light spot, the parameter calibration method further comprising:
calculating the first pose according to the first light spot within the first image, and calculating the second pose according to the first light spot within the second image.
11. The parameter calibration method of claim 10, further comprising:
detecting a plurality of space feature points from the first image and the second image; and
selecting the first light spot from an area circled by the plurality of space feature points.
12. The parameter calibration method of claim 8, further comprising:
determining whether the SLAM module is working properly with the plurality of extrinsic parameters;
performing the calibration process when the SLAM module is working properly; and
performing a reset process to reset the plurality of extrinsic parameters until the SLAM module is working properly with the plurality of extrinsic parameters.
13. The parameter calibration method of claim 12, further comprising:
obtaining a first pose of a first camera according to a first image captured by the first camera of the plurality of cameras;
obtaining a second pose of a second camera according to a second image captured by the second camera of the plurality of cameras; and
adjusting a first extrinsic parameters between the first camera and the second camera until a difference between the first pose and the second pose is smaller than a threshold value.
14. The parameter calibration method of claim 8, further comprising:
generating the plurality of light spots are generated with a frequency; and
adjusting a plurality of exposures of the plurality of cameras so that the plurality of cameras are able to capture the plurality of images with the plurality of light spots.
15. A non-transitory computer readable storage medium, wherein the non-transitory computer readable storage medium comprises one or more computer programs stored therein, and the one or more computer programs can be executed by one or more processors so as to be configured to operate a parameter calibration method, wherein the parameter calibration method comprises:
capturing a plurality of images of a real space by a plurality of cameras of an electronic device;
processing a SLAM module to establish an environment coordinate system in correspondence to the real space and to track a device pose of the electronic device within the environment coordinate system according to the plurality of images; and
performing a calibration process, comprising:
calculating a plurality of poses of the plurality of cameras within the environment coordinate system according to a plurality of light spots within each of the plurality of images, wherein the plurality of light spots are generated by a structured light generation device; and
calibrating a plurality of extrinsic parameters between the plurality of cameras according to the plurality of poses.
16. The non-transitory computer readable storage medium of claim 15, wherein the parameter calibration method further comprises:
capturing a first image of the plurality of images by a first camera of the plurality of cameras;
capturing a second image of the plurality of images by a second camera of the plurality of cameras;
calculating a first pose of the first camera according to the first image;
calculating a second pose of the second camera according to the second image;
calculating a difference between the first pose and the second pose; and
taking the difference as a first extrinsic parameter between the first camera and the second camera.
17. The non-transitory computer readable storage medium of claim 16, wherein the parameter calibration method further comprises:
calculating the first pose according to a first light spot within the first image, and calculating the second pose according to the first light spot within the second image;
wherein the plurality of light spots comprise the first light spot, wherein both of the first image and the second image comprise the first light spot.
18. The non-transitory computer readable storage medium of claim 17, wherein the parameter calibration method further comprises:
detecting a plurality of space feature points from the first image and the second image; and
selecting the first light spot from an area circled by the plurality of space feature points.
19. The non-transitory computer readable storage medium of claim 15, wherein the parameter calibration method further comprises:
determining whether the SLAM module is working properly with the plurality of extrinsic parameters;
performing the calibration process when the SLAM module is working properly; and
performing a reset process to reset the plurality of extrinsic parameters until the SLAM module is working properly with the plurality of extrinsic parameters, comprising:
obtaining a first pose of a first camera according to a first image captured by the first camera of the plurality of cameras;
obtaining a second pose of a second camera according to a second image captured by the second camera of the plurality of cameras; and
adjusting a first extrinsic parameters between the first camera and the second camera until a difference between the first pose and the second pose is smaller than a threshold value.
20. The non-transitory computer readable storage medium of claim 15, wherein the parameter calibration method further comprises:
generating the plurality of light spots are generated with a frequency; and
adjusting a plurality of exposures of the plurality of cameras so that the plurality of cameras are able to capture the plurality of images with the plurality of light spots.
US18/432,065 2023-02-08 2024-02-05 Electronic device, parameter calibration method, and non-transitory computer readable storage medium Pending US20240265579A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US18/432,065 US20240265579A1 (en) 2023-02-08 2024-02-05 Electronic device, parameter calibration method, and non-transitory computer readable storage medium

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US202363483760P 2023-02-08 2023-02-08
US18/432,065 US20240265579A1 (en) 2023-02-08 2024-02-05 Electronic device, parameter calibration method, and non-transitory computer readable storage medium

Publications (1)

Publication Number Publication Date
US20240265579A1 true US20240265579A1 (en) 2024-08-08

Family

ID=92119973

Family Applications (1)

Application Number Title Priority Date Filing Date
US18/432,065 Pending US20240265579A1 (en) 2023-02-08 2024-02-05 Electronic device, parameter calibration method, and non-transitory computer readable storage medium

Country Status (3)

Country Link
US (1) US20240265579A1 (en)
CN (1) CN118470124A (en)
TW (1) TWI883813B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20250329052A1 (en) * 2024-04-18 2025-10-23 Htc Corporation Electronic device, parameter calibration method, and non-transitory computer readable storage medium

Citations (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100165116A1 (en) * 2008-12-30 2010-07-01 Industrial Technology Research Institute Camera with dynamic calibration and method thereof
US20160012588A1 (en) * 2014-07-14 2016-01-14 Mitsubishi Electric Research Laboratories, Inc. Method for Calibrating Cameras with Non-Overlapping Views
US20160112588A1 (en) * 2014-10-16 2016-04-21 Fuji Xerox Co., Ltd. Maintenance necessity estimation apparatus and non-transitory computer readable medium
US9392262B2 (en) * 2014-03-07 2016-07-12 Aquifi, Inc. System and method for 3D reconstruction using multiple multi-channel cameras
US20170249752A1 (en) * 2016-02-29 2017-08-31 Canon Kabushiki Kaisha Device for measuring position and orientation of imaging apparatus and method therefor
US20170359573A1 (en) * 2016-06-08 2017-12-14 SAMSUNG SDS CO., LTD., Seoul, KOREA, REPUBLIC OF; Method and apparatus for camera calibration using light source
US20180124387A1 (en) * 2016-10-28 2018-05-03 Daqri, Llc Efficient augmented reality display calibration
US20180208311A1 (en) * 2017-01-23 2018-07-26 Hangzhou Zero Zero Technology Co., Ltd. System and method for omni-directional obstacle avoidance in aerial systems
US20190004178A1 (en) * 2016-03-16 2019-01-03 Sony Corporation Signal processing apparatus and signal processing method
US20190147625A1 (en) * 2017-11-15 2019-05-16 Magic Leap, Inc. System and methods for extrinsic calibration of cameras and diffractive optical elements
US20190158813A1 (en) * 2016-06-10 2019-05-23 Lucid VR, Inc. Real Time Re-Calibration of Stereo Cameras
US20190243388A1 (en) * 2018-02-07 2019-08-08 Hangzhou Zero Zero Technology Co., Ltd. Unmanned aerial vehicle including an omnidirectional depth sensing and obstacle avoidance aerial system and method of operating same
US10484665B2 (en) * 2017-04-20 2019-11-19 Panasonic Intellectual Property Management Co., Ltd. Camera parameter set calculation method, recording medium, and camera parameter set calculation apparatus
US20190364206A1 (en) * 2018-05-25 2019-11-28 Aquifi, Inc. Systems and methods for multi-camera placement
US20200145588A1 (en) * 2017-07-14 2020-05-07 Canon Kabushiki Kaisha Information processing apparatus presenting information, information processing method, and storage medium
US20200143603A1 (en) * 2017-07-11 2020-05-07 Canon Kabushiki Kaisha Information processing apparatus, information processing method, and non-transitory computer-readable storage medium
US20210124174A1 (en) * 2018-07-17 2021-04-29 Sony Corporation Head mounted display, control method for head mounted display, information processor, display device, and program
US11360375B1 (en) * 2020-03-10 2022-06-14 Rockwell Collins, Inc. Stereoscopic camera alignment via laser projection
US20220198677A1 (en) * 2020-12-18 2022-06-23 Qualcomm Incorporated Object segmentation and feature tracking
US20220222856A1 (en) * 2021-01-13 2022-07-14 Ambarella International Lp Fixed pattern calibration for multi-view stitching
US20220309761A1 (en) * 2019-12-12 2022-09-29 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Target detection method, device, terminal device, and medium
US20230021721A1 (en) * 2020-01-14 2023-01-26 Kyocera Corporation Image processing device, imager, information processing device, detector, roadside unit, image processing method, and calibration method
US20240233180A1 (en) * 2023-01-10 2024-07-11 Verb Surgical Inc. Method and system for calibrating cameras
US12205328B2 (en) * 2021-07-28 2025-01-21 Htc Corporation System for tracking camera and control method thereof

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111207774B (en) * 2020-01-17 2021-12-03 山东大学 Method and system for laser-IMU external reference calibration
CN113701745B (en) * 2020-05-21 2024-03-08 杭州海康威视数字技术股份有限公司 An external parameter change detection method, device, electronic equipment and detection system
US12236616B2 (en) * 2020-09-01 2025-02-25 Georgia Tech Research Corporation Method and system for automatic extraction of virtual on-body inertial measurement units
CN112527102B (en) * 2020-11-16 2022-11-08 青岛小鸟看看科技有限公司 Head-mounted all-in-one machine system and 6DoF tracking method and device thereof
CN112330756B (en) * 2021-01-04 2021-03-23 中智行科技有限公司 Camera calibration method and device, intelligent vehicle and storage medium
CN115409955A (en) * 2021-05-26 2022-11-29 Oppo广东移动通信有限公司 Pose determination method, device, electronic device and storage medium
CN113345028B (en) * 2021-06-01 2022-04-26 亮风台(上海)信息科技有限公司 Method and equipment for determining target coordinate transformation information
CN113989377B (en) * 2021-09-23 2025-07-11 深圳市联洲国际技术有限公司 A method, device, storage medium and terminal device for extrinsic parameter calibration of a camera
CN114663519A (en) * 2022-02-18 2022-06-24 奥比中光科技集团股份有限公司 Multi-camera calibration method and device and related equipment
CN115205399A (en) * 2022-07-13 2022-10-18 深圳市优必选科技股份有限公司 Method, device, robot and storage medium for calibrating multi-objective camera without common sight
CN115601438A (en) * 2022-07-29 2023-01-13 北京易航远智科技有限公司(Cn) External parameter calibration method, device and autonomous mobile equipment
CN115508814B (en) * 2022-09-28 2025-09-23 广州高新兴机器人有限公司 Camera and lidar joint calibration method, device, medium and robot

Patent Citations (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100165116A1 (en) * 2008-12-30 2010-07-01 Industrial Technology Research Institute Camera with dynamic calibration and method thereof
US9392262B2 (en) * 2014-03-07 2016-07-12 Aquifi, Inc. System and method for 3D reconstruction using multiple multi-channel cameras
US20160012588A1 (en) * 2014-07-14 2016-01-14 Mitsubishi Electric Research Laboratories, Inc. Method for Calibrating Cameras with Non-Overlapping Views
US20160112588A1 (en) * 2014-10-16 2016-04-21 Fuji Xerox Co., Ltd. Maintenance necessity estimation apparatus and non-transitory computer readable medium
US20170249752A1 (en) * 2016-02-29 2017-08-31 Canon Kabushiki Kaisha Device for measuring position and orientation of imaging apparatus and method therefor
US20190004178A1 (en) * 2016-03-16 2019-01-03 Sony Corporation Signal processing apparatus and signal processing method
US20170359573A1 (en) * 2016-06-08 2017-12-14 SAMSUNG SDS CO., LTD., Seoul, KOREA, REPUBLIC OF; Method and apparatus for camera calibration using light source
US20190158813A1 (en) * 2016-06-10 2019-05-23 Lucid VR, Inc. Real Time Re-Calibration of Stereo Cameras
US20180124387A1 (en) * 2016-10-28 2018-05-03 Daqri, Llc Efficient augmented reality display calibration
US20180208311A1 (en) * 2017-01-23 2018-07-26 Hangzhou Zero Zero Technology Co., Ltd. System and method for omni-directional obstacle avoidance in aerial systems
US10484665B2 (en) * 2017-04-20 2019-11-19 Panasonic Intellectual Property Management Co., Ltd. Camera parameter set calculation method, recording medium, and camera parameter set calculation apparatus
US20200143603A1 (en) * 2017-07-11 2020-05-07 Canon Kabushiki Kaisha Information processing apparatus, information processing method, and non-transitory computer-readable storage medium
US20200145588A1 (en) * 2017-07-14 2020-05-07 Canon Kabushiki Kaisha Information processing apparatus presenting information, information processing method, and storage medium
US20190147625A1 (en) * 2017-11-15 2019-05-16 Magic Leap, Inc. System and methods for extrinsic calibration of cameras and diffractive optical elements
US20190243388A1 (en) * 2018-02-07 2019-08-08 Hangzhou Zero Zero Technology Co., Ltd. Unmanned aerial vehicle including an omnidirectional depth sensing and obstacle avoidance aerial system and method of operating same
US20190364206A1 (en) * 2018-05-25 2019-11-28 Aquifi, Inc. Systems and methods for multi-camera placement
US20210124174A1 (en) * 2018-07-17 2021-04-29 Sony Corporation Head mounted display, control method for head mounted display, information processor, display device, and program
US20220309761A1 (en) * 2019-12-12 2022-09-29 Guangdong Oppo Mobile Telecommunications Corp., Ltd. Target detection method, device, terminal device, and medium
US20230021721A1 (en) * 2020-01-14 2023-01-26 Kyocera Corporation Image processing device, imager, information processing device, detector, roadside unit, image processing method, and calibration method
US11360375B1 (en) * 2020-03-10 2022-06-14 Rockwell Collins, Inc. Stereoscopic camera alignment via laser projection
US20220198677A1 (en) * 2020-12-18 2022-06-23 Qualcomm Incorporated Object segmentation and feature tracking
US20220222856A1 (en) * 2021-01-13 2022-07-14 Ambarella International Lp Fixed pattern calibration for multi-view stitching
US12205328B2 (en) * 2021-07-28 2025-01-21 Htc Corporation System for tracking camera and control method thereof
US20240233180A1 (en) * 2023-01-10 2024-07-11 Verb Surgical Inc. Method and system for calibrating cameras

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20250329052A1 (en) * 2024-04-18 2025-10-23 Htc Corporation Electronic device, parameter calibration method, and non-transitory computer readable storage medium

Also Published As

Publication number Publication date
TWI883813B (en) 2025-05-11
CN118470124A (en) 2024-08-09
TW202433406A (en) 2024-08-16

Similar Documents

Publication Publication Date Title
US11830216B2 (en) Information processing apparatus, information processing method, and storage medium
JP5961945B2 (en) Image processing apparatus, projector and projector system having the image processing apparatus, image processing method, program thereof, and recording medium recording the program
KR101766603B1 (en) Image processing apparatus, image processing system, image processing method, and computer program
US9684928B2 (en) Foot tracking
JP6735592B2 (en) Image processing apparatus, control method thereof, and image processing system
US9171379B2 (en) Hybrid precision tracking
US12462422B2 (en) Calibration method and calibration apparatus
WO2017022033A1 (en) Image processing device, image processing method, and image processing program
JP2009017480A (en) Camera calibration apparatus and program thereof
US20200364900A1 (en) Point marking using virtual fiducial elements
WO2021129305A1 (en) Calibration rod testing method for optical motion capture system, device, apparatus, and storage medium
JP2004235934A (en) Calibration processing device, calibration processing method, and computer program
KR102546346B1 (en) Apparatus and method for omni-directional camera calibration
US20240265579A1 (en) Electronic device, parameter calibration method, and non-transitory computer readable storage medium
JP2001091232A (en) Three-dimensional shape measuring apparatus and method, and recording medium
US20240012238A1 (en) Tracking apparatus, method, and non-transitory computer readable storage medium thereof
JP6027952B2 (en) Augmented reality image generation system, three-dimensional shape data generation device, augmented reality presentation device, augmented reality image generation method, and program
US11166005B2 (en) Three-dimensional information acquisition system using pitching practice, and method for calculating camera parameters
JP4902564B2 (en) Marker detection and identification device and program thereof
JPH10289315A (en) Parallax calculation device and method, and distance calculation device and method
US20240087149A1 (en) Pattern-based depth mapping with extended reference image
JP7582554B2 (en) Generating complete depth data for 6-DOF video
EP2953096B1 (en) Information processing device, information processing method, system and carrier means
JP2024175823A (en) Reliability calculation device, three-dimensional information creation device, and reliability calculation method
WO2024043055A1 (en) Camera calibration device, camera calibration method, and recording medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: HTC CORPORATION, TAIWAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:HUANG, CHUNKAI;REEL/FRAME:066340/0085

Effective date: 20240125

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED