[go: up one dir, main page]

WO2018056802A1 - A method for estimating three-dimensional depth value from two-dimensional images - Google Patents

A method for estimating three-dimensional depth value from two-dimensional images Download PDF

Info

Publication number
WO2018056802A1
WO2018056802A1 PCT/MY2017/050057 MY2017050057W WO2018056802A1 WO 2018056802 A1 WO2018056802 A1 WO 2018056802A1 MY 2017050057 W MY2017050057 W MY 2017050057W WO 2018056802 A1 WO2018056802 A1 WO 2018056802A1
Authority
WO
WIPO (PCT)
Prior art keywords
dimensional
feature point
image
depth value
dimensional feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/MY2017/050057
Other languages
French (fr)
Inventor
Rahmita Wirza O.K. Rahmat
Ng SENG BENG
Fatimah KHALID
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Universiti Putra Malaysia (UPM)
Original Assignee
Universiti Putra Malaysia (UPM)
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Universiti Putra Malaysia (UPM) filed Critical Universiti Putra Malaysia (UPM)
Publication of WO2018056802A1 publication Critical patent/WO2018056802A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery
    • G06T7/55Depth or shape recovery from multiple images
    • G06T7/579Depth or shape recovery from multiple images from motion
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/246Analysis of motion using feature-based methods, e.g. the tracking of corners or segments
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/20Analysis of motion
    • G06T7/269Analysis of motion using gradient-based methods
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10024Color image

Definitions

  • This invention relates to a method for estimating depth value from at least two images, and more particularly to a method of three-dimensional depth value estimation from at least two images using optical flow and trigonometry method.
  • Three dimensional (3D) modeling of physical objects can be very useful in many areas, such as computer graphics and animation, robot vision, reverse engineering, and medical.
  • 3D modeling can be represented from the scratch using modeling software, or digitized from real world objects.
  • the basic information required for this representation is the x, y, and z coordinates. Further manipulation of these coordinates can deduce the objects' dimensions (width, height, and depth). Other attributes such as the models' surface colour, texture, lighting, shading and shadow contribute to a more realistic representation.
  • Conventional digitization methods utilize Coordinate Measuring Machine (CMMs) or laser scanners to obtain coordinate for each good feature from an object to a digital data. Nevertheless, both of these devices are very costly and require a certain amount of technical knowledge during usage and maintenance.
  • CCMMs Coordinate Measuring Machine
  • US Patent No. 7,342,669 B2 disclosed a 3D measurement system that uses laser-light and estimates depth information of an object by applying a triangulation method.
  • the system consists of an apparatus comprising a laser projecting, an image capturing device and a computer.
  • LEDs light-emitting diodes
  • a user can quickly acquire a precise 3D shape without using complex equipment and can also interactively check the regions whose positions are not measured during the measuring process by displaying the acquired 3D positions in real-time.
  • measurement of 3D shape of the target object becomes efficient.
  • said laser projecting device may face problems on objects with shiny surfaces or objects that do not reflect light, which include black colour and transparent surface.
  • the user may have the difficulties to verify the depth of the object.
  • US Patent No. 8,330,803 B2 disclosed a method for 3D digitization of an object, in which a plurality of camera images of the object are recorded and assembled to determine the 3D coordinates of the object.
  • An apparatus for performing the method for 3D digitization of the object comprises a projector and one or more cameras.
  • the projector projects a light pattern onto the object, in more particular white -light strips.
  • 2D feature points from the plurality of camera images of the objects are determined without human intervention.
  • the 2D point correspondences between the 2D feature points of a picture and the 2D feature points of another picture are determined. Several of these 2D point correspondences are selected, and an associated 3D transformation is determined.
  • US Patent No. 7,573,475 B2 disclosed a method of converting 2D image to 3D image. The method includes receiving a first 2D image comprising image data, where the first 2D image is captured from a first camera location. The method also includes projecting at least a portion of the first 2D image onto computer-generated geometry. The image data has depth value associated with the computer-generated geometry.
  • the system includes rendering, using the computer-generated geometry and a second camera location differs from the first camera location, a second 2D image that is stereoscopically complementary to the first 2D image, and infilling image data that absent from the second 2D image.
  • the cited patent uses two different images captured from two different locations for example image from the left and the right of the object. Then, geometry map is built for both objects and compared to each other to calculate depth for the geometry map. Geometry map is accurate to be used if the object is simple but when the object is complicated, said method becomes inaccurate to determine the depth of the object.
  • the present invention relates to a method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of placing an object on a rotatable plate; acquiring a first view of the object comprising a first image and a second image of the object, wherein the first image of the object is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1 ° to 35° relative to the first image; obtaining two-dimensional feature point coordinates by applying Good Features to Track technique and extracting colour information, simultaneously, from the first image and the second image; filtering two-dimensional feature point coordinates; applying Pyramidal Lucas-Kanade Optical Flow technique to obtain displacement magnitudes from the two-dimensional feature point coordinates; calculating World Coordinate; estimating the three-dimensional depth value of the acquired images; and implementing inverse perspective mapping algorithm on the three-dimensional feature point coordinates.
  • Fig. 1 is a flow chart of a method for estimating three-dimensional depth value from a two-dimensional image
  • Fig. 2 is diagram showing preferred arrangement of a kit for estimating three-dimensional depth value. Detailed Description of the Invention
  • the present invention relates to a method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of:
  • acquiring a first view of the object (102) comprising a first image and a second image of the object (102), wherein the first image of the object (102) is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1 ° to 35° relative to the first image;
  • zw is the three-dimensional depth value for the two-dimensional feature point coordinate (x w1 , y w1 ),
  • objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
  • (x w 2, yw2) is a matching coordinate of the two-dimensional feature point in the second image
  • a is a rotation angle
  • At least two views of the object (102) are acquired to reflect all sides of the object (102).
  • said first image is captured at an angle of 0°, 60°, 120°, 180°, 240°, 360° or a combination thereof.
  • the three-dimensional feature point coordinates from all views are merged and collected into a single set of three-dimensional feature point coordinates.
  • the views are merged by using an inverse rotation matrix with equation (8):
  • further filtering step on the single set of three-dimensional feature point coordinates is required to remove redundant points using a second filtering means.
  • the colour information is extracted using equations (2), (3) and (4):
  • Bi,j is a blue colour information of pixel (/ ' ,/)
  • Gi,j is a green colour information of pixel (/,/)
  • Ri,j is a red colour information of pixel (/,/).
  • the world coordinate is obtained using equations (5) and (6): (5)
  • (xpi, y P i) is a projected coordinate for the two-dimensional feature point coordinate i on the acquired image.
  • the inverse perspective mapping algorithm is implemented using equation (7): wherein,
  • ⁇ x'wi, y'wi) is a corrected coordinate for the two-dimensional feature point coordinate from the first image
  • (x w , 1 ywi) is the world coordinate for the two-dimensional feature point coordinate in the first image.
  • z 1 is an approximated three-dimensional depth value for the two-dimensional feature point coordinate (x w , 1 ywi),
  • objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
  • the step of filtering two-dimensional feature point coordinates using the first filtering means is by applying Euclidean distance method.
  • the estimation of three-dimensional depth value is carried out in a controlled environment includes controlling lighting and background colour.
  • the estimation of three-dimensional depth value is carried out in an open space.
  • further filtering step on the three-dimensional feature point coordinates is carried out to eliminate noise using a third filtering means, before the merging step.
  • a kit (100) mainly comprises a rotatable plate (103) and an image capture apparatus (101 ), where an object (102) is a blue toy bird which placed on the rotatable plate (103) with a distance from the image capture apparatus (101 ).
  • the blue toy bird has a width, height and depth of 134.03 mm, 94.90 mm, 35.12 mm respectively.
  • the distance from the image capture apparatus (101 ) to a centre of rotatable plate (103) is set at a length that is able to capture the whole image of the object.
  • An embodiment shows either use of two pieces of 180° half circle protractor for forming a circle or one piece of 360° protractor, which is placed underneath the rotatable plate (103) for ease of reference when measuring the angle of rotation of the rotatable plate (103).
  • the kit (100) of the present invention is for estimating three-dimensional depth value of the object (102) by carrying a method as shown in Figure 1 .
  • the kit (100) is placed in a controlled environment which includes controlling lighting and background colour.
  • the background colour is preferably black colour to eliminate reflections during image capturing.
  • a first image of the object (102) is acquired by rotating the rotatable plate (103) to a predetermined angle between 0° to 360°.
  • said predetermined angle is 0°, 60°, 120°, 180°, 240° and 360°.
  • a combination of predetermined angle may be selected for acquiring multiple images at different views of the object (102), for example 0° and 180° in a pair angles and 60° and 240°in a pair angles.
  • the pair angles is preferably a vertically opposite angles.
  • There is at least two views of the object (102) are acquired to obtain at least four images to reflect all sides of the object (102).
  • the rotatable plate (103) is rotated to an angle in a range of 1 ° to 35° relative to the predetermined angle of the first image for capturing a second image.
  • the first image and the second image are captured using the image capture apparatus (101 ), wherein the image capture apparatus (101 ) is preferably a camera and more preferably a webcam.
  • Two-dimensional feature point coordinates is obtained by applying Good Features to Track technique and the colour information is extracted from the two-dimensional feature point coordinates, simultaneously, from the first image and the second image. Said colour information is extracted using equations (2), (3), and (4) as followings:
  • Bi,j is a blue colour information of pixel (/ ' ,_/),
  • Gi,j is a green colour information of pixel (/ ' ,_/),
  • Ri,j is a red colour information of pixel (/,/).
  • the two-dimensional feature point coordinates is filtered to eliminate noise using first filtering means, where the detected two-dimensional feature point coordinates which are located further away from the centre of the images are assumed to have a higher probability of noise.
  • the two-dimensional feature point coordinates is filtered by applying Euclidean distance method.
  • detectedheight is a distance of highest point and lowest point of the object (102) detected in the acquired image, measured in pixels:
  • the three-dimensional depth value of the acquired images is estimated by calculating each of the two-dimensional feature point coordinates using equation (1 ), thereby producing three-dimensional feature point coordinates: (1) wherein,
  • z w is the three-dimensional depth value for the two-dimensional feature point coordinate (xwi, ywi),
  • objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
  • (x w , 1 ywi) is a coordinate of the two-dimensional feature point from the first image
  • (x w 2, yw2) is a matching coordinate of the two-dimensional feature point in the second image
  • a is a rotation angle
  • ⁇ x'wi, y'wi) is a corrected coordinate for the two-dimensional feature point coordinate from the first image
  • (x w , 1 ywi) is the world coordinate for the two-dimensional feature point coordinate in the first image.
  • zi is an approximated three-dimensional depth value for the two-dimensional feature point coordinate (xwi, ywi),
  • objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center).
  • the three-dimensional feature point coordinates from all individual views collected are merged into a single set of three-dimensional feature point coordinates.
  • the merging step is performed using an inverse rotation matrix with equation (8):
  • (x' y', z') are merged three-dimensional feature point coordinate after applying the inverse rotation matrix on three-dimensional feature point coordinate (x, y, z).
  • the single set of three-dimensional feature point coordinates is filtered to remove redundant points using a second filtering means.
  • the kit (100) is placed in an open space for carrying out the estimation of three-dimensional depth value of the object (102). If the kit (100) is placed in the open space, a further filtering step is required on the three-dimensional feature point coordinates to eliminate noise using a third filtering means, before the merging step of the present invention.
  • Point 2 is a matching two-dimensional feature point of Point 1 in the second image after the a° rotation.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

The present invention relates to a method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of placing an object (102) on a rotatable plate (103); acquiring a first view of the object (102) comprising a first image and a second image of the object (102), wherein the first image of the object (102) is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1° to 35° relative to the first image; obtaining two-dimensional feature point coordinates by applying Good Features to Track technique and extracting colour information, simultaneously, from the first image and the second image; filtering two-dimensional feature point coordinates; applying Pyramidal Lucas-Kanade Optical Flow technique to obtain displacement magnitudes from the two-dimensional feature point coordinates; calculating World Coordinate; estimating the three-dimensional depth value of the acquired images; and implementing inverse perspective mapping algorithm on the three-dimensional feature point coordinates.

Description

A METHOD FOR ESTIMATING THREE-DIMENSIONAL DEPTH VALUE FROM
TWO-DIMENSIONAL IMAGES
Background of the Invention Field of the Invention
This invention relates to a method for estimating depth value from at least two images, and more particularly to a method of three-dimensional depth value estimation from at least two images using optical flow and trigonometry method. Description of Related Arts
Three dimensional (3D) modeling of physical objects can be very useful in many areas, such as computer graphics and animation, robot vision, reverse engineering, and medical. 3D modeling can be represented from the scratch using modeling software, or digitized from real world objects. The basic information required for this representation is the x, y, and z coordinates. Further manipulation of these coordinates can deduce the objects' dimensions (width, height, and depth). Other attributes such as the models' surface colour, texture, lighting, shading and shadow contribute to a more realistic representation. Conventional digitization methods utilize Coordinate Measuring Machine (CMMs) or laser scanners to obtain coordinate for each good feature from an object to a digital data. Nevertheless, both of these devices are very costly and require a certain amount of technical knowledge during usage and maintenance. For example, US Patent No. 7,342,669 B2 disclosed a 3D measurement system that uses laser-light and estimates depth information of an object by applying a triangulation method. The system consists of an apparatus comprising a laser projecting, an image capturing device and a computer. The cited art also provides a method for 3D measurement using said apparatus comprises projecting a line-laser to an object where the apparatus having light-emitting diodes (LEDs) attached to a line-laser projector for estimating the position and orientation of the laser projecting device; then, capturing projected line-laser light and the LEDs at the same time using the image capturing device; then, calculating, using the computer, a 3D shape of the object from the captured image using the triangulation method; and outputting the calculated 3D shape. Meanwhile, by applying the teachings in this cited art, a user can quickly acquire a precise 3D shape without using complex equipment and can also interactively check the regions whose positions are not measured during the measuring process by displaying the acquired 3D positions in real-time. Thus, measurement of 3D shape of the target object becomes efficient. However, said laser projecting device may face problems on objects with shiny surfaces or objects that do not reflect light, which include black colour and transparent surface. Thus, the user may have the difficulties to verify the depth of the object.
Another example such as US Patent No. 8,330,803 B2 disclosed a method for 3D digitization of an object, in which a plurality of camera images of the object are recorded and assembled to determine the 3D coordinates of the object. An apparatus for performing the method for 3D digitization of the object comprises a projector and one or more cameras. The projector projects a light pattern onto the object, in more particular white -light strips. In order to improve the method for 3D digitization of the object, 2D feature points from the plurality of camera images of the objects are determined without human intervention. The 2D point correspondences between the 2D feature points of a picture and the 2D feature points of another picture are determined. Several of these 2D point correspondences are selected, and an associated 3D transformation is determined. The quality of this 3D transformation is determined with reference to the transformed 3D coordinates of the 2D feature points. Valid 3D feature points are determined therefrom. For assembling the camera images of the object, the 3D coordinates of the valid 3D feature points are use. However, the cited patent does not disclosed in detail how to calculate or determine the depth value of each features across said image-set. US Patent No. 7,573,475 B2 disclosed a method of converting 2D image to 3D image. The method includes receiving a first 2D image comprising image data, where the first 2D image is captured from a first camera location. The method also includes projecting at least a portion of the first 2D image onto computer-generated geometry. The image data has depth value associated with the computer-generated geometry. The system includes rendering, using the computer-generated geometry and a second camera location differs from the first camera location, a second 2D image that is stereoscopically complementary to the first 2D image, and infilling image data that absent from the second 2D image. The cited patent uses two different images captured from two different locations for example image from the left and the right of the object. Then, geometry map is built for both objects and compared to each other to calculate depth for the geometry map. Geometry map is accurate to be used if the object is simple but when the object is complicated, said method becomes inaccurate to determine the depth of the object.
Accordingly, it can be seen in the prior arts that there exists a need to provide a simple depth estimation technique to merge the 2D camera images into a set of 3D surface points with colour information without using high cost gadgets such as coordinate measuring machine (CMM) or laser scanner.
Summary of Invention
It is an objective of the present invention to provide to a method of three-dimensional depth value estimation from at least two images using optical flow and trigonometry method.
It is also an objective of the present invention to provide to a method of three-dimensional depth value estimation and colour information extraction from at least two images. It is yet another objective of the present invention to provide a method of three-dimensional depth value estimation in either a control environment or open space environment. Accordingly, these objectives may be achieved by following the teachings of the present invention. The present invention relates to a method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of placing an object on a rotatable plate; acquiring a first view of the object comprising a first image and a second image of the object, wherein the first image of the object is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1 ° to 35° relative to the first image; obtaining two-dimensional feature point coordinates by applying Good Features to Track technique and extracting colour information, simultaneously, from the first image and the second image; filtering two-dimensional feature point coordinates; applying Pyramidal Lucas-Kanade Optical Flow technique to obtain displacement magnitudes from the two-dimensional feature point coordinates; calculating World Coordinate; estimating the three-dimensional depth value of the acquired images; and implementing inverse perspective mapping algorithm on the three-dimensional feature point coordinates.
Brief Description of the Drawings
The features of the invention will be more readily understood and appreciated from the following detailed description when read in conjunction with the accompanying drawings of the preferred embodiment of the present invention, in which:
Fig. 1 is a flow chart of a method for estimating three-dimensional depth value from a two-dimensional image; and
Fig. 2 is diagram showing preferred arrangement of a kit for estimating three-dimensional depth value. Detailed Description of the Invention
As required, detailed embodiments of the present invention are disclosed herein; however, it is to be understood that the disclosed embodiments are merely exemplary of the invention, which may be embodied in various forms. Therefore, specific structural and functional details disclosed herein are not to be interpreted as limiting but merely as a basis for claims. It should be understood that the drawings and detailed description thereto are not intended to limit the invention to the particular form disclosed, but on the contrary, the invention is to cover all modifications, equivalents and alternatives falling within the scope of the present invention as defined by the appended claims. As used throughout this application, the word "may" is used in a permissive sense (i.e., meaning having the potential to), rather than the mandatory sense (i.e., meaning must). Similarly, the words "include," "including," and "includes" mean including, but not limited to. Further, the words "a" or "an" mean "at least one" and the word "plurality" means one or more, unless otherwise mentioned. Where the abbreviations or technical terms are used, these indicate the commonly accepted meanings as known in the technical field. For ease of reference, common reference numerals will be used throughout the figures when referring to the same or similar features common to the figures. The present invention will now be described with reference to Figs. 1 and 2.
The present invention relates to a method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of:
placing an object (102) on a rotatable plate (103) with a distance from an image capture apparatus (101 );
acquiring a first view of the object (102) comprising a first image and a second image of the object (102), wherein the first image of the object (102) is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1 ° to 35° relative to the first image;
obtaining two-dimensional feature point coordinates by applying Good
Features to Track technique and extracting colour information, simultaneously, from the first image and the second image; filtering two-dimensional feature point coordinates to eliminate noise using first filtering means;
applying Pyramidal Lucas-Kanade Optical Flow technique to obtain displacement magnitudes from the two-dimensional feature point coordinates; calculating World Coordinate for each of the two-dimensional feature point coordinates;
estimating the three-dimensional depth value of the acquired images by calculating each of the two-dimensional feature point coordinates using equation (1 ), thereby producing three-dimensional feature point coordinates:
Figure imgf000008_0001
wherein,
zw is the three-dimensional depth value for the two-dimensional feature point coordinate (xw1 , yw1 ),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
(xw1, ywi) is a coordinate of the two-dimensional feature point from the first image,
(xw 2, yw2) is a matching coordinate of the two-dimensional feature point in the second image,
a is a rotation angle; implementing inverse perspective mapping algorithm on the three-dimensional feature point coordinates to remove perspective distortion effects.
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, at least two views of the object (102) are acquired to reflect all sides of the object (102). In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, said first image is captured at an angle of 0°, 60°, 120°, 180°, 240°, 360° or a combination thereof. In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the three-dimensional feature point coordinates from all views are merged and collected into a single set of three-dimensional feature point coordinates. In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the views are merged by using an inverse rotation matrix with equation (8):
Figure imgf000009_0001
wherein,
(x' y', z') are merged three-dimensional feature point coordinate after applying the inverse rotation matrix on three-dimensional feature point coordinate (x, y, z).
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, further filtering step on the single set of three-dimensional feature point coordinates is required to remove redundant points using a second filtering means.
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the colour information is extracted using equations (2), (3) and (4):
Figure imgf000009_0002
(2)
(3) (4)
Figure imgf000010_0001
wherein,
Bi,j is a blue colour information of pixel (/' ,/),
Gi,j is a green colour information of pixel (/,/),
Ri,j is a red colour information of pixel (/,/).
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the world coordinate is obtained using equations (5) and (6): (5)
Figure imgf000010_0002
wherein,
objheight \s an actual object (102) height measured using linear callipers, detectedheight is a distance of highest point and lowest point of the object
(102) detected in the acquired image, measured in pixels:
(6) wherein,
(xw, 1 yw!) is the world coordinate for the two-dimensional feature point coordinate i,
(xpi, yPi) is a projected coordinate for the two-dimensional feature point coordinate i on the acquired image. In a preferred embodiment of the method for estimating three-dimensional depth value from a two-dimensional image, the inverse perspective mapping algorithm is implemented using equation (7):
Figure imgf000011_0001
wherein,
{x'wi, y'wi) is a corrected coordinate for the two-dimensional feature point coordinate from the first image,
(xw,1 ywi) is the world coordinate for the two-dimensional feature point coordinate in the first image.
z1 is an approximated three-dimensional depth value for the two-dimensional feature point coordinate (xw,1 ywi),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the step of filtering two-dimensional feature point coordinates using the first filtering means is by applying Euclidean distance method.
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the estimation of three-dimensional depth value is carried out in a controlled environment includes controlling lighting and background colour.
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, the estimation of three-dimensional depth value is carried out in an open space.
In a preferred embodiment of the method for estimating three-dimensional depth value from two-dimensional images, further filtering step on the three-dimensional feature point coordinates is carried out to eliminate noise using a third filtering means, before the merging step.
Examples
In an exemplary embodiment of the present invention as illustrated in Figure 2, a kit (100) mainly comprises a rotatable plate (103) and an image capture apparatus (101 ), where an object (102) is a blue toy bird which placed on the rotatable plate (103) with a distance from the image capture apparatus (101 ). In a preferred embodiment, the blue toy bird has a width, height and depth of 134.03 mm, 94.90 mm, 35.12 mm respectively. The distance from the image capture apparatus (101 ) to a centre of rotatable plate (103) is set at a length that is able to capture the whole image of the object. An embodiment shows either use of two pieces of 180° half circle protractor for forming a circle or one piece of 360° protractor, which is placed underneath the rotatable plate (103) for ease of reference when measuring the angle of rotation of the rotatable plate (103).
The kit (100) of the present invention is for estimating three-dimensional depth value of the object (102) by carrying a method as shown in Figure 1 . In an exemplary embodiment, the kit (100) is placed in a controlled environment which includes controlling lighting and background colour. The background colour is preferably black colour to eliminate reflections during image capturing. A first image of the object (102) is acquired by rotating the rotatable plate (103) to a predetermined angle between 0° to 360°. In a preferred embodiment, said predetermined angle is 0°, 60°, 120°, 180°, 240° and 360°. In another preferred embodiment, a combination of predetermined angle may be selected for acquiring multiple images at different views of the object (102), for example 0° and 180° in a pair angles and 60° and 240°in a pair angles. The pair angles is preferably a vertically opposite angles. There is at least two views of the object (102) are acquired to obtain at least four images to reflect all sides of the object (102).
Then, the rotatable plate (103) is rotated to an angle in a range of 1 ° to 35° relative to the predetermined angle of the first image for capturing a second image. In a preferred embodiment, the first image and the second image are captured using the image capture apparatus (101 ), wherein the image capture apparatus (101 ) is preferably a camera and more preferably a webcam.
Two-dimensional feature point coordinates is obtained by applying Good Features to Track technique and the colour information is extracted from the two-dimensional feature point coordinates, simultaneously, from the first image and the second image. Said colour information is extracted using equations (2), (3), and (4) as followings:
(2) (3) (4)
Figure imgf000013_0001
wherein,
Bi,j is a blue colour information of pixel (/',_/),
Gi,j is a green colour information of pixel (/',_/),
Ri,j is a red colour information of pixel (/,/).
The two-dimensional feature point coordinates is filtered to eliminate noise using first filtering means, where the detected two-dimensional feature point coordinates which are located further away from the centre of the images are assumed to have a higher probability of noise. The two-dimensional feature point coordinates is filtered by applying Euclidean distance method.
Pyramidal Lucas-Kanade optical Flow technique is applied to the two-dimensional feature point coordinates to obtain displacement magnitudes and to find matching two-dimensional feature point coordinates from the first image and second image. If any of the two-dimensional feature point coordinates from the first and second image do not match, the unmatched two-dimensional feature point coordinates are discarded. Then, a World Coordinate for each of the two-dimensional feature point coordinates is calculated using equations (5) and (6) as followings: (5)
Figure imgf000014_0003
wherein,
objheight \s an actual object (102) height measured using linear callipers,
detectedheight is a distance of highest point and lowest point of the object (102) detected in the acquired image, measured in pixels:
Figure imgf000014_0001
wherein,
(xw, 1 ywi) is the world coordinate for the two-dimensional feature point coordinate i,
(xpi, yPj) is a projected coordinate for the two-dimensional feature
Figure imgf000014_0004
point coordinate i on the acquired image.
The three-dimensional depth value of the acquired images is estimated by calculating each of the two-dimensional feature point coordinates using equation (1 ), thereby producing three-dimensional feature point coordinates:
Figure imgf000014_0002
(1) wherein,
zw is the three-dimensional depth value for the two-dimensional feature point coordinate (xwi, ywi),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
(xw,1 ywi) is a coordinate of the two-dimensional feature point from the first image,
(xw 2, yw2) is a matching coordinate of the two-dimensional feature point in the second image,
a is a rotation angle.
Then, an inverse perspective mapping algorithm is implemented on the three-dimensional feature point coordinates to remove perspective distortion effects using equation (7):
Figure imgf000015_0001
wherein,
{x'wi, y'wi) is a corrected coordinate for the two-dimensional feature point coordinate from the first image,
(xw,1 ywi) is the world coordinate for the two-dimensional feature point coordinate in the first image.
zi is an approximated three-dimensional depth value for the two-dimensional feature point coordinate (xwi, ywi),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center).
The three-dimensional feature point coordinates from all individual views collected are merged into a single set of three-dimensional feature point coordinates. The merging step is performed using an inverse rotation matrix with equation (8):
Figure imgf000015_0002
wherein, (x' y', z') are merged three-dimensional feature point coordinate after applying the inverse rotation matrix on three-dimensional feature point coordinate (x, y, z). The single set of three-dimensional feature point coordinates is filtered to remove redundant points using a second filtering means.
In another exemplary embodiment of the present invention, the kit (100) is placed in an open space for carrying out the estimation of three-dimensional depth value of the object (102). If the kit (100) is placed in the open space, a further filtering step is required on the three-dimensional feature point coordinates to eliminate noise using a third filtering means, before the merging step of the present invention. Example of calculation for three-dimensional depth value approximation
User input (known values):
objdist = 300 mm
a = 3°
objectheight = 94.90 mm
Detected two-dimensional feature point pair:
Point 1 (xP1, yP1) = (12.3456, 123.4567)
Point 2 (XP2, yP2) = (15.1234, 120.3333)
*Point 2 is a matching two-dimensional feature point of Point 1 in the second image after the a° rotation.
A minimum value and a maximum value of y is extracted from the two-dimensional feature points collected to calculate a detected height. For example, min (y) = 50.1932
max ( ) = 313.4567
detectedheight = 313.4567 = 263.2635
Figure imgf000017_0001
So, the 3D coordinate before inverse perspective projection for Point 1 is (4.4506, 44.5061 , 0.9883).
By applying the perspective distortion, inverse perspective projection is applied on the point to find the corrected coordinates,
Figure imgf000018_0002
Figure imgf000018_0003
Figure imgf000018_0001
So, the three-dimensional coordinate for Point 1 after inverse perspective projection is (4.4359, 44.3592, 0.9883). Although the present invention has been described with reference to specific embodiments, also shown in the appended figures, it will be apparent for those skilled in the art that many variations and modifications can be done within the scope of the invention as described in the specification and defined in the following claims.
Description of the reference numerals used in the accompanying drawings according to the present invention:
Figure imgf000019_0001

Claims

Claims
I/We claim: 1 . A method for estimating three-dimensional depth value from two-dimensional images, characterised by the steps of:
placing an object (102) on a rotatable plate (103) with a distance from an image capture apparatus (101 );
acquiring a first view of the object (102) comprising a first image and a second image of the object (102), wherein the first image of the object (102) is captured at an angle between 0° to 360°, and the second image is captured at an angle in a range of 1 ° to 35° relative to the first image; obtaining two-dimensional feature point coordinates by applying Good Features to Track technique and extracting colour information, simultaneously, from the first image and the second image;
filtering two-dimensional feature point coordinates to eliminate noise using first filtering means;
applying Pyramidal Lucas-Kanade Optical Flow technique to obtain displacement magnitudes from the two-dimensional feature point coordinates and to find matching two-dimensional feature point coordinates from the first image and second image;
calculating World Coordinate for each of the two-dimensional feature point coordinates;
estimating the three-dimensional depth value of the acquired images by calculating each of the two-dimensional feature point coordinates using equation (1 ), thereby producing three-dimensional feature point coordinates:
Figure imgf000020_0001
wherein,
zw is the three-dimensional depth value for the two-dimensional feature point coordinate (xw1, ywi),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center),
(xw,1 ywi) is a coordinate of the two-dimensional feature point from the first image,
(xw2 , Yw2) is a matching coordinate of the two-dimensional feature point in the second image,
a is a rotation angle; implementing inverse perspective mapping algorithm on the three-dimensional feature point coordinates to remove perspective distortion effects.
2. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein at least two views of the object (102) are acquired to reflect all sides of the object (102).
3. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 and claim 2, wherein said first image is captured at an angle of 0°, 60°, 120°, 180°, 240°, 360° or a combination thereof.
4. The method for estimating three-dimensional depth value from two-dimensional images according to claim 2, wherein the three-dimensional feature point coordinates from all views are merged and collected into a single set of three-dimensional feature point coordinates.
5. The method for estimating three-dimensional depth value from two-dimensional images according to claim 4, wherein the views are merged by using an inverse rotation matrix with equation (8):
Figure imgf000021_0001
Figure imgf000022_0001
wherein,
x' y', z') are merged three-dimensional feature point coordinate after applying the inverse rotation matrix on three-dimensional feature point coordinate (x, y, z).
6. The method for estimating three-dimensional depth value from two-dimensional images according to claim 4, wherein further filtering step on the single set of three-dimensional feature point coordinates is required to remove redundant points using a second filtering means.
7. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein the colour information is extracted using equations (2), (3) and (4):
Figure imgf000022_0002
wherein,
Bi,j is a blue colour information of pixel (/,/),
Gi,j is a green colour information of pixel (/,/),
R Rii,,jj is a red colour information of pixel (/,/).
8. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein the world coordinate is obtained using equations (5) and (6):
Figure imgf000023_0002
(5) wherein,
objheight \s an actual object (102) height measured using linear callipers,
detectedheight is a distance of highest point and lowest point of the object (102) detected in the acquired image, measured in pixels:
(6)
Figure imgf000023_0003
wherein,
(xw, 1 ywi) is the world coordinate for the two-dimensional feature point coordinate i,
(xpi, yPj) is a projected coordinate for the two-dimensional feature point coordinate i on the acquired image.
9. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein inverse perspective mapping algorithm is implemented using equation (7):
Figure imgf000023_0001
wherein,
{x'wi, y'wi) is a corrected coordinate for the two-dimensional feature point coordinate from the first image,
(xw,1 ywi) is the world coordinate for the two-dimensional feature point coordinate in the first image.
z-i is an approximated three-dimensional depth value for the two-dimensional feature point coordinate {xw1, yw1),
objdist is a distance of center of rotation (COR) from the image capture apparatus (101 ) (optic center).
10. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein the step of filtering two-dimensional feature point coordinates using the first filtering means is by applying Euclidean distance method.
1 1 . The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein the estimation of three-dimensional depth value is carried out in a controlled environment includes controlling lighting and background colour.
12. The method for estimating three-dimensional depth value from two-dimensional images according to claim 1 , wherein the estimation of three-dimensional depth value is carried out in an open space.
13. The method for estimating three-dimensional depth value from two-dimensional images according to claim 4 and claim 12, wherein further filtering step on the three-dimensional feature point coordinates is carried out to eliminate noise using a third filtering means, before the merging step.
PCT/MY2017/050057 2016-09-21 2017-09-14 A method for estimating three-dimensional depth value from two-dimensional images Ceased WO2018056802A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
MYPI2016703427 2016-09-21
MYPI2016703427 2016-09-21

Publications (1)

Publication Number Publication Date
WO2018056802A1 true WO2018056802A1 (en) 2018-03-29

Family

ID=61689941

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/MY2017/050057 Ceased WO2018056802A1 (en) 2016-09-21 2017-09-14 A method for estimating three-dimensional depth value from two-dimensional images

Country Status (1)

Country Link
WO (1) WO2018056802A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109978982A (en) * 2019-04-02 2019-07-05 广东电网有限责任公司 A kind of quick painting methods of point cloud based on inclination image
CN110533036A (en) * 2019-08-28 2019-12-03 湖南长城信息金融设备有限责任公司 A kind of bill scan image quick slant correction method and system
CN112241977A (en) * 2019-07-16 2021-01-19 北京京东乾石科技有限公司 A method and device for depth estimation of feature points
CN112771576A (en) * 2020-05-06 2021-05-07 深圳市大疆创新科技有限公司 Position information acquisition method, device and storage medium
CN114418996A (en) * 2022-01-19 2022-04-29 北京林业大学 Method for analyzing root configuration of populus tremuloides based on three-dimensional spatial angle algorithm
CN116522556A (en) * 2023-04-21 2023-08-01 国网冀北电力有限公司承德供电公司 Two-dimensional modeling-based meter loading electricity connection optimal wiring method and terminal

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4969036A (en) * 1989-03-31 1990-11-06 Bir Bhanu System for computing the self-motion of moving images devices
US20090226094A1 (en) * 2006-09-13 2009-09-10 Pioneer Corporation Image correcting device and method, and computer program
JP2009222568A (en) * 2008-03-17 2009-10-01 Konica Minolta Sensing Inc Method, device, and computer program of generating three-dimensional shape data

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4969036A (en) * 1989-03-31 1990-11-06 Bir Bhanu System for computing the self-motion of moving images devices
US20090226094A1 (en) * 2006-09-13 2009-09-10 Pioneer Corporation Image correcting device and method, and computer program
JP2009222568A (en) * 2008-03-17 2009-10-01 Konica Minolta Sensing Inc Method, device, and computer program of generating three-dimensional shape data

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109978982A (en) * 2019-04-02 2019-07-05 广东电网有限责任公司 A kind of quick painting methods of point cloud based on inclination image
CN112241977A (en) * 2019-07-16 2021-01-19 北京京东乾石科技有限公司 A method and device for depth estimation of feature points
CN110533036A (en) * 2019-08-28 2019-12-03 湖南长城信息金融设备有限责任公司 A kind of bill scan image quick slant correction method and system
CN110533036B (en) * 2019-08-28 2022-06-07 长城信息股份有限公司 Rapid inclination correction method and system for bill scanned image
CN112771576A (en) * 2020-05-06 2021-05-07 深圳市大疆创新科技有限公司 Position information acquisition method, device and storage medium
CN114418996A (en) * 2022-01-19 2022-04-29 北京林业大学 Method for analyzing root configuration of populus tremuloides based on three-dimensional spatial angle algorithm
CN116522556A (en) * 2023-04-21 2023-08-01 国网冀北电力有限公司承德供电公司 Two-dimensional modeling-based meter loading electricity connection optimal wiring method and terminal
CN116522556B (en) * 2023-04-21 2024-05-24 国网冀北电力有限公司承德供电公司 An optimal wiring method and terminal for meter installation and power connection based on two-dimensional modeling

Similar Documents

Publication Publication Date Title
CN114998499B (en) Binocular three-dimensional reconstruction method and system based on line laser galvanometer scanning
US6781618B2 (en) Hand-held 3D vision system
CN106802138B (en) A kind of three-dimensional scanning system and its scanning method
CN108876926B (en) Navigation method and system in panoramic scene and AR/VR client equipment
Furukawa et al. Accurate camera calibration from multi-view stereo and bundle adjustment
US8447099B2 (en) Forming 3D models using two images
JP5132832B1 (en) Measuring apparatus and information processing apparatus
CN106408556B (en) A Calibration Method for Small Object Measurement System Based on General Imaging Model
WO2018056802A1 (en) A method for estimating three-dimensional depth value from two-dimensional images
US20100328308A1 (en) Three Dimensional Mesh Modeling
WO2021077720A1 (en) Method, apparatus, and system for acquiring three-dimensional model of object, and electronic device
US20120177284A1 (en) Forming 3d models using multiple images
CN110827392B (en) Monocular image three-dimensional reconstruction method, system and device
US20060210146A1 (en) Creating 3D images of objects by illuminating with infrared patterns
CN107155341A (en) 3 D scanning system and framework
WO2018235163A1 (en) Calibration device, calibration chart, chart pattern generation device, and calibration method
WO2012096747A1 (en) Forming range maps using periodic illumination patterns
Ding et al. Fusing structure from motion and lidar for dense accurate depth map estimation
WO2020075252A1 (en) Information processing device, program, and information processing method
CN118918265A (en) Three-dimensional reconstruction method and system based on monocular camera and line laser
JP6285686B2 (en) Parallax image generation device
JP7033294B2 (en) Imaging system, imaging method
TWI595446B (en) Method for improving the quality of shadowed edges based on depth camera in augmented reality
JP6061631B2 (en) Measuring device, information processing device, measuring method, information processing method, and program
Popescu et al. The modelcamera

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 17853513

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 17853513

Country of ref document: EP

Kind code of ref document: A1