US9366759B2 - Apparatus and method for generating depth image - Google Patents
Apparatus and method for generating depth image Download PDFInfo
- Publication number
- US9366759B2 US9366759B2 US12/929,805 US92980511A US9366759B2 US 9366759 B2 US9366759 B2 US 9366759B2 US 92980511 A US92980511 A US 92980511A US 9366759 B2 US9366759 B2 US 9366759B2
- Authority
- US
- United States
- Prior art keywords
- pulse
- difference
- light
- gate
- depth
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active, expires
Links
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/254—Image signal generators using stereoscopic image cameras in combination with electromagnetic radiation sources for illuminating objects
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S7/00—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00
- G01S7/48—Details of systems according to groups G01S13/00, G01S15/00, G01S17/00 of systems according to group G01S17/00
- G01S7/483—Details of pulse systems
- G01S7/486—Receivers
- G01S7/4865—Time delay measurement, e.g. time-of-flight measurement, time of arrival measurement or determining the exact position of a peak
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/02—Systems using the reflection of electromagnetic waves other than radio waves
- G01S17/06—Systems determining position data of a target
- G01S17/08—Systems determining position data of a target for measuring distance only
- G01S17/10—Systems determining position data of a target for measuring distance only using transmission of interrupted, pulse-modulated waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S17/00—Systems using the reflection or reradiation of electromagnetic waves other than radio waves, e.g. lidar systems
- G01S17/88—Lidar systems specially adapted for specific applications
- G01S17/89—Lidar systems specially adapted for specific applications for mapping or imaging
- G01S17/894—3D imaging with simultaneous measurement of time-of-flight at a 2D array of receiver pixels, e.g. time-of-flight cameras or flash lidar
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/10—Processing, recording or transmission of stereoscopic or multi-view image signals
- H04N13/106—Processing image signals
- H04N13/128—Adjusting depth or disparity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/271—Image signal generators wherein the generated image signals comprise depth maps or disparity maps
Definitions
- Embodiments relate to a depth image generating apparatus and Method, and more particularly, to a depth image generating apparatus and method that may extend a maximal measurement distance of a depth and may maintain an accuracy of the measured depth.
- a three-dimensional (3D) image may include a color image and a depth image.
- the depth image may be obtained based on a time of flight (TOF) scheme.
- An infrared ray-based depth camera may emit an infrared ray to an object, may sense a reflected light from the object, and may measure a TOF to calculate a distance between the camera and the object. The calculated distance may be used as a depth of the depth image.
- a maximal measurement distance is limited and thus, the measured distance may often exceed the maximal measurement distance.
- the measured distance may be recorded as a smaller value than the maximal measurement distance. Therefore, sometimes, the conventional depth camera may not accurately measure a distance to the object due to the limited maximal measurement distance.
- the maximal measurement distance is extended, an accuracy of the measured distance may decrease.
- a depth image generating apparatus which may include a light-receiving unit including a first gate and a second gate, and a depth calculator to calculate a depth as the depth of the depth image based on first light information through fourth light information, and the first gate may obtain the first light information from a reflected light of a light emitted based on a first pulse and the second gate may obtain the second light information from a reflected light of a light emitted based on a second pulse, and then the first gate may obtain the third light information from a reflected light of a light emitted based on a third pulse and the second gate may obtain the fourth light information from a reflected light of a light emitted based on a fourth pulse.
- a depth calculator may calculate the depth based on a first difference and a second difference, the first difference being a difference between the first light information and the second light information and the second difference being a difference between the third light information and the fourth light information.
- a depth calculator may calculate, based on different schemes, the depth for each of a condition where both the first difference and the second difference are greater than or equal to zero, a condition where the first difference is less than zero and the second difference is greater than or equal to zero, a condition where both the first difference and the second difference are less than zero, and a condition where the first difference is greater than or equal to zero and the second difference is less than zero.
- One or more embodiments of a depth image generating apparatus may further include a pulse controller to simultaneously provide the first pulse and the second pulse to the first gate and the second gate, respectively, and then, to simultaneously provide the third pulse and the fourth pulse to the first gate and the second gate, respectively.
- a phase difference between the first pulse and the second pulse may be 180 degrees
- a phase difference between the third pulse and the fourth pulse may be 180 degrees
- a phase difference between the first pulse and the third pulse may be 90 degrees
- a phase difference between the second pulse and the fourth pulse may be 90 degrees.
- An on-off period of the first pulse may be the same as an on-off period of light emission, and the third pulse may be generated after the light emission, wherein the period of the third pulse is one-half of the period of the first pulse.
- a depth image generating apparatus of one or more embodiments may be included in a depth camera.
- the first and second gates of a depth image generating apparatus of one or more embodiments may be included in a pixel.
- At least one of a shuttering time of the first gate and a shuttering time of the second gate is adjusted to extend a maximal measurement distance and to calculate the depth of the image.
- a depth image generating method which may include obtaining first light information and second light information respectively from a reflected light of a light emitted based on a first pulse and from a reflected light of a light emitted based on a second pulse, obtaining third light information and fourth light information respectively from a reflected light of a light emitted based on a third pulse and from a reflected light of a light emitted based on a fourth pulse, and calculating a depth as the depth of the depth image based on the first light information through the fourth light information.
- Calculating based on the first light information through the fourth light information may include calculating the depth based on a first difference and a second difference, the first difference being a difference between the first light information and the second light information and the second difference being a difference between the third light information and the fourth light information.
- Calculating based on the first difference and the second difference may include calculating, based on different schemes, the depth for each of a condition where both the first difference and the second difference are greater than or equal to zero, a condition where the first difference is less than zero and the second difference is greater than or equal to zero, a condition where both the first difference and the second difference are less than zero, and a condition where the first difference is greater than or equal to zero and the second difference is less than zero.
- the first light information may include a total quantity of electric charge from the reflected light of the light emitted on the first pulse signal; the second light information comprises a total quantity of electric charge from the reflected light of the light emitted on the second pulse signal.
- the third light information may include a total quantity of electric charge from the reflected light of the light emitted on the third pulse signal; the fourth light information comprises a total quantity of electric charge from the reflected light of the light emitted on the fourth pulse signal.
- the depth may be calculated based on a difference between the first total quantity of electric charge and the second total quantity of electric charge, and based on the difference between the third total quantity of electric charge and the fourth total quantity of electric charge.
- a depth image generating method may further include adjusting at least one of a shuttering time of the first gate and a shuttering time of the second gate to extend a maximal measurement distance and to calculate the depth.
- a depth image generating method which may include measuring, during a first predetermined time, a first total quantity of electric charge in a first gate from a first reflected light of light emitted on a first pulse control signal, and a second total quantity of electric charge in a second gate from a second reflected light of a light emitted on a second pulse control signal; calculating a first difference between the first total quantity of electric charge and the second total quantity of electric charge; measuring, during a second predetermined time, a third total quantity of electric charge in the first gate from a third reflected light of a light emitted on a third pulse control signal, and a fourth total quantity of electric charge in the second gate from a fourth reflected light of a light emitted on a fourth pulse control signal; calculating a second difference between the third total quantity of electric charge and the fourth total quantity of electric charge; and calculating a maximal measurement distance of a depth as the depth of the depth image based on the first difference and the
- a depth image generating method may further include adjusting at least one of a shuttering time of the first gate and a shuttering time of the second gate to extend a maximal measurement distance and to calculate the depth.
- One or more embodiments may include a depth image generating apparatus and method.
- One or more embodiments of a depth image generating apparatus and method may enhance a maximal measurement distance of a depth more than twice a conventional one, and may enhance an accuracy of the measured depth.
- One or more embodiments of a depth image generating apparatus may calculate a time taken in a round-trip flight of a light to an object using a change in a quantity of electric charge, and may calculate the depth based on the calculated time and thus, may increase an accuracy of the depth.
- a three-dimensional (3D) image may be restored based on a depth image and a brightness image.
- the depth image and the brightness image may be generated based on a maximal measurement distance that is extended compared with a conventional one and thus, the 3D image may be more realistic and stereoscopic.
- the brightness image may be obtained by a general two-dimensional (2D) camera or a 3D camera.
- Two different brightness images having difference view points may be generated based on a depth image and a brightness image with respect to an object and may be applied to a 3D display device.
- At least one computer readable medium storing computer readable instructions to implement methods of one or more embodiments.
- FIG. 1 is a block diagram illustrating a depth image generating apparatus according to embodiments.
- FIG. 2 is a diagram illustrating a pixel array of a light-receiving unit according to embodiments.
- FIG. 3 is a diagram illustrating a configuration of a pixel in a pixel array of a light-receiving unit according to embodiments.
- FIGS. 4 and 5 are diagrams illustrating an example of a shuttering time of a G 1 and a G 2 and a movement of a quantity of electric charge according to embodiments.
- FIG. 6 is a diagram illustrating an example of an object and a depth image of the object according to embodiments.
- FIG. 7 is a diagram illustrating an example of generating a three-dimensional (3D) image according to embodiments.
- FIGS. 8A through 8D are diagrams illustrating an example of a quantity of electric charge generated based on t d according to embodiments.
- FIG. 9 is a graph illustrating an Y 1 and an Y 2 varying based on t d of FIGS. 8A through 8D .
- FIG. 10 is a flowchart illustrating a depth image generating method of a depth image generating apparatus according to embodiments.
- FIG. 1 illustrates a depth image generating apparatus 100 according to an embodiment.
- the depth image generating apparatus 100 may measure a depth of an object 10 based on a time of flight (TOF) scheme, and may generate a depth image to be used for generating a three-dimensional (3D) image of the object 10 .
- the depth-image generating apparatus 100 may be included in a depth camera.
- the depth image generating apparatus 100 may measure a distance to the object 10 , namely, a depth of a depth image, based on a pixel unit, using a TOF depth sensor.
- the depth image generating apparatus 100 may include a pulse controller 110 , a light-emitting unit 120 , a lens unit 130 , a light-receiving unit 140 , a depth calculator 150 , and a depth image generator 160 .
- the pulse controller 110 may provide various pulse control signals used for measuring the depth of the object 10 , to the light-emitting unit 120 and the light-receiving unit 140 .
- the pulse controller 110 may generate a waveform of a pulse that the light-emitting unit 120 uses to emit a light, namely, a pulse control signal, and may provide the pulse control signal to the light-emitting unit 120 .
- the pulse controller 110 may generate a waveform of a pulse that the light-receiving unit 140 uses to receive a reflected light, namely, a pulse control signal, and may provide the pulse control signal to the light-receiving unit 140 .
- the light-emitting unit 120 may emit a light to the object 10 to measure the depth of the object 10 .
- the light-emitting unit 120 may emit a light to the object 10 based on a pulse control signal provided form the pulse controller 110 .
- a reflected light reflected from the object 10 may be returned to the lens unit 130 .
- the light-emitting unit 120 may be embodied as a light emitting diode (LED) array or a laser device. Examples of emitted light may include an ultraviolet ray, a infrared ray, a visible ray, and the like.
- the lens unit 130 may collect a reflected light reflected from the object 10 , and may transmit the collected reflected light to the light-receiving unit 140 .
- the light-receiving unit 140 may transform the reflected light inputted from the lens unit 130 into electric charge to accumulate the electric charge, and may obtain light information based on an amount of the accumulated electric charge.
- the inputted reflected light may be the same as the received reflected light.
- the light information may be described with reference to FIGS. 4 and 5 .
- the light-receiving unit 140 may receive a reflected light while a waveform of a pulse is in an ‘on’ state and may accumulate electric charge.
- the light-receiving unit 140 may be embodied as a photo diode array including pixels arranged in two-dimensional (2D) array or as a photo gate array, as illustrated in FIG. 2 .
- FIG. 3 illustrates a configuration of a pixel in a pixel array of a light-receiving unit 140 according to an embodiment.
- the pixel may include a plurality of gates to be used to enable electric charge generated from a reflected light to move (through which electric charge generated from a reflected light may pass?).
- a first gate (G 1 ) and a second gate (G 2 ) may receive a reflected light from the lens unit 130 and may generate electric charge.
- the electric charge may pass through a zero gate (G 0 ), the G 1 and the G 2 , and may move to an integrating gate (IG 1 ) or an integrating gate (IG 2 ).
- IG 1 integrating gate
- IG 2 integrating gate
- Electric charge accumulated in IG 1 or IG 2 may move to a storage gate (SG 1 ) or a storage gate (SG 2 ) via a transmission gate 1 (TG 1 ) or a transmission gate 2 (TG 2 ).
- the SG 1 and the SG 2 may be a gate to accumulate electric charge, and a quantity of the accumulated electric charge may be used, by the depth calculator 150 , to calculate a depth.
- FIGS. 4 and 5 illustrate an example of a shuttering time of a G 1 and a G 2 and a movement of a quantity of electric charge according to an embodiment.
- the pulse controller 110 may provide, to the light-emitting unit 120 , a pulse control signal for light emission, to measure a depth of the object 10 .
- the light-emitting unit 120 may emit a light based on an on-off period set in the pulse control signal.
- T 0 may be a time where a light is emitted from the light-emitting unit 120 during a first period of the pulse.
- t d may be a time taken in a round-trip flight of a light, which may be a time where a light is emitted to the object 10 and a reflected light from the object 10 is projected to the light-receiving unit 140 .
- the pulse control signal provided by the pulse controller 110 is ‘on’, the light-emitting unit 120 may emit a light to the object 10 during T 0 , and when the provided pulse control signal is ‘off’, the light-emitting unit 120 may not emit a light to the object 10 .
- a 0 may denote an intensity of a light emitted by the light-emitting unit 120
- r may denote a reflexibility of the object 10
- rA 0 may denote an intensity of a reflected light arriving at the light-receiving unit 140 .
- the reflexibility may vary based on a color or a quality of the material of the object 10 .
- At least two quantities of electric charge, namely, Q 1 and Q 2 may be used to compensate for a distortion due to the reflexibility of the object 10 .
- Q 1 may denote a quantity of electric charge transformed from a reflected light that a G 1 receives.
- Q 2 may denote a quantity of electric charge transformed from a reflected light that a G 2 receives
- a shuttering time (t s ) may be a time where the G 1 or the G 2 is open during a single period. Therefore, a first and a second pulse control signal may be a signal to control t s .
- the pulse controller 110 may provide, to the G 1 of the light-receiving unit 140 , the first pulse control signal that is a control signal of a first pulse for receiving a reflected light, and may provide, to the G 2 (of the light receiving unit 140 ?), the second pulse control signal that is a control signal of a second pulse.
- the pulse controller 110 may simultaneously provide the first pulse control signal and the second pulse control signal to the G 1 and the G 2 , respectively.
- a phase difference between the second pulse control signal and the first pulse control signal may exist.
- the phase difference between the first and the second pulse control signals may be 180 degrees.
- the pulse controller 110 may emit a light, and may turn the G 1 on to enable the G 1 to receive a reflected light generated by the emitted light. Therefore, the pulse controller 110 may provide, to the G 1 , the first pulse control signal of which T 0 is the same as t s of the G 1 , which means that an on-off period of the first pulse control signal is the same as an on-off period of light emission.
- the G 1 may receive a reflected light during (T 0 -t d ), after t d .
- the IG 1 may obtain first light information (nQ 1 ) from a reflected light of a light emitted based on the first pulse control signal.
- the nQ 1 may be a total quantity of electric charge transformed from the reflected light that the G 1 receives.
- the pulse controller 110 may turn the G 1 off, when the light emission is turned off, and may simultaneously turn the G 2 on to enable the G 2 to receive a reflected light.
- the turning off of the G 1 indicates closing of the G 1
- the turning on of the G 2 indicates opening of the G 2 .
- the pulse controller 110 may provide, to the G 2 , the second pulse control signal having a phase difference of 180 degrees with the first pulse control signal.
- the IG 2 may obtain second light information (nQ 2 ) from a reflected light of a light emitted based on the second pulse control signal.
- the nQ 2 may be a total quantity of electric charge transformed from the reflected light that the G 2 receives.
- a first measurement measures during a predetermined time, the quantities of electric charge generated in the G 1 and the G 2 , namely, the nQ 1 and the nQ 2 .
- the nQ 1 and the nQ 2 may pass through IG 1 ⁇ TG 1 ⁇ SG 1 and IG 2 ⁇ TG 2 ⁇ SG 2 , respectively, and may be stored in a frame memory.
- the total quantities of electric charge stored in the frame memory may be nQ 1 and nQ 2 .
- n may denote a number of opens of G 1 or G 2 , or may denote a number of turning-‘on’s of the first pulse control signal or the second pulse control signal.
- the pulse controller 110 may operates as illustrated in FIG. 5 .
- the pulse controller 110 may provide, to the G 1 , a third pulse control signal that is a control signal of a third pulse to receive a reflected light, and may provide, to the G 2 , a fourth pulse control signal that is a control signal of a fourth pulse.
- the pulse controller 110 may simultaneously provide the third pulse control signal and the fourth pulse control signal to the G 1 and the G 2 , respectively.
- a phase difference between the third pulse control signal and the fourth control signal may exist.
- the phase difference between the third pulse control signal and the fourth control signal may be 180 degrees.
- a predetermined phase difference for example, a phase difference of 90 degrees between the first pulse control signal and the third pulse control signal may exist, and a phase difference of 90 degrees between the second pulse control signal and the fourth pulse control signal may exist.
- the pulse controller 110 may provide, to the G 1 , the third pulse control signal that turns the G 1 on
- the pulse controller 110 may provide, to the G 1 , the third pulse control signal that enables the to receive a reflected light.
- the IG 1 may obtain third light information (nQ 3 ) from a reflected light of a light emitted based on the third pulse control signal.
- the nQ 3 may be a total quantity of electric charge that is transformed from the reflected light that the G 1 receives.
- the pulse controller 110 may simultaneously turn the G 2 on to enable the G 2 to receive a reflected light, when the G 1 is turned off.
- the pulse controller 110 may provide, to the G 2 , a fourth control signal having a phase difference of 180 degrees with the third pulse control signal.
- the IG 2 may obtain fourth light information (nQ 4 ) from a reflected light of a light emitted based on the fourth pulse control signal.
- the nQ 4 may be a total quantity of electric charge that is transformed from the reflected light that the G 2 receives.
- a second measurement measures during a predetermined time, the quantities of electric charge generated in the G 1 and the G 2 , namely, the nQ 3 and the nQ 4 .
- the nQ 3 and the nQ 4 may pass through IG 1 ⁇ TG 1 ⁇ SG 1 and IG 2 ⁇ TG 2 ⁇ SG 2 , respectively, and may be stored in a frame memory.
- the total quantities of electric charge stored in the frame memory may be nQ 3 and nQ 4 .
- a maximal measurement distance of a depth may be enhanced more than twice a conventional maximal measurement distance and an accuracy of a measured depth may be maintained or enhanced.
- the depth calculator 150 may calculate a depth of the object 10 based on the nQ 1 through the nQ 4 obtained from the G 1 and the G 2 of the light-receiving unit 140 , namely, based on accumulated total quantity of electric charge.
- the depth may be calculated based on a pixel unit.
- the depth calculator 150 may calculate a first difference (Y 1 ) and a second difference (Y 2 ).
- the first difference may be a difference between the nQ 1 and the nQ 2
- the Y 2 may be a difference between the nQ 3 and the nQ 4 .
- Y 1 nQ 1 ⁇ nQ 2
- Y 2 nQ 3 ⁇ nQ 4 [Equation 1]
- nQ 1 may denote the first light information obtained by the G 1 based on the first pulse control signal
- nQ 2 may denote the second light information obtained by the G 2 based on the second pulse control signal
- nQ 3 may denote the third light information obtained by the G 1 based on the third pulse control signal
- nQ 4 may denote the fourth light information obtained by the G 2 based on the fourth pulse control signal.
- the depth calculator 150 may adaptively calculate the depth based on the calculated Y 1 and the calculated Y 2 .
- the depth calculator 150 may calculate, based on difference schemes, the depth for each of a condition where both the Y 1 and the Y 2 are greater than or equal to zero, a condition where the Y 1 is less than zero and the Y 2 is greater than or equal to zero, a condition where both the Y 1 and the Y 2 are less than zero, and a condition where the Y 1 is greater than or equal to zero and the Y 2 is less than zero.
- the depth calculator 150 may calculate t d based on Equation 2.
- t d may be a time taken in a round-trip flight of a light, which may be a time where a light is emitted to the object 10 and a reflected light from the object 10 is projected to the light-receiving unit 140 .
- the depth calculator 150 may calculate t d based on Equation 3.
- the calculator 150 may calculate t d based on Equation 4.
- the depth calculator 150 may calculate t d 0 based on Equation 5.
- the depth calculator 150 may calculate a dept of a pixel by substituting the calculated t d in Equation 6.
- the depth may denote the depth of the pixel
- C may denote a speed of an emitted light or a reflected light
- t d may denote time taken in a round-trip flight of a light that is calculated based on one of Equation 2 through Equation 5.
- C may use a speed of light, for example, 3 ⁇ 10 8 m/s, which is merely an example.
- the depth calculator 150 may calculate a depth of each of pixels of the object 10 based on the described method.
- the depth image generator unit 160 may generate a depth image corresponding to the object 10 based on the calculated depth of each of the pixels.
- a depth of a pixel which may be a distance between the depth image generating apparatus 100 and the corresponding pixel, may be a real number.
- a brightness image of n-bit may be represented by an integer number in a display device.
- the brightness image of eight-bit may be represented by an integer number in a range between 0 through 255 in the display device. Therefore, the depth image generator 160 may normalize the depth to be eight-bit.
- the normalization may be performed based on a maximal distance value that is available in measuring the object 10 .
- a maximal distance value that is available in measuring the object 10 .
- the depth image generator 160 may normalize the depth into based on eight-bit value.
- 10 m that is the maximal measurement distance is changed into 255 that is a maximal value of the eight-bit value
- 5 m may be changed into 127 or 128.
- FIG. 6 illustrates an example of the object 10 and a depth image of the object 10 according to an embodiment.
- the depth image generator 160 may apply normalization with respect to all pixels, to generate a depth image as illustrated in FIG. 6 .
- a bright portion in the depth image may indicate that a distance between the object 10 and the depth image generating apparatus 100 is relatively short, and a dark portion in the depth image may indicate that a distance between the object 10 and the depth image generating apparatus 100 is relatively long.
- FIG. 7 illustrates an example of generating a 3D image according to an embodiment.
- the 3D image may be generated based on a brightness image of the object 10 and a depth image generated by the depth image generator 160 .
- Equation 2 through Equation 5 A method of generalizing Equation 2 through Equation 5 may be described, Equation 2 through Equation 5 being used to calculate t d based on an Y 1 and an Y 2 .
- FIGS. 8A through 8D illustrate examples of a quantity of electric charge generated based on t d according to an embodiment.
- FIG. 8A illustrates quantities of electric charge generated in a G 1 and a G 2 , namely, nQ 1 through nQ 4 , when
- both the Y 1 and the Y 2 are greater than or equal to zero.
- FIG. 8B illustrates quantities of electric charge generated in a G 1 and a G 2 , namely, nQ 1 through nQ 4 , when
- FIG. 8C illustrates quantities of electric charge generated in a G 1 and a G 2 , namely, nQ 1 through nQ 4 , when
- FIG. 8D illustrates quantities of electric charge generated in a G 1 and a G 2 , namely, nQ 1 and nQ 4 , when
- the Y 1 and the Y 2 are calculated based on the quantities of electric charge of FIG. 8D , the Y 1 is greater than or equal to zero and the Y 2 is less than zero.
- FIG. 9 illustrates a graph with respect to an Y 1 and an Y 2 varying based on t d of FIGS. 8A through 8D
- FIGS. 9 , A, B, C, and D correspond to FIG. 8A , FIG. 8B , FIG. 8C , and FIG. 8D , respectively.
- the Y 1 and the Y 2 in a period A may show a change in the quantities of electric charge of FIG. 8A .
- the Y 1 decreases and the Y 2 increases in its initial period where a first pulse control signal and a second pulse control signal are provided to a G 1 and a G 2 , respectively.
- a period B both the Y 1 and the Y 2 decrease.
- a period C the Y 1 increases and the Y 2 decreases.
- both the Y 1 and the Y 2 increase. Accordingly, the Y 1 and the Y 2 vary based on t d , which may indicate that t d affects nQ 1 through nQ 4 obtained by the G 1 and the G 2 .
- Equation 2 through Equation 5 correspond to the period A through period D, respectively.
- FIG. 10 illustrates a depth image generating method of a depth image generating apparatus according to an embodiment.
- the depth image generating method may be performed by the depth image generating apparatus 100 of FIG. 1 .
- the depth image generating apparatus 100 provides a first pulse control signal to a G 1 of the light-receiving unit 140 , and provides a second pulse control signal to a G 2 , during a predetermined time.
- the first pulse control signal may have an on-off period same as a pulse provided to the light-emitting unit 120 .
- a phase difference between the first pulse control signal and the second pulse control signal may be 180 degrees. Therefore, the second pulse control signal is turned off, when the first pulse control signal is turned on.
- the depth image generating apparatus 100 obtains nQ 1 through the G 1 and obtains nQ 2 through the G 2 .
- the depth image generating apparatus 100 provides, to the G 1 , a third pulse control signal to receive a reflected light, and provide, to the G 2 , a fourth pulse control signal, during a predetermined time.
- the third pulse control signal may have a waveform that is to be turned on T 0 /2 after an emitted light.
- a phase difference between the third pulse control signal and the fourth pulse control signal may be 180 degrees. Therefore, the fourth pulse control signal may be turned off, when the third pulse control signal is turned on.
- the depth image generating apparatus 100 obtains the nQ 3 through the G 1 and obtains nQ 4 through the G 2 .
- the depth image generating apparatus 100 calculates an Y 1 and an Y 2 based on Equation 1, the Y 1 being a difference between nQ 1 and nQ 2 and the Y 2 being a difference between nQ 3 and nQ 4 .
- the Y 1 and the Y 2 may be used to calculate t d that is a time taken in round-trip flight of a light to a pixel.
- the depth image generating apparatus 100 calculates t d based on Equation 2 in operation 1040 .
- the depth image generating apparatus 100 calculates t d based on Equation 3 in operation 1050 .
- the depth image generating apparatus 100 calculates t d based on Equation 4 in operation 1060 .
- the depth image generating apparatus 100 calculates t d based on Equation 5 in operation 1070 .
- the depth image generating apparatus 100 may calculate a depth of each pixel based on the calculated t d and Equation 6 in operation 1075 .
- the depth image generating apparatus 100 normalizes the calculated depth of each pixel to generate a depth image of the object 10 .
- the depth image generating apparatus 100 and the method thereof may obtain, based on a configuration of a pixel, a quantity of electric charge generated due to a reflected light, and may calculate a depth based on the obtained quantity of electric charge.
- the depth image generating apparatus 100 and the method thereof may adjust a shuttering time of a gate, such as the G 1 and the G 2 , included in the pixel, to calculate the depth and thus, a maximal measurement distance may be extended without a complexity in operations and an accuracy of the depth may be maintained.
- Non-transitory computer-readable media including program instructions to implement various operations embodied by a computer.
- the media may also include, alone or in combination with the program instructions, data files, data structures, and the like.
- Examples of non-transitory computer-readable media include magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD ROM disks and DVDs; magneto-optical media such as optical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory (ROM), random access memory (RAM), flash memory, and the like.
- Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.
- the described hardware devices may be configured to act as one or more software modules in order to perform the operations of the above-described embodiments, or vice versa.
- the non-transitory computer-readable media may also be a distributed network, so that the program instructions are stored and executed in a distributed fashion.
- the program instructions may be executed by one or more processors or processing devices.
- the computer-readable media may also be embodied in at least one application specific integrated circuit (ASIC) or Field Programmable Gate Array (FPGA).
- ASIC application specific integrated circuit
- FPGA Field Programmable Gate Array
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Electromagnetism (AREA)
- Computer Networks & Wireless Communication (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Optical Radar Systems And Details Thereof (AREA)
- Measurement Of Optical Distance (AREA)
Abstract
Description
after emission of a light. Therefore,
after the light is emitted, the
Y1=nQ1−nQ2
Y2=nQ3−nQ4 [Equation 1]
When the Y1 and the Y2 are calculated based on the quantities of electric charge of
When the Y1 and the Y2 are calculated based on the quantities of electric charge of
When the Y1 and the Y2 are calculated based on the quantities of electric charge of
When the Y1 and the Y2 are calculated based on the quantities of electric charge of
Claims (20)
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR10-2010-0060597 | 2010-06-25 | ||
| KR1020100060597A KR101666020B1 (en) | 2010-06-25 | 2010-06-25 | Apparatus and Method for Generating Depth Image |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US20110317878A1 US20110317878A1 (en) | 2011-12-29 |
| US9366759B2 true US9366759B2 (en) | 2016-06-14 |
Family
ID=44773227
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/929,805 Active 2033-06-12 US9366759B2 (en) | 2010-06-25 | 2011-02-16 | Apparatus and method for generating depth image |
Country Status (3)
| Country | Link |
|---|---|
| US (1) | US9366759B2 (en) |
| EP (1) | EP2402784B1 (en) |
| KR (1) | KR101666020B1 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2021262987A1 (en) * | 2020-06-25 | 2021-12-30 | Smiths Detection Inc. | Systems and methods for real-time configurable backscatter scanners |
Families Citing this family (22)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| TWI470485B (en) * | 2012-03-29 | 2015-01-21 | Wistron Corp | Stylus capable of detecting pressure on a tip |
| KR101871235B1 (en) | 2012-06-05 | 2018-06-27 | 삼성전자주식회사 | Depth image generating method and apparatus, depth image processing method and apparatus |
| US10324033B2 (en) | 2012-07-20 | 2019-06-18 | Samsung Electronics Co., Ltd. | Image processing apparatus and method for correcting an error in depth |
| WO2014157847A1 (en) * | 2013-03-25 | 2014-10-02 | 엘지전자 주식회사 | Depth image obtaining device and display device using same |
| US9799117B2 (en) | 2013-09-30 | 2017-10-24 | Lenovo (Beijing) Co., Ltd. | Method for processing data and apparatus thereof |
| CN104519342B (en) | 2013-09-30 | 2017-07-21 | 联想(北京)有限公司 | A kind of image processing method and device |
| KR102158212B1 (en) * | 2014-02-05 | 2020-09-22 | 엘지전자 주식회사 | Multispectral camera and control method of the camera for detecting 3d image |
| JP6231940B2 (en) * | 2014-05-08 | 2017-11-15 | 浜松ホトニクス株式会社 | Ranging device and driving method of ranging device |
| CN105872520A (en) * | 2016-04-25 | 2016-08-17 | 京东方科技集团股份有限公司 | Display device and display method |
| KR101946941B1 (en) * | 2016-06-13 | 2019-04-29 | 엘지전자 주식회사 | Night vision image dispay apparatus |
| WO2019054099A1 (en) * | 2017-09-14 | 2019-03-21 | パナソニックIpマネジメント株式会社 | Solid-state imaging device and imaging device equipped with same |
| US10585176B2 (en) | 2017-09-19 | 2020-03-10 | Rockwell Automation Technologies, Inc. | Pulsed-based time of flight methods and system |
| US10663565B2 (en) | 2017-09-19 | 2020-05-26 | Rockwell Automation Technologies, Inc. | Pulsed-based time of flight methods and system |
| JP7228509B2 (en) * | 2017-10-18 | 2023-02-24 | ソニーセミコンダクタソリューションズ株式会社 | Identification device and electronic equipment |
| US11002836B2 (en) | 2018-05-14 | 2021-05-11 | Rockwell Automation Technologies, Inc. | Permutation of measuring capacitors in a time-of-flight sensor |
| US10996324B2 (en) | 2018-05-14 | 2021-05-04 | Rockwell Automation Technologies, Inc. | Time of flight system and method using multiple measuring sequences |
| US10969476B2 (en) | 2018-07-10 | 2021-04-06 | Rockwell Automation Technologies, Inc. | High dynamic range for sensing systems and methods |
| US10789506B2 (en) | 2018-09-24 | 2020-09-29 | Rockwell Automation Technologies, Inc. | Object intrusion detection system and method |
| JP7510815B2 (en) * | 2020-02-20 | 2024-07-04 | 浜松ホトニクス株式会社 | Optical coherence tomography |
| CN111580119B (en) * | 2020-05-29 | 2022-09-02 | Oppo广东移动通信有限公司 | Depth camera, electronic device and control method |
| DE102021107903A1 (en) * | 2021-03-29 | 2022-09-29 | Conti Temic Microelectronic Gmbh | Method and system for estimating depth information |
| CN113298778B (en) * | 2021-05-21 | 2023-04-07 | 奥比中光科技集团股份有限公司 | Depth calculation method and system based on flight time and storage medium |
Citations (10)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE19704496A1 (en) | 1996-09-05 | 1998-03-12 | Rudolf Prof Dr Ing Schwarte | Method and device for determining the phase and / or amplitude information of an electromagnetic wave |
| KR20070020005A (en) | 2004-07-30 | 2007-02-16 | 마츠시다 덴코 가부시키가이샤 | Image processing unit |
| US20090079955A1 (en) * | 2005-05-02 | 2009-03-26 | Fumi Tsunesada | Spatial information detection device and spatial information detection system using the same |
| US20090114802A1 (en) * | 2007-11-06 | 2009-05-07 | Samsung Electronics Co., Ltd. | Image generating method and apparatus |
| KR20090049322A (en) | 2007-11-13 | 2009-05-18 | 삼성전자주식회사 | Method and apparatus for generating depth image |
| EP2116864A1 (en) | 2008-05-09 | 2009-11-11 | Vrije Universiteit Brussel | TOF range finding with background radiation suppression |
| US20090284731A1 (en) | 2008-05-13 | 2009-11-19 | Samsung Electronics Co., Ltd. | Distance measuring sensor including double transfer gate and three dimensional color image sensor including the distance measuring sensor |
| WO2010013779A1 (en) | 2008-07-30 | 2010-02-04 | 国立大学法人静岡大学 | Distance image sensor and method for generating image signal by time-of-flight method |
| KR20100025228A (en) | 2008-08-27 | 2010-03-09 | 삼성전자주식회사 | Apparatus and method for obtaining a depth image |
| US20100128109A1 (en) * | 2008-11-25 | 2010-05-27 | Banks Paul S | Systems And Methods Of High Resolution Three-Dimensional Imaging |
-
2010
- 2010-06-25 KR KR1020100060597A patent/KR101666020B1/en active Active
-
2011
- 2011-02-16 US US12/929,805 patent/US9366759B2/en active Active
- 2011-02-25 EP EP11155932.4A patent/EP2402784B1/en active Active
Patent Citations (12)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE19704496A1 (en) | 1996-09-05 | 1998-03-12 | Rudolf Prof Dr Ing Schwarte | Method and device for determining the phase and / or amplitude information of an electromagnetic wave |
| KR20070020005A (en) | 2004-07-30 | 2007-02-16 | 마츠시다 덴코 가부시키가이샤 | Image processing unit |
| US20070237363A1 (en) * | 2004-07-30 | 2007-10-11 | Matsushita Electric Works, Ltd. | Image Processing Device |
| US20090079955A1 (en) * | 2005-05-02 | 2009-03-26 | Fumi Tsunesada | Spatial information detection device and spatial information detection system using the same |
| US20090114802A1 (en) * | 2007-11-06 | 2009-05-07 | Samsung Electronics Co., Ltd. | Image generating method and apparatus |
| KR20090046535A (en) | 2007-11-06 | 2009-05-11 | 삼성전자주식회사 | Image generation method and device |
| KR20090049322A (en) | 2007-11-13 | 2009-05-18 | 삼성전자주식회사 | Method and apparatus for generating depth image |
| EP2116864A1 (en) | 2008-05-09 | 2009-11-11 | Vrije Universiteit Brussel | TOF range finding with background radiation suppression |
| US20090284731A1 (en) | 2008-05-13 | 2009-11-19 | Samsung Electronics Co., Ltd. | Distance measuring sensor including double transfer gate and three dimensional color image sensor including the distance measuring sensor |
| WO2010013779A1 (en) | 2008-07-30 | 2010-02-04 | 国立大学法人静岡大学 | Distance image sensor and method for generating image signal by time-of-flight method |
| KR20100025228A (en) | 2008-08-27 | 2010-03-09 | 삼성전자주식회사 | Apparatus and method for obtaining a depth image |
| US20100128109A1 (en) * | 2008-11-25 | 2010-05-27 | Banks Paul S | Systems And Methods Of High Resolution Three-Dimensional Imaging |
Non-Patent Citations (2)
| Title |
|---|
| Extended European Search Report issued Mar. 3, 2014 in European Patent Application No. 11155932.4. |
| R. Schwarte et al., "New optical four-quadrant phase-detector integrated into a photogate array for small and precise 3D-cameras", SPIE, vol. 3023, pp. 119-128. |
Cited By (3)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| WO2021262987A1 (en) * | 2020-06-25 | 2021-12-30 | Smiths Detection Inc. | Systems and methods for real-time configurable backscatter scanners |
| US20230236141A1 (en) * | 2020-06-25 | 2023-07-27 | Joseph Bendahan | Systems and methods for real-time configurable backscatter scanners |
| US12332191B2 (en) * | 2020-06-25 | 2025-06-17 | Smiths Detection Inc. | Systems and methods for real-time configurable backscatter scanners |
Also Published As
| Publication number | Publication date |
|---|---|
| KR20120000299A (en) | 2012-01-02 |
| EP2402784A2 (en) | 2012-01-04 |
| US20110317878A1 (en) | 2011-12-29 |
| EP2402784A3 (en) | 2014-04-02 |
| EP2402784B1 (en) | 2019-08-28 |
| KR101666020B1 (en) | 2016-10-25 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9366759B2 (en) | Apparatus and method for generating depth image | |
| US8217327B2 (en) | Apparatus and method of obtaining depth image | |
| US11209528B2 (en) | Time-of-flight depth image processing systems and methods | |
| US8339582B2 (en) | Apparatus and method to correct image | |
| US10545237B2 (en) | Method and device for acquiring distance information | |
| CN108370438B (en) | Range gated depth camera assembly | |
| US9123164B2 (en) | 3D image acquisition apparatus and method of extracting depth information in 3D image acquisition apparatus | |
| US11196919B2 (en) | Image processing method, electronic apparatus, and computer-readable storage medium | |
| KR102561099B1 (en) | ToF(time of flight) capturing apparatus and method for reducing of depth distortion caused by multiple reflection thereof | |
| CN104272731B (en) | Apparatus and method for processing 3d information | |
| US8369575B2 (en) | 3D image processing method and apparatus for improving accuracy of depth measurement of an object in a region of interest | |
| US9237333B2 (en) | Method and apparatus of measuring depth information for 3D camera | |
| JP7094937B2 (en) | Built-in calibration of time-of-flight depth imaging system | |
| US10708514B2 (en) | Blending depth images obtained with multiple exposures | |
| CN105190426A (en) | Time of flight sensor binning | |
| US20130307933A1 (en) | Method of recording an image and obtaining 3d information from the image, camera system | |
| CN103959089A (en) | Depth imaging method and apparatus with adaptive illumination of an object of interest | |
| CN108663682A (en) | Barrier range-measurement system and the vehicle with it and TOF measurement method | |
| US10877238B2 (en) | Bokeh control utilizing time-of-flight sensor to estimate distances to an object | |
| US20240134053A1 (en) | Time-of-flight data generation circuitry and time-of-flight data generation method | |
| CN109146906A (en) | Image processing method and device, electronic equipment and computer readable storage medium | |
| WO2022242348A1 (en) | Dtof depth image acquisition method and apparatus, electronic device, and medium | |
| US20160065942A1 (en) | Depth image acquisition apparatus and method of acquiring depth information | |
| CN111896971A (en) | TOF sensing device and distance detection method thereof | |
| US9247124B2 (en) | Imaging apparatus, semiconductor integrated circuit, and imaging method |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: SAMSUNG ELECTRONICS CO., LTD., KOREA, REPUBLIC OF Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:KANG, BYONG MIN;LEE, KEE CHANG;KIM, SEONG JIN;REEL/FRAME:025921/0301 Effective date: 20110214 |
|
| FEPP | Fee payment procedure |
Free format text: PAYOR NUMBER ASSIGNED (ORIGINAL EVENT CODE: ASPN); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |
|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |