US20170006276A1 - Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle - Google Patents
Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle Download PDFInfo
- Publication number
- US20170006276A1 US20170006276A1 US15/248,294 US201615248294A US2017006276A1 US 20170006276 A1 US20170006276 A1 US 20170006276A1 US 201615248294 A US201615248294 A US 201615248294A US 2017006276 A1 US2017006276 A1 US 2017006276A1
- Authority
- US
- United States
- Prior art keywords
- image
- lens
- sensor
- processing unit
- imaging system
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/207—Image signal generators using stereoscopic image cameras using a single 2D image sensor
- H04N13/218—Image signal generators using stereoscopic image cameras using a single 2D image sensor using spatial multiplexing
-
- H04N13/0217—
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
- B60R1/23—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle with a predetermined field of view
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/31—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles providing stereoscopic vision
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R11/00—Arrangements for holding or mounting articles, not otherwise provided for
- B60R11/04—Mounting of cameras operative during drive; Arrangement of controls thereof relative to the vehicle
-
- G—PHYSICS
- G02—OPTICS
- G02B—OPTICAL ELEMENTS, SYSTEMS OR APPARATUS
- G02B13/00—Optical objectives specially designed for the purposes specified below
- G02B13/06—Panoramic objectives; So-called "sky lenses" including panoramic objectives having reflecting surfaces
-
- G06K9/00201—
-
- G06K9/00798—
-
- G06T7/002—
-
- G06T7/2006—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/56—Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
- G06V20/588—Recognition of the road, e.g. of lane markings; Recognition of the vehicle driving pattern in relation to the road
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- H04N13/0246—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N13/00—Stereoscopic video systems; Multi-view video systems; Details thereof
- H04N13/20—Image signal generators
- H04N13/204—Image signal generators using stereoscopic image cameras
- H04N13/246—Calibration of cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/107—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using stereoscopic cameras
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/10—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used
- B60R2300/108—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of camera system used using 'non-standard' camera systems, e.g. camera sensor used for additional purposes i.a. rain sensor, camera sensor split in multiple image areas
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/40—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the details of the power supply or the coupling to vehicle components
- B60R2300/402—Image calibration
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/804—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for lane monitoring
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8093—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for obstacle warning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
- G06T2207/10012—Stereo images
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30248—Vehicle exterior or interior
- G06T2207/30268—Vehicle interior
Definitions
- the invention relates to an imaging system for providing an image of the surroundings of a motor vehicle, such that this image can be displayed to the driver and simultaneously used for further evaluations by image analysis.
- U.S. Pat. No. 6,922,292 B2 discloses a wide angle imaging system for providing an image of the surroundings of a vehicle (for example a reversing image display device).
- a camera inserted into the front side or rear side of the vehicle is used together with a lens arrangement downstream thereof, in order to present the driver with a view of the surroundings imaged by the camera on a display screen in the vehicle's interior.
- the image is prepared by refracting and diffracting lenses together with distortion compensation by way of correction.
- the imaging device used has a non-uniform array of electro-optical pixels, and the display screen is provided with a uniform pixilation.
- DE 10 2004 053 416 A1 discloses a stereoscopic distance measurement system for use in a motor vehicle which has the following elements arranged in the specified sequence in the beam path of a virtual beam: a) a digital camera having an imaging optics, and having an aperture solid angle range that delimits the virtual beam emanating from the camera; b) at least one deflecting mirror element entirely or partially deflecting the virtual beam emanating from the camera; and c) at least one image field divider element having at least one divider mirror element deflecting at least a first component beam of the virtual beam emanating from the camera, the first component beam overlapping at least a second component beam of the virtual beam emanating from the camera in a restricted spatial region.
- the corresponding CCD element is split up into two separate regions lying one above another, of which a lower region is used for the driver display. This region and an upper region (invisible to the driver) are used together for a stereoscopic evaluation, both regions having the same resolution.
- an imaging system for a motor vehicle comprises a digital camera having a sensor, a first lens directing a first image onto a first region of the sensor, and a second lens directing a second image onto a second region of the sensor.
- a display screen displays to a driver of the vehicle the first image and a processing unit processes data originating from the second region.
- the lens system is configured with at least two separate lenses in order to direct separate images of the surroundings on the imaging sensor as two optically different regions, of which a first region is associated with the driver display, and a second region is associated with the data processing unit for the purpose of evaluation. It is therefore possible to adapt the regions to the different requirements, and to optimize them.
- the lenses can be optimized independently of one another for the respective use of the regions.
- the region for the driver display can be a large central region of the virtual image, directed on the sensor, of the surroundings in high resolution.
- the driver can be provided with a central, more effectively resolved image, and the other region can be optimized for digital image processing, such as detection of relative movement between the vehicle and objects exterior to the vehicle.
- the lower image segment which is preferably used for this purpose, of the image of the surroundings on the sensor permits this in a favorable way, since the observed region comprises the substratum of the road that, in turn, includes structures (lane markings, etc.) which permit simple determination of the path and/or a deviation therefrom.
- the evaluation of this image region also permits the determination of the vehicle's longitudinal and transverse speeds.
- the digital camera further comprises a third lens directing a third image onto a third region of the sensor, and the processing unit processes data originating from the second and third regions of the sensor and performing stereoscopic image evaluation.
- the lens system is configured with three independent lenses directing three separate images of the surroundings onto the imaging sensor in three different regions, of which the second and third regions are associated with the data processing unit for the purpose of stereoscopic evaluation
- the second and third lenses and image regions can be optimized for stereoscopic evaluation, in particular as concerns their distortion and spacing from one another. The spacing between the lenses is important for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings or their data from the sensor.
- the lens system is configured even with a further lens in order to direct the surroundings on the imaging sensor with the aid of an additional (fourth) optical region, and this additional region is associated with the data processing unit for the purpose of evaluations intended to detect movement
- a direction detection unit based on an image processing unit in addition to the driver display and the stereoscopic evaluation of the determination of distance.
- the processing unit may detects movement and/or direction of the vehicle by analyzing lane markings, curbs, or other structures appearing in the fourth image.
- the left and right hand lateral edge regions are suitable for the stereoscopic evaluation, and the lower edge region is suitable for detecting relative movement.
- the middle region important to the driver therefore remains in association with the driver display.
- the first region lies in the middle region of the image, for which the driver usually has the most interest, and so said region is associated and used with the driver display on a display screen. Additionally, information of interest to the driver such as, for example, data on path and distance etc, can be overlaid onto the visual display. Again, this region can be used for the image calibration and even detection of collision with an object, without the main function of the driver display being influenced.
- the second region preferably lies at the middle lower edge.
- This region has also, if appropriate, already been used in the first region, or been overlapped therewith.
- the road is included directly in front of or behind the vehicle so that the structures present there (compare above) can be used to evaluate movement, that is to say to detect the path of the vehicle by image processing.
- the third and fourth regions preferably lie respectively at the left and right hand lateral edge outside the middle region. As already described above, these regions can be used for the stereoscopic evaluation. Moreover, they are also suitable for monitoring the dead angle as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
- the lens system and/or respective lens may therefore comprise prisms and/or mirrors for increasing the effective spacing of the optical regions for the purpose of evaluation, and for reducing the distortions that occur.
- An increase in the resolution of the stereoscopy may also be achieved.
- the data processing unit may have a buffer for the sensor data such that the data can be processed sequentially, for example during the image processing for the purpose of detecting movement and path or for stereoscopic determination of distances.
- FIG. 1 is an overview schematic of a vehicle with an embodiment of a wide angle imaging system
- FIG. 2 is a schematic of the division of an imaging sensor
- FIG. 3 is a schematic of the design of an embodiment of a lens system
- FIG. 4 is a schematic of the design of an alternative embodiment of a lens system.
- FIG. 1 is a schematic of a vehicle 1 having an imaging system 2 for providing an image of the vehicle surroundings U.
- the imaging system is able to display this image to the driver on a screen and to simultaneously digitally process the image for further evaluations.
- the imaging system 2 comprises a camera 3 for imaging the surroundings U, here in particular the road S in front of or behind the vehicle 1 .
- the camera 3 comprises a lens system 4 and a sensor 5 , the lens system 4 directing multiple images of the surroundings onto the sensor 5 .
- Sensor 5 is preferably an electronic optical sensor and may be a charge coupled device (CCD) or a complimentary metal-oxide semiconductor (CMOS) image sensor, as is well known in the field of artificial vision.
- CCD charge coupled device
- CMOS complimentary metal-oxide semiconductor
- the imaging system 2 further comprises a visual display screen 6 for displaying to the driver of the vehicle the surroundings, as imaged by the camera 3 and directed onto the sensor 5 by the lens system 4 .
- the driver is thereby enabled when driving backward to search the surroundings for objects, etc., visible on the display screen in a known way.
- the imaging system 2 additionally comprises a data processing unit 7 for processing the electronic data originating from the sensor 5 for the purpose of further evaluations.
- the data processing unit 7 is “switched” logically between the sensor 5 and display 6 , and can be embodied, with appropriate suitability, by a microprocessor, which is present in any case in the onboard electronics.
- the lens system 4 is configured with four lenses 4 a , 4 b , 4 c , 4 d in order to direct images of the surroundings onto the imaging sensor 5 as four optically different regions 5 a , 5 b , 5 c , 5 d (compare FIG. 2 ).
- the first region 5 a is associated with the visual display screen 6 .
- the other regions 5 b , 5 c , 5 d are, by contrast, associated with the data processing unit 7 for digital image evaluation purposes.
- the lenses 4 a , 4 b , 4 c , 4 d can be optimized independently of one another depending on the respective use of each particular region. Characteristics that may be optimized may, for example, relate to their distortion, resolution, etc.
- the region for the driver's visual display can be a large middle region allowing for an image of the surroundings in high resolution. The driver can thereby be provided with a central, more effectively resolved image, and it is possible for the other regions to be optimized for detection of direction and movement and/or for stereoscopic evaluation by image processing.
- a vehicle 11 located behind the driver's own vehicle is shown with little or no distortion in region 5 a .
- the roadway directly behind the driver's own vehicle is illustrated in region 5 b with height distortion for the purpose of detecting obstacles by processing unit 7 .
- the region behind the driver's own vehicle is illustrated in regions 5 c and 5 d with width distortion, with in each case the same resolution for the purpose of stereoscopic evaluation performed by processing unit 7 .
- a quasi-stereoscopic evaluation is being carried out here, by means of a single camera with the aid of different camera positions (on the basis of the movement of the vehicle).
- the differences between the respective images on the two regions of the sensor can be used to determine angular differences between the observed objects, and to calculate a pixel displacement from them.
- the mutual inter-axis spacing between the two regions is known per se, and is used for the further evaluation by the processing unit.
- the two lens portions 4 a , 4 b have different optical characteristics to produce the two different images in regions 5 a and 5 b respectively.
- the first region 5 a is located in this embodiment in the middle region of the sensor 5 which is usually of most interest for the driver, and so is associated with the driver's visual display on the display screen 6 .
- the display of this image on display screen 6 can be overlaid with information of interest to the driver such as, for example, vehicle path and distance data, etc.
- This region 5 a can also be used for the image calibration, and even for detection of collision with objects, without the main function of the driver display being influenced.
- the data of region 5 a can also be processed, analyzed, and/or conditioned by the data processing unit 7 .
- the second region 5 b lies at the middle lower edge, below the first region 5 a .
- This region may, if appropriate, also already be used for the driver display together with the first region 5 a.
- the region 5 b of may include the road directly in front of or behind the vehicle, and so the structures present there (compare above) may be used to evaluate movement, that is to say to detect the path of the vehicle, by image processing.
- the lower section of the image of the surroundings, which is preferably used for this purpose, on the sensor 5 a permits this in a favorable way, since the observed region comprises the surface of the road which, in turn, includes structure (lane markings, lines, curbs, etc.) that allow a determination of the desired vehicle path or a deviation therefrom, and/or the determination of the longitudinal and transverse speeds.
- the third and fourth regions 5 c , 5 d lie respectively at the left and right lateral edge outboard of the middle regions 5 a and 5 b .
- the regions 5 c , 5 d may be used for stereoscopic evaluation, performed by processing unit 7 . Furthermore, they are also suitable for monitoring a blind spot as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
- the lens system is configured with independent lenses 4 a , 4 b , 4 c , 4 d in order to direct multiple images of the surroundings onto the three or four optically different regions 5 a , 5 b , 5 c , 5 d of the imaging sensor 5 .
- the inter-axis spacing A between the respective central axes of lenses 4 c and 4 d and between the respective regions 5 c and 5 d is significant for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings, or their data from the sensor.
- a quasi-stereoscopic evaluation may be carried out here by means of a single camera with the aid of different camera points-of-view (on the basis of the movement of the vehicle).
- the greater this inter-axis spacing A the more effective or more accurate is the subsequent evaluation.
- the spacing between the two regions 5 c , 5 d of the image of the surroundings U is used in this case to determine angular differences between the objects observed in the regions, and to calculate a pixel displacement from them.
- the spacing A between the two regions is known per se and is used for the further evaluation by processing unit 7 .
- each lens 4 ′ c , 4 ′ d comprises a pair of mirrors 9 c , 10 c and 9 d , 10 d respectively.
- the mirror pairs 9 c , 10 c and 9 d , 10 d are optically aligned to direct and focus the images onto regions 5 c and 5 d respectively.
- Other focusing lenses and/or prisms may also be employed along with the mirrors to provide the desired optical characteristics.
- the outboard placement of mirrors 9 c , 9 d results in an increase in the effective spacing A between the optical regions for the purpose of stereoscopic evaluation. An increase in the resolution of the stereoscopy may be achieved thereby.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Mechanical Engineering (AREA)
- Physics & Mathematics (AREA)
- Signal Processing (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Optics & Photonics (AREA)
- Closed-Circuit Television Systems (AREA)
- Studio Devices (AREA)
- Testing, Inspecting, Measuring Of Stereoscopic Televisions And Televisions (AREA)
Abstract
An imaging system includes a digital camera having a sensor (such as a charge coupled device), a first lens directing a first image onto a first region of the sensor, a second lens directing a second image onto a second region of the sensor, and a third lens directing a third image onto a third region of the sensor. A display screen displays to a driver of the vehicle the first image, and a processing unit performs stereoscopic image analysis on data originating from the second and third regions. A fourth lens may be used to direct a fourth image onto a fourth region of the sensor, and the processing unit performs calculations on data from the fourth region for the detection of movement of the vehicle.
Description
- This application is a Divisional claiming priority on U.S. application Ser. No. 12/695,554, filed on Jan. 28, 2010, which claims foreign priority benefits under 35 U.S.C. §119(a)-(d) to DE 10 2009 000 550.1, filed on Feb. 2, 2009, which is hereby incorporated by reference in its entirety.
- 1. Technical Field
- The invention relates to an imaging system for providing an image of the surroundings of a motor vehicle, such that this image can be displayed to the driver and simultaneously used for further evaluations by image analysis.
- 2. Background Art
- U.S. Pat. No. 6,922,292 B2 discloses a wide angle imaging system for providing an image of the surroundings of a vehicle (for example a reversing image display device). In this case, a camera inserted into the front side or rear side of the vehicle is used together with a lens arrangement downstream thereof, in order to present the driver with a view of the surroundings imaged by the camera on a display screen in the vehicle's interior. The image is prepared by refracting and diffracting lenses together with distortion compensation by way of correction. The imaging device used has a non-uniform array of electro-optical pixels, and the display screen is provided with a uniform pixilation.
- Another wide angle imaging system is disclosed in
DE 10 2004 053 416 A1 which discloses a stereoscopic distance measurement system for use in a motor vehicle which has the following elements arranged in the specified sequence in the beam path of a virtual beam: a) a digital camera having an imaging optics, and having an aperture solid angle range that delimits the virtual beam emanating from the camera; b) at least one deflecting mirror element entirely or partially deflecting the virtual beam emanating from the camera; and c) at least one image field divider element having at least one divider mirror element deflecting at least a first component beam of the virtual beam emanating from the camera, the first component beam overlapping at least a second component beam of the virtual beam emanating from the camera in a restricted spatial region. The corresponding CCD element is split up into two separate regions lying one above another, of which a lower region is used for the driver display. This region and an upper region (invisible to the driver) are used together for a stereoscopic evaluation, both regions having the same resolution. - In a disclosed embodiment of the invention, an imaging system for a motor vehicle comprises a digital camera having a sensor, a first lens directing a first image onto a first region of the sensor, and a second lens directing a second image onto a second region of the sensor. A display screen displays to a driver of the vehicle the first image and a processing unit processes data originating from the second region. The lens system is configured with at least two separate lenses in order to direct separate images of the surroundings on the imaging sensor as two optically different regions, of which a first region is associated with the driver display, and a second region is associated with the data processing unit for the purpose of evaluation. It is therefore possible to adapt the regions to the different requirements, and to optimize them. Thus, the lenses can be optimized independently of one another for the respective use of the regions. Thus, the region for the driver display can be a large central region of the virtual image, directed on the sensor, of the surroundings in high resolution. Thus, the driver can be provided with a central, more effectively resolved image, and the other region can be optimized for digital image processing, such as detection of relative movement between the vehicle and objects exterior to the vehicle.
- The lower image segment, which is preferably used for this purpose, of the image of the surroundings on the sensor permits this in a favorable way, since the observed region comprises the substratum of the road that, in turn, includes structures (lane markings, etc.) which permit simple determination of the path and/or a deviation therefrom. The evaluation of this image region also permits the determination of the vehicle's longitudinal and transverse speeds.
- In another embodiment, the digital camera further comprises a third lens directing a third image onto a third region of the sensor, and the processing unit processes data originating from the second and third regions of the sensor and performing stereoscopic image evaluation. When the lens system is configured with three independent lenses directing three separate images of the surroundings onto the imaging sensor in three different regions, of which the second and third regions are associated with the data processing unit for the purpose of stereoscopic evaluation, the second and third lenses and image regions can be optimized for stereoscopic evaluation, in particular as concerns their distortion and spacing from one another. The spacing between the lenses is important for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings or their data from the sensor.
- When the lens system is configured even with a further lens in order to direct the surroundings on the imaging sensor with the aid of an additional (fourth) optical region, and this additional region is associated with the data processing unit for the purpose of evaluations intended to detect movement, it is further possible to provide a direction detection unit based on an image processing unit in addition to the driver display and the stereoscopic evaluation of the determination of distance. The processing unit may detects movement and/or direction of the vehicle by analyzing lane markings, curbs, or other structures appearing in the fourth image.
- In particular, the left and right hand lateral edge regions are suitable for the stereoscopic evaluation, and the lower edge region is suitable for detecting relative movement. The middle region important to the driver therefore remains in association with the driver display.
- In the disclosed embodiment of the invention, the first region lies in the middle region of the image, for which the driver usually has the most interest, and so said region is associated and used with the driver display on a display screen. Additionally, information of interest to the driver such as, for example, data on path and distance etc, can be overlaid onto the visual display. Again, this region can be used for the image calibration and even detection of collision with an object, without the main function of the driver display being influenced.
- The second region preferably lies at the middle lower edge. This region has also, if appropriate, already been used in the first region, or been overlapped therewith. In this region of the image, the road is included directly in front of or behind the vehicle so that the structures present there (compare above) can be used to evaluate movement, that is to say to detect the path of the vehicle by image processing.
- The third and fourth regions preferably lie respectively at the left and right hand lateral edge outside the middle region. As already described above, these regions can be used for the stereoscopic evaluation. Moreover, they are also suitable for monitoring the dead angle as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods.
- It may therefore be advantageous for the lens system and/or respective lens to comprise prisms and/or mirrors for increasing the effective spacing of the optical regions for the purpose of evaluation, and for reducing the distortions that occur. An increase in the resolution of the stereoscopy may also be achieved.
- For evaluation of the data supplied by the sensor, the data processing unit may have a buffer for the sensor data such that the data can be processed sequentially, for example during the image processing for the purpose of detecting movement and path or for stereoscopic determination of distances.
- Further details, features and advantages of the invention emerge from the following description of exemplary embodiments with the aid of the single drawing, in which:
-
FIG. 1 is an overview schematic of a vehicle with an embodiment of a wide angle imaging system; -
FIG. 2 is a schematic of the division of an imaging sensor; -
FIG. 3 is a schematic of the design of an embodiment of a lens system; and -
FIG. 4 is a schematic of the design of an alternative embodiment of a lens system. -
FIG. 1 is a schematic of avehicle 1 having animaging system 2 for providing an image of the vehicle surroundings U. The imaging system is able to display this image to the driver on a screen and to simultaneously digitally process the image for further evaluations. - The
imaging system 2 comprises acamera 3 for imaging the surroundings U, here in particular the road S in front of or behind thevehicle 1. Thecamera 3 comprises alens system 4 and asensor 5, thelens system 4 directing multiple images of the surroundings onto thesensor 5.Sensor 5 is preferably an electronic optical sensor and may be a charge coupled device (CCD) or a complimentary metal-oxide semiconductor (CMOS) image sensor, as is well known in the field of artificial vision. - The
imaging system 2 further comprises a visual display screen 6 for displaying to the driver of the vehicle the surroundings, as imaged by thecamera 3 and directed onto thesensor 5 by thelens system 4. The driver is thereby enabled when driving backward to search the surroundings for objects, etc., visible on the display screen in a known way. - The
imaging system 2 additionally comprises adata processing unit 7 for processing the electronic data originating from thesensor 5 for the purpose of further evaluations. Thedata processing unit 7 is “switched” logically between thesensor 5 and display 6, and can be embodied, with appropriate suitability, by a microprocessor, which is present in any case in the onboard electronics. - As seen in
FIG. 3 , thelens system 4 is configured with fourlenses 4 a, 4 b, 4 c, 4 d in order to direct images of the surroundings onto theimaging sensor 5 as four optically 5 a, 5 b, 5 c, 5 d (comparedifferent regions FIG. 2 ). Thefirst region 5 a is associated with the visual display screen 6. The 5 b, 5 c, 5 d are, by contrast, associated with theother regions data processing unit 7 for digital image evaluation purposes. - Thus, the
lenses 4 a, 4 b, 4 c, 4 d can be optimized independently of one another depending on the respective use of each particular region. Characteristics that may be optimized may, for example, relate to their distortion, resolution, etc. Thus, the region for the driver's visual display can be a large middle region allowing for an image of the surroundings in high resolution. The driver can thereby be provided with a central, more effectively resolved image, and it is possible for the other regions to be optimized for detection of direction and movement and/or for stereoscopic evaluation by image processing. - As seen in
FIG. 2 , avehicle 11 located behind the driver's own vehicle is shown with little or no distortion inregion 5 a. The roadway directly behind the driver's own vehicle is illustrated inregion 5 b with height distortion for the purpose of detecting obstacles by processingunit 7. The region behind the driver's own vehicle is illustrated in 5 c and 5 d with width distortion, with in each case the same resolution for the purpose of stereoscopic evaluation performed by processingregions unit 7. - A quasi-stereoscopic evaluation is being carried out here, by means of a single camera with the aid of different camera positions (on the basis of the movement of the vehicle). The greater this inter-axis spacing between the respective central axes of the lenses, the more effective or accurate is the subsequent evaluation. The differences between the respective images on the two regions of the sensor can be used to determine angular differences between the observed objects, and to calculate a pixel displacement from them. The mutual inter-axis spacing between the two regions is known per se, and is used for the further evaluation by the processing unit.
- It is also possible to combine two of more of the lenses so as, as illustrated here in the exemplary embodiment, to combine the two central lenses 4 a and 4 b into a combined
central lens 8 a/b that consists of an upper lens portion 4 a and a lower lens portion 4 b. The two lens portions 4 a, 4 b have different optical characteristics to produce the two different images in 5 a and 5 b respectively.regions - The
first region 5 a is located in this embodiment in the middle region of thesensor 5 which is usually of most interest for the driver, and so is associated with the driver's visual display on the display screen 6. The display of this image on display screen 6 can be overlaid with information of interest to the driver such as, for example, vehicle path and distance data, etc. Thisregion 5 a can also be used for the image calibration, and even for detection of collision with objects, without the main function of the driver display being influenced. The data ofregion 5 a can also be processed, analyzed, and/or conditioned by thedata processing unit 7. - The
second region 5 b lies at the middle lower edge, below thefirst region 5 a. This region may, if appropriate, also already be used for the driver display together with thefirst region 5 a. - The
region 5 b of may include the road directly in front of or behind the vehicle, and so the structures present there (compare above) may be used to evaluate movement, that is to say to detect the path of the vehicle, by image processing. The lower section of the image of the surroundings, which is preferably used for this purpose, on thesensor 5 a permits this in a favorable way, since the observed region comprises the surface of the road which, in turn, includes structure (lane markings, lines, curbs, etc.) that allow a determination of the desired vehicle path or a deviation therefrom, and/or the determination of the longitudinal and transverse speeds. - The third and
5 c, 5 d lie respectively at the left and right lateral edge outboard of thefourth regions 5 a and 5 b. As already described above, themiddle regions 5 c, 5 d may be used for stereoscopic evaluation, performed by processingregions unit 7. Furthermore, they are also suitable for monitoring a blind spot as well as for determining a rotation (angle and angular velocity) of the vehicle by means of image processing methods. - As best seen in
FIGS. 3 and 4 , the lens system is configured withindependent lenses 4 a, 4 b, 4 c, 4 d in order to direct multiple images of the surroundings onto the three or four optically 5 a, 5 b, 5 c, 5 d of thedifferent regions imaging sensor 5. - The inter-axis spacing A between the respective central axes of
lenses 4 c and 4 d and between the 5 c and 5 d is significant for the stereoscopic evaluation with the aid of sequentially stored images of the surroundings, or their data from the sensor. A quasi-stereoscopic evaluation may be carried out here by means of a single camera with the aid of different camera points-of-view (on the basis of the movement of the vehicle). The greater this inter-axis spacing A, the more effective or more accurate is the subsequent evaluation. The spacing between the tworespective regions 5 c, 5 d of the image of the surroundings U is used in this case to determine angular differences between the objects observed in the regions, and to calculate a pixel displacement from them. The spacing A between the two regions is known per se and is used for the further evaluation byregions processing unit 7. - Referring now to
FIG. 4 , eachlens 4′c, 4′d comprises a pair of 9 c, 10 c and 9 d, 10 d respectively. The mirror pairs 9 c, 10 c and 9 d, 10 d are optically aligned to direct and focus the images ontomirrors 5 c and 5 d respectively. Other focusing lenses and/or prisms (not shown) may also be employed along with the mirrors to provide the desired optical characteristics. The outboard placement ofregions 9 c, 9 d results in an increase in the effective spacing A between the optical regions for the purpose of stereoscopic evaluation. An increase in the resolution of the stereoscopy may be achieved thereby.mirrors
Claims (19)
1. An imaging system comprising:
an electro-optical sensor;
a first lens directing an image onto the sensor;
a display screen displaying the image;
second and third inter-axially spaced lenses directing respective second and third images onto the sensor on opposite sides of the image, the second and third lenses producing purposeful width distortion of the respective images adapted for stereoscopic analysis; and
a processing unit performing stereoscopic analysis on the second and third images.
2. The imaging system of claim 1 , further comprising a fourth lens inter-axially spaced from the first, the second, and the third lenses to direct a fourth image onto a fourth region of the sensor located adjacent a third edge of the sensor, the fourth image exhibiting purposeful height distortion.
3. The imaging system of claim 2 , wherein the processing unit performs calculations on data from the fourth region to detect lateral movement of a host vehicle.
4. The imaging system of claim 3 , wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
5. The imaging system of claim 1 , wherein at least one of the lenses comprises at least two mirrors optically aligned with one another.
6. The imaging system of claim 1 , further comprising a data buffer associated with the processing unit.
7. An imaging system for a motor vehicle comprising:
an electro-optical sensor;
a first lens directing a first image onto a central region of the sensor;
a display screen for displaying the first image to a driver of the vehicle;
a second lens inter-axially spaced from the first lens to direct a second image onto a second region of the sensor located adjacent a first lateral edge of the sensor;
a third lens inter-axially spaced from the first lens and the second lens to direct a third image onto a third region of the sensor located adjacent a second lateral edge of the sensor, the second lateral edge positioned opposite from the first lateral edge, the second and third images having substantially equal amounts of purposeful width distortion adapted for stereoscopic image analysis; and
a processing unit performing stereoscopic image analysis on data originating from the second and third regions.
8. The imaging system of claim 7 , further comprising a fourth lens inter-axially spaced from the lens, the second lens, and the third lens to direct a fourth image onto the sensor, the fourth image exhibiting height distortion.
9. The imaging system of claim 8 , wherein the processing unit performs analysis on the fourth image to detect lateral movement of a host vehicle.
10. The imaging system of claim 9 , wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
11. The imaging system of claim 7 , further comprising a data buffer associated with the processing unit.
12. Apparatus comprising:
a first lens directing a first image onto a first region of an electro-optical sensor; and
second and third lenses spaced inter-axially from one another and located on opposite sides of the first lens and directing respective second and third images having substantially equal amounts of purposeful width distortion adapted for stereoscopic evaluation onto respective second and third regions of the sensor.
13. The apparatus of claim 12 , further comprising a processing unit performing stereoscopic analysis on the second and third images.
14. The apparatus of claim 13 , further comprising a data buffer associated with the processing unit.
15. The apparatus of claim 12 , further comprising a display screen for displaying the first image to a driver of a host vehicle.
16. The apparatus of claim 12 , wherein at least one of the lenses comprises at least two mirrors optically aligned with one another.
17. The apparatus of claim 12 , further comprising a fourth lens inter-axially spaced from the first lens, the second lens, and the third lens to direct a fourth image onto the sensor, the fourth image exhibiting purposeful height distortion.
18. The apparatus of claim 17 , further comprising a processing unit performing analysis on the fourth image to detect lateral movement of a host vehicle.
19. The apparatus of claim 18 , wherein the processing unit detects lateral movement of the vehicle by analyzing lane markings appearing in the fourth image.
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US15/248,294 US20170006276A1 (en) | 2009-02-02 | 2016-08-26 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| DE102009000550.1A DE102009000550B4 (en) | 2009-02-02 | 2009-02-02 | Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle |
| DE102009000550.1 | 2009-02-02 | ||
| US12/695,554 US9469249B2 (en) | 2009-02-02 | 2010-01-28 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
| US15/248,294 US20170006276A1 (en) | 2009-02-02 | 2016-08-26 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/695,554 Division US9469249B2 (en) | 2009-02-02 | 2010-01-28 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20170006276A1 true US20170006276A1 (en) | 2017-01-05 |
Family
ID=42317242
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/695,554 Expired - Fee Related US9469249B2 (en) | 2009-02-02 | 2010-01-28 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
| US15/248,294 Abandoned US20170006276A1 (en) | 2009-02-02 | 2016-08-26 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
Family Applications Before (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US12/695,554 Expired - Fee Related US9469249B2 (en) | 2009-02-02 | 2010-01-28 | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle |
Country Status (3)
| Country | Link |
|---|---|
| US (2) | US9469249B2 (en) |
| CN (1) | CN101881923B (en) |
| DE (1) | DE102009000550B4 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20230027882A1 (en) * | 2021-07-22 | 2023-01-26 | Argo AI, LLC | Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use |
Families Citing this family (17)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| DE102009000550B4 (en) * | 2009-02-02 | 2018-10-04 | Ford Global Technologies, Llc | Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle |
| JP2010263272A (en) * | 2009-04-30 | 2010-11-18 | Koito Mfg Co Ltd | Monitoring camera for vehicle and monitoring camera system for vehicle |
| JP6019998B2 (en) * | 2012-02-17 | 2016-11-02 | ソニー株式会社 | Imaging apparatus, imaging control program, and imaging method |
| WO2013126715A2 (en) | 2012-02-22 | 2013-08-29 | Magna Electronics, Inc. | Vehicle camera system with image manipulation |
| US9242602B2 (en) * | 2012-08-27 | 2016-01-26 | Fotonation Limited | Rearview imaging systems for vehicle |
| KR101428296B1 (en) | 2012-12-18 | 2014-08-08 | 현대오트론 주식회사 | Camera module for vehicle and monitoring system provided with the same |
| DE102013012216B4 (en) * | 2013-07-23 | 2016-12-29 | Audi Ag | Optical sensor arrangement for a motor vehicle and motor vehicle |
| JP6200481B2 (en) * | 2015-11-25 | 2017-09-20 | 株式会社Subaru | Outside environment recognition device |
| JP6310899B2 (en) | 2015-11-25 | 2018-04-11 | 株式会社Subaru | Outside environment recognition device |
| US10023120B2 (en) * | 2016-03-30 | 2018-07-17 | Delphi Technologies, Inc. | Multi-purpose camera device for use on a vehicle |
| CN107021016A (en) * | 2016-11-08 | 2017-08-08 | 王向阳 | Motor vehicle light guide blind-area-free panoramic display device |
| CN106791617B (en) * | 2016-12-02 | 2019-11-05 | 湖南农业大学 | Farm machinery wireless video control loop and its video switching method |
| DE112017005118A5 (en) * | 2016-12-06 | 2019-06-13 | Conti Temic Microelectronic Gmbh | Camera apparatus and method for situation-adapted detection of an environmental area of a vehicle |
| JP6837931B2 (en) * | 2017-06-26 | 2021-03-03 | アルパイン株式会社 | Display control device, display control method and camera monitoring system |
| DE102017210845A1 (en) * | 2017-06-27 | 2018-12-27 | Conti Temic Microelectronic Gmbh | Camera apparatus and method for environmental detection of an environmental area of a vehicle |
| US11341607B2 (en) * | 2019-06-07 | 2022-05-24 | Texas Instruments Incorporated | Enhanced rendering of surround view images |
| CN110667477A (en) * | 2019-11-11 | 2020-01-10 | 西南交通大学 | A vehicle night video processing system and method |
Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20060029256A1 (en) * | 2004-08-09 | 2006-02-09 | Takashi Miyoshi | Method of generating image and device |
| US20070090311A1 (en) * | 2005-10-21 | 2007-04-26 | C.R.F. Societa Consortile Per Azioni Orbassano (Torino), Italy | Optical sensor device to be installed on board a motor-vehicle for aid in driving and/or for automatic activation of systems provided on the motor-vehicle |
Family Cites Families (34)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US4457584A (en) * | 1982-03-08 | 1984-07-03 | Pryor Eugene F | Stereoscopic viewer with variable fields of vision |
| DE3441745C2 (en) * | 1984-11-15 | 1986-11-13 | Jos. Schneider Optische Werke Kreuznach GmbH & Co KG, 6550 Bad Kreuznach | Room image viewing device |
| US4989078A (en) * | 1988-08-15 | 1991-01-29 | Eastman Kodak Company | Still video camera for recording stereo images on a video disk |
| US5142357A (en) * | 1990-10-11 | 1992-08-25 | Stereographics Corp. | Stereoscopic video camera with image sensors having variable effective position |
| US5063441A (en) * | 1990-10-11 | 1991-11-05 | Stereographics Corporation | Stereoscopic video cameras with image sensors having variable effective position |
| JP3167752B2 (en) * | 1991-10-22 | 2001-05-21 | 富士重工業株式会社 | Vehicle distance detection device |
| AU5620296A (en) * | 1995-06-27 | 1997-01-09 | Yoshihisa Furuta | Device for checking lateral views at front/rear ends of vehicle |
| CA2255382A1 (en) * | 1997-12-05 | 1999-06-05 | Mcgill University | Stereoscopic gaze controller |
| US6201642B1 (en) | 1999-07-27 | 2001-03-13 | Donnelly Corporation | Vehicular vision system with a wide angle lens including a diffractive element |
| US6717610B1 (en) * | 1998-11-25 | 2004-04-06 | Donnelly Corporation | Wide angle image capture system for vehicle |
| JP2000295604A (en) * | 1999-04-06 | 2000-10-20 | Yazaki Corp | Rear and side monitoring device for vehicle |
| US6757109B2 (en) * | 1999-07-27 | 2004-06-29 | Donnelly Corporation | Plastic lens system for vehicle imaging system |
| US6963661B1 (en) * | 1999-09-09 | 2005-11-08 | Kabushiki Kaisha Toshiba | Obstacle detection system and method therefor |
| US7852462B2 (en) * | 2000-05-08 | 2010-12-14 | Automotive Technologies International, Inc. | Vehicular component control methods based on blind spot monitoring |
| EP1303140A4 (en) * | 2000-07-19 | 2007-01-17 | Matsushita Electric Industrial Co Ltd | MONITORING SYSTEM |
| JP3759429B2 (en) * | 2001-05-23 | 2006-03-22 | 株式会社東芝 | Obstacle detection apparatus and method |
| US7697027B2 (en) * | 2001-07-31 | 2010-04-13 | Donnelly Corporation | Vehicular video system |
| DE10149115A1 (en) * | 2001-10-05 | 2003-04-17 | Bosch Gmbh Robert | Object detection device for motor vehicle driver assistance systems checks data measured by sensor systems for freedom from conflict and outputs fault signal on detecting a conflict |
| US7253833B2 (en) * | 2001-11-16 | 2007-08-07 | Autonetworks Technologies, Ltd. | Vehicle periphery visual recognition system, camera and vehicle periphery monitoring apparatus and vehicle periphery monitoring system |
| JP2003186085A (en) * | 2001-12-14 | 2003-07-03 | Auto Network Gijutsu Kenkyusho:Kk | Vehicle periphery monitoring device |
| JP4052650B2 (en) * | 2004-01-23 | 2008-02-27 | 株式会社東芝 | Obstacle detection device, method and program |
| US7131740B2 (en) * | 2004-03-17 | 2006-11-07 | Olympus Corporation | Optical system and optical apparatus provided with the same |
| JP2006050263A (en) * | 2004-08-04 | 2006-02-16 | Olympus Corp | Image generation method and device |
| DE102004053416A1 (en) | 2004-11-05 | 2006-05-11 | Robert Bosch Gmbh | Stereoscopic distance measurement system to determine distance of object from motor vehicle has splitter mirror element to deflect first part of virtual beam bundle from camera which is then overlapped by second part of beam bundle |
| JP2006159933A (en) * | 2004-12-02 | 2006-06-22 | Denso Corp | Vehicle periphery display assistance system |
| DE102005043411A1 (en) | 2005-09-13 | 2007-03-15 | Robert Bosch Gmbh | Low-cost stereo optics system |
| ITTO20060214A1 (en) * | 2006-03-22 | 2007-09-23 | Kria S R L | VEHICLE DETECTION SYSTEM |
| JP2008180808A (en) * | 2007-01-23 | 2008-08-07 | Funai Electric Co Ltd | Stereoscopic imaging apparatus |
| US8358332B2 (en) * | 2007-07-23 | 2013-01-22 | Disney Enterprises, Inc. | Generation of three-dimensional movies with improved depth control |
| KR100882011B1 (en) * | 2007-07-29 | 2009-02-04 | 주식회사 나노포토닉스 | Method and apparatus for obtaining omnidirectional image using rotationally symmetrical wide angle lens |
| DE102007044536A1 (en) * | 2007-09-18 | 2009-03-19 | Bayerische Motoren Werke Aktiengesellschaft | Device for monitoring the environment of a motor vehicle |
| DE102007044535B4 (en) * | 2007-09-18 | 2022-07-14 | Bayerische Motoren Werke Aktiengesellschaft | Method for driver information in a motor vehicle |
| DE102009000550B4 (en) * | 2009-02-02 | 2018-10-04 | Ford Global Technologies, Llc | Wide-angle imaging system for providing an image of the surroundings of a vehicle, in particular of a motor vehicle |
| DE102010041490A1 (en) * | 2010-09-27 | 2012-03-29 | Carl Zeiss Microimaging Gmbh | Optical instrument and method for optical monitoring |
-
2009
- 2009-02-02 DE DE102009000550.1A patent/DE102009000550B4/en active Active
-
2010
- 2010-01-22 CN CN201010100130.4A patent/CN101881923B/en not_active Expired - Fee Related
- 2010-01-28 US US12/695,554 patent/US9469249B2/en not_active Expired - Fee Related
-
2016
- 2016-08-26 US US15/248,294 patent/US20170006276A1/en not_active Abandoned
Patent Citations (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20060029256A1 (en) * | 2004-08-09 | 2006-02-09 | Takashi Miyoshi | Method of generating image and device |
| US20070090311A1 (en) * | 2005-10-21 | 2007-04-26 | C.R.F. Societa Consortile Per Azioni Orbassano (Torino), Italy | Optical sensor device to be installed on board a motor-vehicle for aid in driving and/or for automatic activation of systems provided on the motor-vehicle |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20230027882A1 (en) * | 2021-07-22 | 2023-01-26 | Argo AI, LLC | Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use |
| US11800085B2 (en) * | 2021-07-22 | 2023-10-24 | Argo AI, LLC | Motorized mounting device for positioning an optical element within a field-of-view of an optical sensor and method of use |
| US20230344982A1 (en) * | 2021-07-22 | 2023-10-26 | Argo Al, LLC | Motorized Mounting Device for Positioning an Optical Element Within a Field-of-View of an Optical Sensor and Method of Use |
| US12101461B2 (en) * | 2021-07-22 | 2024-09-24 | Argo AI, LLC | Motorized mounting device for positioning an optical element within a field-of-view of an optical sensor and method of use |
Also Published As
| Publication number | Publication date |
|---|---|
| DE102009000550A1 (en) | 2010-08-12 |
| US9469249B2 (en) | 2016-10-18 |
| CN101881923A (en) | 2010-11-10 |
| CN101881923B (en) | 2015-09-09 |
| US20100194889A1 (en) | 2010-08-05 |
| DE102009000550B4 (en) | 2018-10-04 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US9469249B2 (en) | Wide angle imaging system for providing an image of the surroundings of a vehicle, in particular a motor vehicle | |
| US7385680B2 (en) | Camera module | |
| US10899277B2 (en) | Vehicular vision system with reduced distortion display | |
| US6611202B2 (en) | Vehicle camera display system | |
| US8233045B2 (en) | Method and apparatus for distortion correction and image enhancing of a vehicle rear viewing system | |
| US7538864B2 (en) | Vehicle wheel alignment system scanned beam imaging sensor | |
| US20160379066A1 (en) | Method and Camera System for Distance Determination of Objects from a Vehicle | |
| JP5132249B2 (en) | In-vehicle imaging device | |
| US9426364B2 (en) | Image processing apparatus and image processing method | |
| JP4425495B2 (en) | Outside monitoring device | |
| US20170140542A1 (en) | Vehicular image processing apparatus and vehicular image processing system | |
| US8477191B2 (en) | On-vehicle image pickup apparatus | |
| US10869002B2 (en) | Vehicle camera device for capturing the surroundings of a motor vehicle and driver assistance device for detecting objects with such a vehicle camera device | |
| US9238435B2 (en) | Device for monitoring an environment of a vehicle with pairs of wafer level cameras using different base distances | |
| US10635914B2 (en) | Optical test device for a vehicle camera and testing method | |
| JP2015057341A (en) | Transportation vehicle visible system for displaying statutory visibility by main mirror and wide-angle mirror | |
| EP3667413B1 (en) | Stereo image processing device | |
| US20120081509A1 (en) | Optical instrument system and method | |
| US9649990B2 (en) | Device and method for optically recording the underbody of a vehicle | |
| JP3765862B2 (en) | Vehicle environment recognition device | |
| US20050057651A1 (en) | Apparatus for visually confirming vehicle periphery | |
| CN113596389A (en) | Method for aligning a camera of a vehicle camera surveillance system with a vehicle | |
| US12101455B2 (en) | Multi-aperture zoom digital cameras and methods of using same | |
| US20220089103A1 (en) | Vehicular camera test system using true and simulated targets to determine camera defocus | |
| JP2007333525A (en) | Distance measuring device |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |