US20170280061A1 - Method for enhancing vehicle camera image quality - Google Patents
Method for enhancing vehicle camera image quality Download PDFInfo
- Publication number
- US20170280061A1 US20170280061A1 US15/619,630 US201715619630A US2017280061A1 US 20170280061 A1 US20170280061 A1 US 20170280061A1 US 201715619630 A US201715619630 A US 201715619630A US 2017280061 A1 US2017280061 A1 US 2017280061A1
- Authority
- US
- United States
- Prior art keywords
- vehicle
- image data
- camera
- captured
- frame
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/20—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
- B60R1/22—Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles for viewing an area outside the vehicle, e.g. the exterior of the vehicle
-
- H04N5/243—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/70—Circuitry for compensating brightness variation in the scene
- H04N23/76—Circuitry for compensating brightness variation in the scene by influencing the image signals
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R1/00—Optical viewing arrangements; Real-time viewing arrangements for drivers or passengers using optical image capturing systems, e.g. cameras or video systems specially adapted for use in or on vehicles
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/40—Image enhancement or restoration using histogram techniques
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/90—Dynamic range modification of images or parts thereof
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/61—Control of cameras or camera modules based on recognised objects
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/80—Camera processing pipelines; Components thereof
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/30—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing
- B60R2300/307—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the type of image processing virtually distinguishing relevant parts of a scene from the background of the scene
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B60—VEHICLES IN GENERAL
- B60R—VEHICLES, VEHICLE FITTINGS, OR VEHICLE PARTS, NOT OTHERWISE PROVIDED FOR
- B60R2300/00—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle
- B60R2300/80—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement
- B60R2300/8053—Details of viewing arrangements using cameras and displays, specially adapted for use in a vehicle characterised by the intended use of the viewing arrangement for bad weather conditions or night vision
Definitions
- the present invention relates to imaging systems or vision systems for vehicles.
- the present invention provides a vision system or imaging system for a vehicle that utilizes one or more cameras to capture images exterior of the vehicle, such as forwardly or rearwardly of the vehicle, and provides for enhanced image processing to detect objects in poor visibility conditions, such as in dense fog or the like.
- the vision system may enhance the image processing by amplifying the contrast in the captured images by brightness transfer function filtering and exposure stacking and tracking contrast thresholds or features within the captured images, such as on a frame-by-frame basis as the vehicle travels along a road.
- FIG. 1 is a plan view of a vehicle with a vision system and forward facing imaging sensor or camera that provides a forward exterior field of view in accordance with the present invention
- FIG. 2 shows images captured by the forward facing camera and processed by the vision system of the present invention
- FIGS. 3( a ) and 3( b ) show graphs showing histograms of luminance distribution for the vision system, with FIG. 3( b ) showing the histogram of the original image's graph of FIG. 3( a ) as spread into the highest possible dynamic range of the target system which equates to a contrast amplification of the present invention;
- FIGS. 4( a ) and 4( b ) show graphs of brightness transfer functions A ( FIG. 4( a ) ) and B ( FIG. 4( b ) ) which find use in the flow chart of FIGS. 5 and 6 , with the brightness transfer function A enhancing the brighter areas and dampening the darker ones, and with the brightness transfer function B decreasing the medium illuminated areas, and with the upper end at less of 100 , whereby the overall illumination becomes decreased by this transfer function;
- FIG. 5 shows a flow chart of the image enhancing and processing steps according the invention, and by mapping/stacking an illumination reduced image scene on top of a contrast enhanced image the dynamic range of the image increases: overexposed areas appear less bright and underexposed more bright, which leads to acknowledge details in the scene easier;
- FIG. 6 shows a flow chart of the image enhancing and processing steps according the invention as to be used in a vehicle vision system, supporting machine and human vision driver assistant algorithms;
- FIG. 7 shows an example of how images may be altered when processed according to the flow chart of FIG. 5 , whereby it becomes apparent that the process turns out more contrasts of possible objects on foggy weather conditions when comparing Image(t 0 ) and Image h (t 0 ).
- a vehicle 10 includes an imaging system or vision system 12 that includes at least one imaging sensor or camera 14 (such as forward facing camera at the front (or at the windshield) of the vehicle), which captures images exterior of and forwardly of the vehicle ( FIG. 1 ).
- an imaging system or vision system 12 that includes at least one imaging sensor or camera 14 (such as forward facing camera at the front (or at the windshield) of the vehicle), which captures images exterior of and forwardly of the vehicle ( FIG. 1 ).
- the imaging system 12 is operable to process (such as via an image processor) image data captured by the camera 14 to present these on a display or to detect objects and/or headlights of approaching vehicles and/or taillights of leading vehicles in the field of view of the camera (such as for use in an object detection system of the vehicle or collision avoidance system of the vehicle or headlamp control system of the vehicle or adaptive cruise control system of the vehicle or lane change departure warning system of the vehicle or traffic sign recognition system of the vehicle or driver assistance system of the vehicle the like).
- image data captured by the camera 14 to present these on a display or to detect objects and/or headlights of approaching vehicles and/or taillights of leading vehicles in the field of view of the camera (such as for use in an object detection system of the vehicle or collision avoidance system of the vehicle or headlamp control system of the vehicle or adaptive cruise control system of the vehicle or lane change departure warning system of the vehicle or traffic sign recognition system of the vehicle or driver assistance system of the vehicle the like).
- the image processor of the vision system 12 is operable to process captured image data, such as to detect and identify objects forward (and optionally sideward and/or rearward) of the vehicle during normal operation of the vehicle.
- objects may be difficult for the driver to see and may be difficult even for the image processor to detect, even when image processing algorithms for lens pollution detection (such as similar to that described in U.S. provisional application Ser. No. 61/616,126, filed Mar. 27, 2012, which is hereby incorporated herein by reference in its entirety) come into use.
- image processing algorithms for lens pollution detection such as similar to that described in U.S. provisional application Ser. No. 61/616,126, filed Mar. 27, 2012, which is hereby incorporated herein by reference in its entirety
- the image processor may process the image to detect objects, but, and with reference to image “B” in FIG. 2 , normal image processing may not detect the object of interest (the person and dog in this example) due to the poor visibility conditions. Typically, the object detection may not work feasibly when the image contrast falls under a certain level.
- the vision system of the present invention is operable to enhance or increase the contrast of the captured images so that any objects in the field of view of the camera are darkened to enhance the detectability of the objects by the image processor or to ease the visibility of objects to the driver of the vehicle.
- the image processor can detect the presence of the fixed and moving objects and determine if they are objects of interest to the driver of the vehicle and generate the appropriate signal responsive to such detection and determination or identification.
- the system may generate an alert to the driver or may adjust the headlamps accordingly or may display the detected object on a display screen for viewing by the driver (particularly for backup assist systems where the object is detected rearward of the vehicle during a reversing maneuver).
- the vision system can enhance detection of objects in the camera's field of view that may otherwise go undetected.
- the system may be operable to increase the contrast in the captured images responsive to a user input or to a detection or determination of a low visibility condition, such as responsive to a signal from a rain sensor or the like that is indicative of detection of a foggy condition or such as responsive to image processing of the captured images to determine that the vehicle is in foggy driving conditions (such as by utilizing aspects of the vision systems described in U.S. Pat. Nos. 4,973,844; 5,796,094; 5,877,897 and 6,353,392, which are hereby incorporated herein by reference in their entireties).
- image contrast enhancing for photographs (such as photographs taken by hand held digital cameras or astronomical telescopes or the like), and such enhancements may be done by known computer based tools for editing images.
- Today, nearly every operating system, library, presenting program and/or the like provides at least basic image editing functions.
- Professional photo editing programs like CoralDRAW®, Gimp® or Adobe Photoshop° provide a wide range of image editing and enhancing features.
- contrast enhancing is the editing of the contrast histogram. This can be used to expose objects stronger.
- the base was a physical model on fog reflection.
- infrared systems or low light amplifying systems in vehicles.
- Earlier systems have used infrared cameras alone, and some systems additionally use infrared headlights to light up the area in front of the vehicle (invisible for the human eye) which makes that area easier to detect with the infrared camera.
- Infrared cameras may provide enhanced performance in object detection in dense fog conditions due to its physical principal and the detected wave length have the intrinsic property to interfuse fog, so objects in fog can be detected and/or visualized.
- automotive driver assistance systems typically provide the driver with useful information of the vehicle's environment, including the traffic or objects in front of, to the side of and rearward of the vehicle.
- additional warnings or image overlays for highlighting hazards, especially those in the driving direction of the vehicle and in the anticipated path of travel of the vehicle.
- Obstacles or pedestrians that are in the way or path of the vehicle or tend to step into the path of the vehicle may be highlighted.
- Systems which also do active interventions such as braking or collision avoidance maneuvers are also known. For distinguishing pedestrians from other objects and for predetermining their walking direction and speed, the detected objects need to be tracked over a certain time. Also, analyzing shapes or markers of walking or standing pedestrians is known in the field of automotive vision systems and image processing.
- the objects in the captured images flow or move over successively captured images (optical flow).
- optical flow For example, external or outside objects (even stationary objects) move through the images taken from a front facing vehicle camera as the vehicle travels along the road.
- Algorithms for tracking objects under driving conditions are also known.
- the optical flow When a vehicle drives through a turn, the optical flow also behaves in a turned manner. That turn can be anticipated by the knowledge of the steering wheel's angle and a kinematic model of the vehicle's curve behavior.
- the optical flow speed directly translates from the vehicle's ground speed given by the odometer.
- known algorithms may determine the optical flow direct from the image flow without the previous mentioned input from the vehicle.
- the present invention provides enhanced image quality of visible spectrum cameras, especially the dynamic range of the resulting image, so that the driver assist system algorithms can work properly and/or display the processed image to the driver as an improvement to his or her view in such limited visibility conditions. This is achieved without the need of additional cameras using different light spectrums (such as infrared sensitive cameras or the like) or other sensors for the same purpose or high dynamic range (HDR) cameras.
- the present invention thus provides enhanced image quality in poor visibility conditions captured by a non HDR camera by amplifying the contrast details in the captured images by generating a pseudo HDR image out of current and historical image components by tone mapping.
- the system tracks the contrast thresholds/features within the captured images with respect to the image flow caused by the vehicle's movement. This process is repeated on a frame-by-frame basis to detect and identify objects in the camera's forward field of view, as can be seen in FIGS. 5 and 6 .
- the historically (previously enhanced) image (Image h (t 0 ⁇ n))) passes two individual image transfer functions and then becomes superpositioned (or mapped, merged, blended or stacked) by the currently captured frame (Image t 0 )).
- This tone mapping method is called image stacking, exposure fusion or exposure blending.
- the mapping ratio of how much of the historical image (Image h (t 0 ⁇ n))) becomes kept and how much of the current image (Image t 0 )) becomes mapped in is freely selectable between 0 and 1. In the example in FIG. 5 , 20%/80% was chosen for a data frame rate of 30 frames per second. Slower frame rates might require a shift into a stronger influence of (Image t 0 )).
- the used image enhancements shall not be limited to these shown in the example of FIGS. 5 and 6 .
- the brightness transfer function A ( FIG. 4( a ) ) enhances the brighter areas and dampens the darker ones. This equates to a histogram spread (Dynamic Range Increase), such as shown in FIGS. 3( a ) and 3( b ) , of the historically image (Image h (t 0 ⁇ n))).
- the brightness transfer function B ( FIG. 4( b ) ) decreases the medium illuminated areas of the currently captured image (Image t 0 )). The upper end is at less than 100%. The overall illumination becomes decreased by this transfer function.
- FIG. 7 shows that already after this step the discrimination between the object (person with dog) and surrounding (foggy) area is improved.
- FIG. 2 The result of this image processing and tracking of the features with respect to the optical flow and the vehicle movement is shown in principle in FIG. 2 (and discussed above).
- the algorithm based on already established image processing procedures (non-automotive, image enhancements of photographs and ‘image registration’ and the like), such as tonal value splitting/-buckling/-limiting, histogram equalization and the like, as simplified can be seen with reference to FIGS. 3( a ) and 3( b ) .
- the camera's or cameras parameters as like mounting position and viewing angle optical properties may be reflected in that (combined) look up table or in another mathematical model or table.
- the moving objects/obstacles can thus be distinguished from steady objects relative to the movement of the vehicle that is equipped with the camera system or vision system of the present invention.
- Object classification may work on further distances by feeding enhanced image data.
- Further algorithms may process the image data and may indicate hazards or the like, and/or may actively intervene to avoid collisions and the like.
- the image enhancing algorithm may find use in processing multiple camera images separate or by processing a stitched image which may be arranged as a vehicle top view image or the like.
- the imaging sensor and its photosensor array may comprise any suitable camera or sensing device, such as, for example, an array of a plurality of photosensor elements arranged in 640 columns and 480 rows (a 640 ⁇ 480 imaging array), with a respective lens focusing images onto respective portions of the array.
- the photosensor array may comprise a plurality of photosensor elements arranged in a photosensor array having rows and columns.
- the logic and control circuit of the imaging sensor may function in any known manner, such as in the manner described in U.S. Pat. Nos. 5,550,677; 5,877,897; 6,498,620; 5,670,935; 5,796,094 and/or 6,396,397, and/or U.S. provisional applications, Ser. No. 61/696,416, filed Sep.
- WO 2013/019707 and/or PCT Application No. PCT/US2012/048110, filed Jul. 25, 2012, and published Jan. 31, 2013 as International Publication No. WO 2013/016409, and/or U.S. patent application Ser. No. 13/534,657, filed Jun. 27, 2012, and published Jan. 3, 2013 as U.S. Publication No. US-2013-0002873, which are all hereby incorporated herein by reference in their entireties.
- the system may communicate with other communication systems via any suitable means, such as by utilizing aspects of the systems described in PCT Application No. PCT/US10/038477, filed Jun. 14, 2010, and/or U.S. patent application Ser. No. 13/202,005, filed Aug. 17, 2011, now U.S. Pat.
- the imaging device and control and image processor and any associated illumination source may comprise any suitable components, and may utilize aspects of the cameras and vision systems described in U.S. Pat. Nos. 5,550,677; 5,877,897; 6,498,620; 5,670,935; 5,796,094; 6,396,397; 6,806,452; 6,690,268; 7,005,974; 7,123,168; 7,004,606; 6,946,978; 7,038,577; 6,353,392; 6,320,176; 6,313,454 and 6,824,281, and/or International Publication No. WO 2010/099416, published Sep. 2, 2010, and/or PCT Application No. PCT/US10/47256, filed Aug.
- the camera or cameras may comprise any suitable cameras or imaging sensors or camera modules, and may utilize aspects of the cameras or sensors described in U.S.
- the imaging array sensor may comprise any suitable sensor, and may utilize various imaging sensors or imaging array sensors or cameras or the like, such as a CMOS imaging array sensor, a CCD sensor or other sensors or the like, such as the types described in U.S. Pat. Nos.
- the camera module and circuit chip or board and imaging sensor may be implemented and operated in connection with various vehicular vision-based systems, and/or may be operable utilizing the principles of such other vehicular systems, such as a vehicle headlamp control system, such as the type disclosed in U.S. Pat. Nos. 5,796,094; 6,097,023; 6,320,176; 6,559,435; 6,831,261; 7,004,606; 7,339,149 and/or 7,526,103, which are all hereby incorporated herein by reference in their entireties, a rain sensor, such as the types disclosed in commonly assigned U.S. Pat. Nos.
- a vehicle vision system such as a forwardly, sidewardly or rearwardly directed vehicle vision system utilizing principles disclosed in U.S. Pat. Nos.
- a reverse or sideward imaging system such as for a lane change assistance system or lane departure warning system or for a blind spot or object detection system, such as imaging or detection systems of the types disclosed in U.S. Pat. Nos. 7,881,496; 7,720,580; 7,038,577; 5,929,786 and/or 5,786,772, and/or U.S. provisional applications, Ser. No. 60/628,709, filed Nov. 17, 2004; Ser. No. 60/614,644, filed Sep. 30, 2004; Ser. No. 60/618,686, filed Oct. 14, 2004; Ser. No. 60/638,687, filed Dec.
- a video device for internal cabin surveillance and/or video telephone function such as disclosed in U.S. Pat. Nos. 5,760,962; 5,877,897; 6,690,268 and/or 7,370,983, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are hereby incorporated herein by reference in their entireties, a traffic sign recognition system, a system for determining a distance to a leading or trailing vehicle or object, such as a system utilizing the principles disclosed in U.S. Pat. Nos. 6,396,397 and/or 7,123,168, which are hereby incorporated herein by reference in their entireties, and/or the like.
- the circuit board or chip may include circuitry for the imaging array sensor and or other electronic accessories or features, such as by utilizing compass-on-a-chip or EC driver-on-a-chip technology and aspects such as described in U.S. Pat. Nos. 7,255,451 and/or 7,480,149; and/or U.S. patent applications, Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US-2006-0061008, and/or Ser. No. 12/578,732, filed Oct. 14, 2009, now U.S. Pat. No. 9,487,144, which are hereby incorporated herein by reference in their entireties.
- the vision system may include a display for displaying images captured by one or more of the imaging sensors for viewing by the driver of the vehicle while the driver is normally operating the vehicle.
- the vision system may include a video display device disposed at or in the interior rearview mirror assembly of the vehicle, such as by utilizing aspects of the video mirror display systems described in U.S. Pat. No. 6,690,268 and/or U.S. patent application Ser. No. 13/333,337, filed Dec. 21, 2011, now U.S. Pat. No. 9,264,672, which are hereby incorporated herein by reference in their entireties.
- the video mirror display may comprise any suitable devices and systems and optionally may utilize aspects of the compass display systems described in U.S. Pat. Nos.
- the video mirror display screen or device may be operable to display images captured by a rearward viewing camera of the vehicle during a reversing maneuver of the vehicle (such as responsive to the vehicle gear actuator being placed in a reverse gear position or the like) to assist the driver in backing up the vehicle, and optionally may be operable to display the compass heading or directional heading character or icon when the vehicle is not undertaking a reversing maneuver, such as when the vehicle is being driven in a forward direction along a road (such as by utilizing aspects of the display system described in PCT Application No. PCT/US2011/056295, filed Oct. 14, 2011 and published Apr. 19, 2012 as International Publication No. WO 2012/051500, which is hereby incorporated herein by reference in its entirety).
- the vision system (utilizing the forward facing camera and a rearward facing camera and other cameras disposed at the vehicle with exterior fields of view) may be part of or may provide a display of a top-down view or birds-eye view system of the vehicle or a surround view at the vehicle, such as by utilizing aspects of the vision systems described in PCT Application No. PCT/US10/25545, filed Feb. 26, 2010 and published on Sep. 2, 2010 as International Publication No. WO 2010/099416, and/or PCT Application No. PCT/US10/47256, filed Aug. 31, 2010 and published Mar. 10, 2011 as International Publication No. WO 2011/028686, and/or PCT Application No. PCT/US11/62755, filed Dec. 1, 2011 and published Jun.
- the video mirror display may be disposed rearward of and behind the reflective element assembly and may comprise a display such as the types disclosed in U.S. Pat. Nos. 5,530,240; 6,329,925; 7,855,755; 7,626,749; 7,581,859; 7,338,177; 7,274,501; 7,255,451; 7,195,381; 7,184,190; 5,668,663; 5,724,187 and/or 6,690,268, and/or in U.S. patent applications, Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US-2006-0061008; and/or Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar.
- a display such as the types disclosed in U.S. Pat. Nos. 5,530,240; 6,329,925; 7,855,755; 7,626,749; 7,581,859; 7,338,177; 7,274,501; 7,255,451; 7,195,381
- the display is viewable through the reflective element when the display is activated to display information.
- the display element may be any type of display element, such as a vacuum fluorescent (VF) display element, a light emitting diode (LED) display element, such as an organic light emitting diode (OLED) or an inorganic light emitting diode, an electroluminescent (EL) display element, a liquid crystal display (LCD) element, a video screen display element or backlit thin film transistor (TFT) display element or the like, and may be operable to display various information (as discrete characters, icons or the like, or in a multi-pixel manner) to the driver of the vehicle, such as passenger side inflatable restraint (PSIR) information, tire pressure status, and/or the like.
- PSIR passenger side inflatable restraint
- the mirror assembly and/or display may utilize aspects described in U.S. Pat. Nos. 7,184,190; 7,255,451; 7,446,924 and/or 7,338,177, which are all hereby incorporated herein by reference in their entireties.
- the thicknesses and materials of the coatings on the substrates of the reflective element may be selected to provide a desired color or tint to the mirror reflective element, such as a blue colored reflector, such as is known in the art and such as described in U.S. Pat. Nos. 5,910,854; 6,420,036 and/or 7,274,501, which are hereby incorporated herein by reference in their entireties.
- the display or displays and any associated user inputs may be associated with various accessories or systems, such as, for example, a tire pressure monitoring system or a passenger air bag status or a garage door opening system or a telematics system or any other accessory or system of the mirror assembly or of the vehicle or of an accessory module or console of the vehicle, such as an accessory module or console of the types described in U.S. Pat. Nos. 7,289,037; 6,877,888; 6,824,281; 6,690,268; 6,672,744; 6,386,742 and 6,124,886, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are hereby incorporated herein by reference in their entireties.
- accessories or systems such as, for example, a tire pressure monitoring system or a passenger air bag status or a garage door opening system or a telematics system or any other accessory or system of the mirror assembly or of
- the display or displays may comprise a video display and may utilize aspects of the video display devices or modules described in U.S. Pat. Nos. 6,690,268; 7,184,190; 7,274,501; 7,370,983; 7,446,650 and/or 7,855,755, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are all hereby incorporated herein by reference in their entireties.
- the video display may be operable to display images captured by one or more imaging sensors or cameras at the vehicle.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Signal Processing (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Mechanical Engineering (AREA)
- Studio Devices (AREA)
- Closed-Circuit Television Systems (AREA)
- Traffic Control Systems (AREA)
Abstract
A method of image enhancement for a vehicle vision system includes capturing multiple frames of image data with a camera and executing a first brightness transfer function on a current frame of captured image data to generate a first enhanced image frame. At least one previously captured frame of image data is retrieved, and a second brightness transfer function is executed on the at least one previously captured frame of image data to generate a second enhanced image frame. The first and second enhanced image frames are blended to generate a blended enhanced image frame, and these steps are repeated so as to generate multiple blended enhanced image frames. The presence of an object is detected via processing of blended enhanced image frames by a processor, and an output is generated responsive to detection of the object.
Description
- The present application is a continuation of U.S. patent application Ser. No. 14/343,937, filed Mar. 10, 2014, now U.S. Pat. No. 9,681,062, which is a 371 national phase filing of PCT Application No. PCT/US2012/057007, filed Sep. 25, 2012, which claims the filing benefit of U.S. provisional application Ser. No. 61/539,049, filed Sep. 26, 2012, which is hereby incorporated herein by reference in its entirety.
- The present invention relates to imaging systems or vision systems for vehicles.
- Use of imaging sensors in vehicle imaging systems is common and known. Examples of such known systems are described in U.S. Pat. Nos. 5,877,897; 5,796,094; 5,670,935 and/or 5,550,677, which are hereby incorporated herein by reference in their entireties.
- The present invention provides a vision system or imaging system for a vehicle that utilizes one or more cameras to capture images exterior of the vehicle, such as forwardly or rearwardly of the vehicle, and provides for enhanced image processing to detect objects in poor visibility conditions, such as in dense fog or the like.
- The vision system may enhance the image processing by amplifying the contrast in the captured images by brightness transfer function filtering and exposure stacking and tracking contrast thresholds or features within the captured images, such as on a frame-by-frame basis as the vehicle travels along a road.
- These and other objects, advantages, purposes and features of the present invention will become apparent upon review of the following specification in conjunction with the drawings.
-
FIG. 1 is a plan view of a vehicle with a vision system and forward facing imaging sensor or camera that provides a forward exterior field of view in accordance with the present invention; -
FIG. 2 shows images captured by the forward facing camera and processed by the vision system of the present invention; -
FIGS. 3(a) and 3(b) show graphs showing histograms of luminance distribution for the vision system, withFIG. 3(b) showing the histogram of the original image's graph ofFIG. 3(a) as spread into the highest possible dynamic range of the target system which equates to a contrast amplification of the present invention; -
FIGS. 4(a) and 4(b) show graphs of brightness transfer functions A (FIG. 4(a) ) and B (FIG. 4(b) ) which find use in the flow chart ofFIGS. 5 and 6 , with the brightness transfer function A enhancing the brighter areas and dampening the darker ones, and with the brightness transfer function B decreasing the medium illuminated areas, and with the upper end at less of 100, whereby the overall illumination becomes decreased by this transfer function; -
FIG. 5 shows a flow chart of the image enhancing and processing steps according the invention, and by mapping/stacking an illumination reduced image scene on top of a contrast enhanced image the dynamic range of the image increases: overexposed areas appear less bright and underexposed more bright, which leads to acknowledge details in the scene easier; -
FIG. 6 shows a flow chart of the image enhancing and processing steps according the invention as to be used in a vehicle vision system, supporting machine and human vision driver assistant algorithms; and -
FIG. 7 shows an example of how images may be altered when processed according to the flow chart ofFIG. 5 , whereby it becomes apparent that the process turns out more contrasts of possible objects on foggy weather conditions when comparing Image(t0) and Imageh(t0). - Referring now to the drawings and the illustrative embodiments depicted therein, a
vehicle 10 includes an imaging system orvision system 12 that includes at least one imaging sensor or camera 14 (such as forward facing camera at the front (or at the windshield) of the vehicle), which captures images exterior of and forwardly of the vehicle (FIG. 1 ). Theimaging system 12 is operable to process (such as via an image processor) image data captured by thecamera 14 to present these on a display or to detect objects and/or headlights of approaching vehicles and/or taillights of leading vehicles in the field of view of the camera (such as for use in an object detection system of the vehicle or collision avoidance system of the vehicle or headlamp control system of the vehicle or adaptive cruise control system of the vehicle or lane change departure warning system of the vehicle or traffic sign recognition system of the vehicle or driver assistance system of the vehicle the like). - The image processor of the
vision system 12 is operable to process captured image data, such as to detect and identify objects forward (and optionally sideward and/or rearward) of the vehicle during normal operation of the vehicle. In poor visibility conditions, such as foggy conditions and/or heavy snow fall conditions or the like, objects may be difficult for the driver to see and may be difficult even for the image processor to detect, even when image processing algorithms for lens pollution detection (such as similar to that described in U.S. provisional application Ser. No. 61/616,126, filed Mar. 27, 2012, which is hereby incorporated herein by reference in its entirety) come into use. For example, and with reference to image “A” inFIG. 2 , during low visibility conditions, such as fog conditions as shown, it is difficult for the driver of the vehicle to detect the person and dog at the side of the road ahead of the vehicle and beyond the principal illumination area of the vehicle headlamps (set at low beams for seeing in the fog). The image processor may process the image to detect objects, but, and with reference to image “B” inFIG. 2 , normal image processing may not detect the object of interest (the person and dog in this example) due to the poor visibility conditions. Typically, the object detection may not work feasibly when the image contrast falls under a certain level. In order to increase the detectability of such objects in poor visibility conditions, the vision system of the present invention is operable to enhance or increase the contrast of the captured images so that any objects in the field of view of the camera are darkened to enhance the detectability of the objects by the image processor or to ease the visibility of objects to the driver of the vehicle. - As can be seen with reference to images “C” through “F” in
FIG. 2 , as the contrast is increased, the side markers or posts along the side of the road and the object of interest (the person and dog in this example) become darker and, in this example, the object moves relative to other fixed objects in the captured images (see images B-F inFIG. 2 and note that the person and dog approach the fixed road marker in the captured images), and thus the image processor can detect the presence of the fixed and moving objects and determine if they are objects of interest to the driver of the vehicle and generate the appropriate signal responsive to such detection and determination or identification. For example, the system, responsive to such an object detection, may generate an alert to the driver or may adjust the headlamps accordingly or may display the detected object on a display screen for viewing by the driver (particularly for backup assist systems where the object is detected rearward of the vehicle during a reversing maneuver). Thus, by increasing the contrast in captured images, the vision system can enhance detection of objects in the camera's field of view that may otherwise go undetected. The system may be operable to increase the contrast in the captured images responsive to a user input or to a detection or determination of a low visibility condition, such as responsive to a signal from a rain sensor or the like that is indicative of detection of a foggy condition or such as responsive to image processing of the captured images to determine that the vehicle is in foggy driving conditions (such as by utilizing aspects of the vision systems described in U.S. Pat. Nos. 4,973,844; 5,796,094; 5,877,897 and 6,353,392, which are hereby incorporated herein by reference in their entireties). - It is known to provide image contrast enhancing for photographs (such as photographs taken by hand held digital cameras or astronomical telescopes or the like), and such enhancements may be done by known computer based tools for editing images. Today, nearly every operating system, library, presenting program and/or the like provides at least basic image editing functions. Professional photo editing programs like CoralDRAW®, Gimp® or Adobe Photoshop° provide a wide range of image editing and enhancing features. Typically used for contrast enhancing is the editing of the contrast histogram. This can be used to expose objects stronger. A function used especially to do this is “Contrast Enhancement through Localized Histogram Equalization” (see Cromwell-intl.com: http://www.cromwell-intl.com/3d/histogram/, which is hereby incorporated herein by reference in its entirety). Even night images can become contrast enhanced in a way that low illuminated objects turn out more visible. Such algorithms used in consumer computer programs for image enhancing are typically used in individual pictures, and are not meant to be used in real time applications.
- Image quality improvement in poor visibility conditions is known from airborne weather surveillance pictures for reworking pictures taken in cloudy (foggy) situations. The best results were achieved by Oakley et al. when contrast enhancement algorithm in conjunction with a temporal filters came into use (see Image Processing, IEEE; “Improving Image Quality in Poor Visibility Conditions Using a Physical Model for Contrast Degradation,” http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=660994, by Oakley, J. P. and Satherley, B. L., February 1998, which is hereby incorporated herein by reference in its entirety). The base was a physical model on fog reflection.
- Attempts have been made to do video contrast enhancements such as in “Contrast Enhancement Using Brightness Preserving Bi-Histogram Equalization” by Yeong-Taeg Kim (Consumer Electronics: IEEE: “Contrast Enhancement Using Brightness Preserving Bi-Histogram Equalization,” by Yeong-Taeg Kim, February 1997, which is hereby incorporated herein by reference in its entirety). This requires real time processing. Demand for this was and is in applications for the likes of television images, images providing medical devices, military engineering and/or the like, and Kim et al. suggested “Partially Overlapped Sub-Block Histogram Equalization” to be used in cameras (Circuits and Systems for Video Technology, IEEE: “Partially Overlapped Sub-Block Histogram Equalization” http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=915354, by Joung-Youn Kim, Lee-Sup Kim and Seung-Ho Hwang, April 2001, which is hereby incorporated herein by reference in its entirety). Also, Marsi et al. were able to simplify algorithms by attempting recursive rational filters (Imaging Systems and Techniques, 2004; IEEE International Workshop: “Real Time Video Contrast Enhancement by Using Recursive Rational Filter,” http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=1397276, by Marsi, S., Ramponi, G. and Carrato, S., May 14, 2004, which is hereby incorporated herein by reference in its entirety), and Wang et al. suggested the use of weighted thresholded histogram equalization for fast processing (Consumer Electronics, IEEE: “Real Time Video Contrast Enhancement by using Weighted Thresholded Histogram Equalization” http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=4266969, by Qing Wang and Ward, R. K., May 2007, which is hereby incorporated herein by reference in its entirety). Another challenge is the noise, a common problem on electronic cameras; Starck et al. published a procedure to do noise reduction by curvelet transforms in 2003 (Image Processing, IEEE: “Gray and Color Image Contrast Enhancement by the Curvelet Transform,” http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=1208320, by Starck, J.-L., Murtagh, F., Candes, E. J. and Donoho, D. L., Jun. 2003, which is hereby incorporated herein by reference in its entirety).
- It is also known to use infrared systems or low light amplifying systems in vehicles. Earlier systems have used infrared cameras alone, and some systems additionally use infrared headlights to light up the area in front of the vehicle (invisible for the human eye) which makes that area easier to detect with the infrared camera. Infrared cameras may provide enhanced performance in object detection in dense fog conditions due to its physical principal and the detected wave length have the intrinsic property to interfuse fog, so objects in fog can be detected and/or visualized.
- State of the art automotive driver assistance systems typically provide the driver with useful information of the vehicle's environment, including the traffic or objects in front of, to the side of and rearward of the vehicle. Typically, there are additional warnings or image overlays for highlighting hazards, especially those in the driving direction of the vehicle and in the anticipated path of travel of the vehicle. Obstacles or pedestrians that are in the way or path of the vehicle or tend to step into the path of the vehicle may be highlighted. Systems which also do active interventions such as braking or collision avoidance maneuvers are also known. For distinguishing pedestrians from other objects and for predetermining their walking direction and speed, the detected objects need to be tracked over a certain time. Also, analyzing shapes or markers of walking or standing pedestrians is known in the field of automotive vision systems and image processing. Due to the vehicle's own movement, the objects in the captured images flow or move over successively captured images (optical flow). For example, external or outside objects (even stationary objects) move through the images taken from a front facing vehicle camera as the vehicle travels along the road. Algorithms for tracking objects under driving conditions are also known. When a vehicle drives through a turn, the optical flow also behaves in a turned manner. That turn can be anticipated by the knowledge of the steering wheel's angle and a kinematic model of the vehicle's curve behavior. The optical flow speed directly translates from the vehicle's ground speed given by the odometer. Alternatively known algorithms may determine the optical flow direct from the image flow without the previous mentioned input from the vehicle.
- For enabling the above mentioned pedestrian and obstacle acknowledging and tracking algorithm to work properly, especially to be able to highlight a hazard or warn the driver or intervene (such as via braking or cruise control adjustment or the like), it is necessary to receive sufficient images. In foggy driving conditions or during heavy snow fall driving conditions, cameras in the visible spectrum deliver images of insufficient quality. The present invention provides enhanced image quality of visible spectrum cameras, especially the dynamic range of the resulting image, so that the driver assist system algorithms can work properly and/or display the processed image to the driver as an improvement to his or her view in such limited visibility conditions. This is achieved without the need of additional cameras using different light spectrums (such as infrared sensitive cameras or the like) or other sensors for the same purpose or high dynamic range (HDR) cameras.
- The present invention thus provides enhanced image quality in poor visibility conditions captured by a non HDR camera by amplifying the contrast details in the captured images by generating a pseudo HDR image out of current and historical image components by tone mapping. The system then tracks the contrast thresholds/features within the captured images with respect to the image flow caused by the vehicle's movement. This process is repeated on a frame-by-frame basis to detect and identify objects in the camera's forward field of view, as can be seen in
FIGS. 5 and 6 . At every loop the historically (previously enhanced) image (Imageh(t0−n))) passes two individual image transfer functions and then becomes superpositioned (or mapped, merged, blended or stacked) by the currently captured frame (Image t0)). This tone mapping method is called image stacking, exposure fusion or exposure blending. The mapping ratio of how much of the historical image (Imageh(t0−n))) becomes kept and how much of the current image (Image t0)) becomes mapped in is freely selectable between 0 and 1. In the example inFIG. 5 , 20%/80% was chosen for a data frame rate of 30 frames per second. Slower frame rates might require a shift into a stronger influence of (Image t0)). The used image enhancements shall not be limited to these shown in the example ofFIGS. 5 and 6 . - The brightness transfer function A (
FIG. 4(a) ) enhances the brighter areas and dampens the darker ones. This equates to a histogram spread (Dynamic Range Increase), such as shown inFIGS. 3(a) and 3(b) , of the historically image (Imageh(t0−n))). The brightness transfer function B (FIG. 4(b) ) decreases the medium illuminated areas of the currently captured image (Image t0)). The upper end is at less than 100%. The overall illumination becomes decreased by this transfer function.FIG. 7 shows that already after this step the discrimination between the object (person with dog) and surrounding (foggy) area is improved. By mapping/stacking an illumination reduced image scene (currently captured images) on top of a contrast enhanced image (historical image), the dynamic range of the image increases as to be seen inFIGS. 3(a) and 3(b) . Overexposed areas appear less bright and underexposed more bright which leads to acknowledge details in the scene easier (seeFIG. 7 ). After consecutive loops it may come to a blooming effect or halo at the borderline of areas with high contrast. This effect may be enhanced by some blurring which is caused by unavoidable inaccuracy of the distorting, turning, cropping and moving of the currently captured image to the historical scene. - The result of this image processing and tracking of the features with respect to the optical flow and the vehicle movement is shown in principle in
FIG. 2 (and discussed above). The algorithm based on already established image processing procedures (non-automotive, image enhancements of photographs and ‘image registration’ and the like), such as tonal value splitting/-buckling/-limiting, histogram equalization and the like, as simplified can be seen with reference toFIGS. 3(a) and 3(b) . - Because the yet to be processed images are captured by a camera on a moving vehicle, it is necessary that the optical flow and the according information or data of objects (both steady or moving) moving through the images, including the vehicle speed, the steering angle of the vehicle and the like, be taken into account. There may be a model of the vehicle's cinematic mathematical equations. Its results may be stored in a look up table. The camera's or cameras parameters as like mounting position and viewing angle optical properties may be reflected in that (combined) look up table or in another mathematical model or table. The moving objects/obstacles can thus be distinguished from steady objects relative to the movement of the vehicle that is equipped with the camera system or vision system of the present invention. Object classification may work on further distances by feeding enhanced image data. Further algorithms may process the image data and may indicate hazards or the like, and/or may actively intervene to avoid collisions and the like. The image enhancing algorithm may find use in processing multiple camera images separate or by processing a stitched image which may be arranged as a vehicle top view image or the like.
- The imaging sensor and its photosensor array may comprise any suitable camera or sensing device, such as, for example, an array of a plurality of photosensor elements arranged in 640 columns and 480 rows (a 640×480 imaging array), with a respective lens focusing images onto respective portions of the array. The photosensor array may comprise a plurality of photosensor elements arranged in a photosensor array having rows and columns. The logic and control circuit of the imaging sensor may function in any known manner, such as in the manner described in U.S. Pat. Nos. 5,550,677; 5,877,897; 6,498,620; 5,670,935; 5,796,094 and/or 6,396,397, and/or U.S. provisional applications, Ser. No. 61/696,416, filed Sep. 4, 2012; Ser. No. 61/682,995, filed Aug. 14, 2012; Ser. No. 61/682,486, filed Aug. 13, 2012; Ser. No. 61/680,883, filed Aug. 8, 2012; Ser. No. 61/678,375, filed Aug. 1, 2012; Ser. No. 61/676,405, filed Jul. 27, 2012; Ser. No. 61/666,146, filed Jun. 29, 2012; Ser. No. 61/653,665, filed May 31, 2012; Ser. No. 61/653,664, filed May 31, 2012; Ser. No. 61/648,744, filed May 18, 2012; Ser. No. 61/624,507, filed Apr. 16, 2012; Ser. No. 61/616,126, filed Mar. 27, 2012; Ser. No. 61/615,410, filed Mar. 26, 2012; Ser. No. 61/613,651, filed Mar. 21, 2012; Ser. No. 61/607,229, filed Mar. 6, 2012; Ser. No. 61/605,409, filed Mar. 1, 2012; Ser. No. 61/602,878, filed Feb. 24, 2012; Ser. No. 61/602,876, filed Feb. 24, 2012; Ser. No. 61/600,205, filed Feb. 17, 2012; Ser. No. 61/588,833, filed Jan. 20, 2012; Ser. No. 61/583,381, filed Jan. 5, 2012; Ser. No. 61/579,682, filed Dec. 23, 2011; Ser. No. 61/570,017, filed Dec. 13, 2011; Ser. No. 61/568,791, filed Dec. 9, 2011; Ser. No. 61/567,446, filed Dec. 6, 2011; Ser. No. 61/559,970, filed Nov. 15, 2011; and/or Ser. No. 61/552,167, filed Oct. 27, 2011, and/or PCT Application No. PCT/CA2012/000378, filed Apr. 25, 2012, and published Nov. 1, 2012 as International Publication No. WO 2012/145822, and/or PCT Application No. PCT/US2012/056014, filed Sep. 19, 2012, and published Mar. 28, 2013 as International Publication No. WO 2013/043661, and/or PCT Application No. PCT/US2012/048800, filed Jul. 30, 2012, and published Feb. 7, 2013 as International Publication No. WO 2013/019707, and/or PCT Application No. PCT/US2012/048110, filed Jul. 25, 2012, and published Jan. 31, 2013 as International Publication No. WO 2013/016409, and/or U.S. patent application Ser. No. 13/534,657, filed Jun. 27, 2012, and published Jan. 3, 2013 as U.S. Publication No. US-2013-0002873, which are all hereby incorporated herein by reference in their entireties. The system may communicate with other communication systems via any suitable means, such as by utilizing aspects of the systems described in PCT Application No. PCT/US10/038477, filed Jun. 14, 2010, and/or U.S. patent application Ser. No. 13/202,005, filed Aug. 17, 2011, now U.S. Pat. No. 9,126,525, and/or U.S. provisional applications, Ser. No. 61/650,667, filed May 23, 2012; Ser. No. 61/579,682, filed Dec. 23, 2011; Ser. No. 61/565,713, filed Dec. 1, 2011, which are hereby incorporated herein by reference in their entireties.
- The imaging device and control and image processor and any associated illumination source, if applicable, may comprise any suitable components, and may utilize aspects of the cameras and vision systems described in U.S. Pat. Nos. 5,550,677; 5,877,897; 6,498,620; 5,670,935; 5,796,094; 6,396,397; 6,806,452; 6,690,268; 7,005,974; 7,123,168; 7,004,606; 6,946,978; 7,038,577; 6,353,392; 6,320,176; 6,313,454 and 6,824,281, and/or International Publication No. WO 2010/099416, published Sep. 2, 2010, and/or PCT Application No. PCT/US10/47256, filed Aug. 31, 2010, and/or U.S. patent application Ser. No. 12/508,840, filed Jul. 24, 2009, and published Jan. 28, 2010 as U.S. Pat. Publication No. US 2010-0020170; and/or PCT Application No. PCT/US2012/048110, filed Jul. 25, 2012, and published Jan. 31, 2013 as International Publication No. WO 2013/016409, and/or U.S. patent application Ser. No. 13/534,657, filed Jun. 27, 2012, and published Jan. 3, 2013 as U.S. Publication No. US-2013-0002873, which are all hereby incorporated herein by reference in their entireties. The camera or cameras may comprise any suitable cameras or imaging sensors or camera modules, and may utilize aspects of the cameras or sensors described in U.S. patent applications, Ser. No. 12/091,359, filed Apr. 24, 2008 and published Oct. 1, 2009 as U.S. Publication No. US-2009-0244361; and/or Ser. No. 13/260,400, filed Sep. 26, 2011, now U.S. Pat. No. 8,542,451, and/or U.S. Pat. Nos. 7,965,336 and/or 7,480,149, which are hereby incorporated herein by reference in their entireties. The imaging array sensor may comprise any suitable sensor, and may utilize various imaging sensors or imaging array sensors or cameras or the like, such as a CMOS imaging array sensor, a CCD sensor or other sensors or the like, such as the types described in U.S. Pat. Nos. 5,550,677; 5,670,935; 5,760,962; 5,715,093; 5,877,897; 6,922,292; 6,757,109; 6,717,610; 6,590,719; 6,201,642; 6,498,620; 5,796,094; 6,097,023; 6,320,176; 6,559,435; 6,831,261; 6,806,452; 6,396,397; 6,822,563; 6,946,978; 7,339,149; 7,038,577; 7,004,606; 7,720,580 and/or 7,965,336, and/or PCT Application No. PCT/US2008/076022, filed Sep. 11, 2008 and published Mar. 19, 2009 as International Publication No. WO 2009/036176, and/or PCT Application No. PCT/US2008/078700, filed Oct. 3, 2008 and published Apr. 9, 2009 as International Publication No. WO 2009/046268, which are all hereby incorporated herein by reference in their entireties.
- The camera module and circuit chip or board and imaging sensor may be implemented and operated in connection with various vehicular vision-based systems, and/or may be operable utilizing the principles of such other vehicular systems, such as a vehicle headlamp control system, such as the type disclosed in U.S. Pat. Nos. 5,796,094; 6,097,023; 6,320,176; 6,559,435; 6,831,261; 7,004,606; 7,339,149 and/or 7,526,103, which are all hereby incorporated herein by reference in their entireties, a rain sensor, such as the types disclosed in commonly assigned U.S. Pat. Nos. 6,353,392; 6,313,454; 6,320,176 and/or 7,480,149, which are hereby incorporated herein by reference in their entireties, a vehicle vision system, such as a forwardly, sidewardly or rearwardly directed vehicle vision system utilizing principles disclosed in U.S. Pat. Nos. 5,550,677; 5,670,935; 5,760,962; 5,877,897; 5,949,331; 6,222,447; 6,302,545; 6,396,397; 6,498,620; 6,523,964; 6,611,202; 6,201,642; 6,690,268; 6,717,610; 6,757,109; 6,802,617; 6,806,452; 6,822,563; 6,891,563; 6,946,978 and/or 7,859,565, which are all hereby incorporated herein by reference in their entireties, a trailer hitching aid or tow check system, such as the type disclosed in U.S. Pat. No. 7,005,974, which is hereby incorporated herein by reference in its entirety, a reverse or sideward imaging system, such as for a lane change assistance system or lane departure warning system or for a blind spot or object detection system, such as imaging or detection systems of the types disclosed in U.S. Pat. Nos. 7,881,496; 7,720,580; 7,038,577; 5,929,786 and/or 5,786,772, and/or U.S. provisional applications, Ser. No. 60/628,709, filed Nov. 17, 2004; Ser. No. 60/614,644, filed Sep. 30, 2004; Ser. No. 60/618,686, filed Oct. 14, 2004; Ser. No. 60/638,687, filed Dec. 23, 2004, which are hereby incorporated herein by reference in their entireties, a video device for internal cabin surveillance and/or video telephone function, such as disclosed in U.S. Pat. Nos. 5,760,962; 5,877,897; 6,690,268 and/or 7,370,983, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are hereby incorporated herein by reference in their entireties, a traffic sign recognition system, a system for determining a distance to a leading or trailing vehicle or object, such as a system utilizing the principles disclosed in U.S. Pat. Nos. 6,396,397 and/or 7,123,168, which are hereby incorporated herein by reference in their entireties, and/or the like.
- Optionally, the circuit board or chip may include circuitry for the imaging array sensor and or other electronic accessories or features, such as by utilizing compass-on-a-chip or EC driver-on-a-chip technology and aspects such as described in U.S. Pat. Nos. 7,255,451 and/or 7,480,149; and/or U.S. patent applications, Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US-2006-0061008, and/or Ser. No. 12/578,732, filed Oct. 14, 2009, now U.S. Pat. No. 9,487,144, which are hereby incorporated herein by reference in their entireties.
- Optionally, the vision system may include a display for displaying images captured by one or more of the imaging sensors for viewing by the driver of the vehicle while the driver is normally operating the vehicle. Optionally, for example, the vision system may include a video display device disposed at or in the interior rearview mirror assembly of the vehicle, such as by utilizing aspects of the video mirror display systems described in U.S. Pat. No. 6,690,268 and/or U.S. patent application Ser. No. 13/333,337, filed Dec. 21, 2011, now U.S. Pat. No. 9,264,672, which are hereby incorporated herein by reference in their entireties. The video mirror display may comprise any suitable devices and systems and optionally may utilize aspects of the compass display systems described in U.S. Pat. Nos. 7,370,983; 7,329,013; 7,308,341; 7,289,037; 7,249,860; 7,004,593; 4,546,551; 5,699,044; 4,953,305; 5,576,687; 5,632,092; 5,677,851; 5,708,410; 5,737,226; 5,802,727; 5,878,370; 6,087,953; 6,173,508; 6,222,460; 6,513,252 and/or 6,642,851, and/or European patent application, published Oct. 11, 2000 under Publication No. EP 0 1043566, and/or U.S. patent application Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US-2006-0061008, which are all hereby incorporated herein by reference in their entireties. Optionally, the video mirror display screen or device may be operable to display images captured by a rearward viewing camera of the vehicle during a reversing maneuver of the vehicle (such as responsive to the vehicle gear actuator being placed in a reverse gear position or the like) to assist the driver in backing up the vehicle, and optionally may be operable to display the compass heading or directional heading character or icon when the vehicle is not undertaking a reversing maneuver, such as when the vehicle is being driven in a forward direction along a road (such as by utilizing aspects of the display system described in PCT Application No. PCT/US2011/056295, filed Oct. 14, 2011 and published Apr. 19, 2012 as International Publication No. WO 2012/051500, which is hereby incorporated herein by reference in its entirety).
- Optionally, the vision system (utilizing the forward facing camera and a rearward facing camera and other cameras disposed at the vehicle with exterior fields of view) may be part of or may provide a display of a top-down view or birds-eye view system of the vehicle or a surround view at the vehicle, such as by utilizing aspects of the vision systems described in PCT Application No. PCT/US10/25545, filed Feb. 26, 2010 and published on Sep. 2, 2010 as International Publication No. WO 2010/099416, and/or PCT Application No. PCT/US10/47256, filed Aug. 31, 2010 and published Mar. 10, 2011 as International Publication No. WO 2011/028686, and/or PCT Application No. PCT/US11/62755, filed Dec. 1, 2011 and published Jun. 7, 2012 as International Publication No. WO 2012-075250, and/or PCT Application No. PCT/US2012/048993, filed Jul. 31, 2012, and published Feb. 7, 2013 as International Publication No. WO 2013/019795, and/or PCT Application No. PCT/CA2012/000378, filed Apr. 25, 2012, and published Nov. 1, 2012 as International Publication No. WO 2012/145822, and/or U.S. patent application Ser. No. 13/333,337, filed Dec. 21, 2011, now U.S. Pat. No. 9,264,672, and/or U.S. provisional applications, Ser. No. 61/615,410, filed Mar. 26, 2012; Ser. No. 61/588,833, filed Jan. 20, 2012; Ser. No. 61/570,017, filed Dec. 13, 2011; Ser. No. 61/568,791, filed Dec. 9, 2011; Ser. No. 61/559,970, filed Nov. 15, 2011; Ser. No. 61/540,256, filed Sep. 28, 2011, which are hereby incorporated herein by reference in their entireties.
- Optionally, the video mirror display may be disposed rearward of and behind the reflective element assembly and may comprise a display such as the types disclosed in U.S. Pat. Nos. 5,530,240; 6,329,925; 7,855,755; 7,626,749; 7,581,859; 7,338,177; 7,274,501; 7,255,451; 7,195,381; 7,184,190; 5,668,663; 5,724,187 and/or 6,690,268, and/or in U.S. patent applications, Ser. No. 11/226,628, filed Sep. 14, 2005 and published Mar. 23, 2006 as U.S. Publication No. US-2006-0061008; and/or Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are all hereby incorporated herein by reference in their entireties. The display is viewable through the reflective element when the display is activated to display information. The display element may be any type of display element, such as a vacuum fluorescent (VF) display element, a light emitting diode (LED) display element, such as an organic light emitting diode (OLED) or an inorganic light emitting diode, an electroluminescent (EL) display element, a liquid crystal display (LCD) element, a video screen display element or backlit thin film transistor (TFT) display element or the like, and may be operable to display various information (as discrete characters, icons or the like, or in a multi-pixel manner) to the driver of the vehicle, such as passenger side inflatable restraint (PSIR) information, tire pressure status, and/or the like. The mirror assembly and/or display may utilize aspects described in U.S. Pat. Nos. 7,184,190; 7,255,451; 7,446,924 and/or 7,338,177, which are all hereby incorporated herein by reference in their entireties. The thicknesses and materials of the coatings on the substrates of the reflective element may be selected to provide a desired color or tint to the mirror reflective element, such as a blue colored reflector, such as is known in the art and such as described in U.S. Pat. Nos. 5,910,854; 6,420,036 and/or 7,274,501, which are hereby incorporated herein by reference in their entireties.
- Optionally, the display or displays and any associated user inputs may be associated with various accessories or systems, such as, for example, a tire pressure monitoring system or a passenger air bag status or a garage door opening system or a telematics system or any other accessory or system of the mirror assembly or of the vehicle or of an accessory module or console of the vehicle, such as an accessory module or console of the types described in U.S. Pat. Nos. 7,289,037; 6,877,888; 6,824,281; 6,690,268; 6,672,744; 6,386,742 and 6,124,886, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are hereby incorporated herein by reference in their entireties.
- The display or displays may comprise a video display and may utilize aspects of the video display devices or modules described in U.S. Pat. Nos. 6,690,268; 7,184,190; 7,274,501; 7,370,983; 7,446,650 and/or 7,855,755, and/or U.S. patent application Ser. No. 10/538,724, filed Jun. 13, 2005 and published Mar. 9, 2006 as U.S. Publication No. US-2006-0050018, which are all hereby incorporated herein by reference in their entireties. The video display may be operable to display images captured by one or more imaging sensors or cameras at the vehicle.
- Changes and modifications to the specifically described embodiments may be carried out without departing from the principles of the present invention, which is intended to be limited only by the scope of the appended claims as interpreted according to the principles of patent law.
Claims (36)
1. A method of image enhancement for a vehicle vision system, said method comprising:
(a) providing a camera at a vehicle so that the camera has an exterior field of view;
(b) providing a processor operable to process image data;
(c) capturing multiple frames of image data with the camera;
(d) executing a first brightness transfer function on a current frame of image data captured by the camera to generate a first enhanced image frame;
(e) retrieving at least one previously captured frame of image data;
(f) executing a second brightness transfer function on the at least one previously captured frame of image data to generate a second enhanced image frame, wherein the second brightness transfer function is different from the first brightness transfer function;
(g) blending the first and second enhanced image frames to generate a blended enhanced image frame;
repeating steps (d)-(g) so as to generate multiple blended enhanced image frames;
(h) detecting, via processing of blended enhanced image frames by the processor, presence of an object in the field of view of the camera; and
(i) generating an output responsive to detection of the object present in the field of view of the camera.
2. The method of claim 1 , wherein the step of executing the first brightness transfer function enhances contrast of the current frame of image data captured by the camera.
3. The method of claim 1 , comprising executing tone mapping of multiple frames of captured image data to enhance detection of the object present in the field of view of the camera.
4. The method of claim 1 , comprising classifying the detected object present in the field of view of the camera.
5. The method of claim 4 , comprising generating an output responsive to classification of the detected object.
6. The method of claim 1 , comprising determining a low visibility driving condition and, responsive to determination of the low visibility driving condition, increasing contrast of features in captured image data by brightening brighter areas of captured image data and dampening darker areas of captured image data.
7. The method of claim 6 , comprising increasing contrast of features in captured image data over multiple successive frames of captured image data.
8. The method of claim 7 , comprising tracking, via processing by the processor of multiple successive frames of captured image data during the determined low visibility driving condition, image flow caused by movement of the vehicle to enhance detection and identification of objects present in the field of view of the camera.
9. The method of claim 1 , wherein the at least one previously captured frame of image data is retrieved from memory.
10. The method of claim 1 , comprising determining a low visibility driving condition via processing of captured image data.
11. The method of claim 1 , comprising determining that fog is present in the field of view of the camera via processing of captured image data.
12. The method of claim 1 , comprising providing blended enhanced image frames of image data to a video display screen that is disposed in the vehicle at a location viewable by a driver of the vehicle when operating the vehicle.
13. The method of claim 1 , wherein capturing multiple frames of image data with the camera comprises capturing multiple frames of image data with the camera at a frame rate of at least 30 frames per second.
14. The method of claim 1 , wherein blending the first and second enhanced image frames comprises blending the first and second enhanced image frames to generate a blended enhanced image frame that is up to 20 percent derived from the first enhanced image frame.
15. The method of claim 1 , wherein executing a second brightness transfer function on the at least one previously captured frame of image data comprises executing the second brightness transfer function on a plurality of previously captured frames of image data to generate the second enhanced image frame.
16. The method of claim 1 , comprising providing the generated output to a driver assistance system of the vehicle.
17. The method of claim 16 , wherein the driver assistance system of the vehicle comprises a system selected from the group consisting of a lane change assist system of the vehicle, a lane departure warning system of the vehicle, a blind spot detection system of the vehicle, an adaptive cruise control system of the vehicle, a collision avoidance system of the vehicle, a traffic sign recognition system of the vehicle, and a vehicle headlamp control system of the vehicle.
18. The method of claim 1 , comprising tracking the detected object over successive frames of captured image data to determine if the detected object is an object of interest in the field of view of the camera.
19. The method of claim 1 , wherein processing of captured image data by the processor is responsive at least in part to steering of the vehicle.
20. The method of claim 1 , comprising distinguishing, via processing of captured image data by the processor, moving objects from non-moving objects.
21. The method of claim 20 , wherein distinguishing moving objects comprises distinguishing moving objects responsive at least in part to at least one of (i) speed of the vehicle and (ii) steering of the vehicle.
22. The method of claim 1 , comprising disposing the camera at a rear portion of the vehicle with an exterior field of view rearward of the vehicle, and wherein said method comprises providing a plurality of cameras at the vehicle so as to have respective exterior fields of view, and wherein the plurality of cameras comprises the camera at the rear portion of the vehicle.
23. The method of claim 22 , comprising providing a display for displaying images derived, at least in part, from image data captured by the camera at the rear portion of the vehicle and derived, at least in part, from image data captured by other cameras of the plurality of cameras.
24. The method of claim 1 , comprising disposing the camera at a rear portion of the vehicle with an exterior field of view rearward of the vehicle, and comprising providing a display for displaying images derived, at least in part, from image data captured by the camera during a reversing maneuver of the vehicle.
25. A method of image enhancement for a vehicle vision system, said method comprising:
(a) providing a camera at a vehicle so that the camera has an exterior field of view;
(b) providing a processor operable to process image data;
(c) capturing multiple frames of image data with the camera;
(d) executing a first brightness transfer function on a current frame of image data captured by the camera to generate a first enhanced image frame;
(e) retrieving at least one previously captured frame of image data;
(f) executing a second brightness transfer function on the at least one previously captured frame of image data to generate a second enhanced image frame, wherein the second brightness transfer function is different from the first brightness transfer function;
(g) blending the first and second enhanced image frames to generate a blended enhanced image frame;
repeating steps (d)-(g) so as to generate multiple blended enhanced image frames;
(h) detecting, via processing of blended enhanced image frames by the processor, presence of an object in the field of view of the camera;
(i) tracking the detected object over successive frames of captured image data to determine if the detected object is an object of interest in the field of view of the camera;
(j) generating an output responsive to determination of the detected object being an object of interest; and
(k) providing the generated output to a driver assistance system of the vehicle.
26. The method of claim 25 , wherein the driver assistance system of the vehicle comprises a blind spot detection system of the vehicle.
27. The method of claim 25 , wherein the driver assistance system of the vehicle comprises a collision avoidance system of the vehicle.
28. The method of claim 25 , wherein the driver assistance system of the vehicle comprises an adaptive cruise control system of the vehicle.
29. The method of claim 25 , wherein executing a second brightness transfer function on the at least one previously captured frame of image data comprises executing the second brightness transfer function on a plurality of previously captured frames of image data to generate the second enhanced image frame.
30. The method of claim 29 , wherein blending the first and second enhanced image frames comprises blending the first and second enhanced image frames to generate a blended enhanced image frame that is up to 20 percent derived from the first enhanced image frame.
31. A method of image enhancement for a vehicle vision system, said method comprising:
(a) providing a camera at a rear portion of a vehicle with an exterior field of view rearward of the vehicle;
(b) providing a processor operable to process image data;
(c) capturing multiple frames of image data with the camera;
(d) executing a first brightness transfer function on a current frame of image data captured by the camera to generate a first enhanced image frame;
(e) retrieving at least one previously captured frame of image data;
(f) executing a second brightness transfer function on the at least one previously captured frame of image data to generate a second enhanced image frame, wherein the second brightness transfer function is different from the first brightness transfer function;
(g) blending the first and second enhanced image frames to generate a blended enhanced image frame;
repeating steps (d)-(g) so as to generate multiple blended enhanced image frames; and
(h) providing blended enhanced image frames of image data to a video display screen that is disposed in the vehicle at a location viewable by a driver of the vehicle when operating the vehicle.
32. The method of claim 31 , comprising determining a low visibility driving condition via processing of captured image data.
33. The method of claim 32 , comprising, responsive to determination of the low visibility driving condition, increasing contrast of features in captured image data by brightening brighter areas of captured image data and dampening darker areas of captured image data.
34. The method of claim 31 , comprising determining that fog is present in the field of view of the camera via processing of captured image data.
35. The method of claim 31 , comprising providing a plurality of cameras at the vehicle so as to have respective exterior fields of view, wherein the plurality of cameras comprises the camera at the rear portion of the vehicle, and wherein the video display screen displays images derived, at least in part, from image data captured by the camera at the rear portion of the vehicle and derived, at least in part, from image data captured by other cameras of the plurality of cameras.
36. The method of claim 35 , wherein the plurality of cameras is part of a surround view multi-camera system of the vehicle, and wherein images displayed by the video display screen derived, at least in part, from image data captured by the camera at the rear portion of the vehicle and derived, at least in part, from image data captured by other cameras of the plurality of cameras comprise a bird's-eye view.
Priority Applications (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US15/619,630 US9774790B1 (en) | 2011-09-26 | 2017-06-12 | Method for enhancing vehicle camera image quality |
| US15/713,814 US10257432B2 (en) | 2011-09-26 | 2017-09-25 | Method for enhancing vehicle camera image quality |
Applications Claiming Priority (4)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201161539049P | 2011-09-26 | 2011-09-26 | |
| PCT/US2012/057007 WO2013048994A1 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| US14/343,937 US9681062B2 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| US15/619,630 US9774790B1 (en) | 2011-09-26 | 2017-06-12 | Method for enhancing vehicle camera image quality |
Related Parent Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2012/057007 Continuation WO2013048994A1 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| US14/343,937 Continuation US9681062B2 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
Related Child Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/713,814 Continuation US10257432B2 (en) | 2011-09-26 | 2017-09-25 | Method for enhancing vehicle camera image quality |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| US9774790B1 US9774790B1 (en) | 2017-09-26 |
| US20170280061A1 true US20170280061A1 (en) | 2017-09-28 |
Family
ID=47996325
Family Applications (3)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/343,937 Expired - Fee Related US9681062B2 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| US15/619,630 Active US9774790B1 (en) | 2011-09-26 | 2017-06-12 | Method for enhancing vehicle camera image quality |
| US15/713,814 Active US10257432B2 (en) | 2011-09-26 | 2017-09-25 | Method for enhancing vehicle camera image quality |
Family Applications Before (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/343,937 Expired - Fee Related US9681062B2 (en) | 2011-09-26 | 2012-09-25 | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
Family Applications After (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US15/713,814 Active US10257432B2 (en) | 2011-09-26 | 2017-09-25 | Method for enhancing vehicle camera image quality |
Country Status (2)
| Country | Link |
|---|---|
| US (3) | US9681062B2 (en) |
| WO (1) | WO2013048994A1 (en) |
Cited By (1)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180048867A1 (en) * | 2011-12-06 | 2018-02-15 | Mobileye Vision Technologies Ltd. | Road vertical contour detection |
Families Citing this family (49)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| GB2447672B (en) | 2007-03-21 | 2011-12-14 | Ford Global Tech Llc | Vehicle manoeuvring aids |
| US9723274B2 (en) | 2011-04-19 | 2017-08-01 | Ford Global Technologies, Llc | System and method for adjusting an image capture setting |
| US9283892B2 (en) | 2011-04-19 | 2016-03-15 | Ford Global Technologies, Llc | Method and system for monitoring placement of a target on a trailer |
| US8930140B2 (en) | 2011-04-19 | 2015-01-06 | Ford Global Technologies, Llc | Trailer target placement assist system and method |
| US9102272B2 (en) | 2011-04-19 | 2015-08-11 | Ford Global Technologies, Llc | Trailer target monitoring system and method |
| US9346396B2 (en) | 2011-04-19 | 2016-05-24 | Ford Global Technologies, Llc | Supplemental vehicle lighting system for vision based target detection |
| US9374562B2 (en) | 2011-04-19 | 2016-06-21 | Ford Global Technologies, Llc | System and method for calculating a horizontal camera to target distance |
| US9296422B2 (en) | 2011-04-19 | 2016-03-29 | Ford Global Technologies, Llc | Trailer angle detection target plausibility |
| US9854209B2 (en) | 2011-04-19 | 2017-12-26 | Ford Global Technologies, Llc | Display system utilizing vehicle and trailer dynamics |
| US9555832B2 (en) | 2011-04-19 | 2017-01-31 | Ford Global Technologies, Llc | Display system utilizing vehicle and trailer dynamics |
| US9102271B2 (en) | 2011-04-19 | 2015-08-11 | Ford Global Technologies, Llc | Trailer monitoring system and method |
| US9926008B2 (en) | 2011-04-19 | 2018-03-27 | Ford Global Technologies, Llc | Trailer backup assist system with waypoint selection |
| US9683848B2 (en) | 2011-04-19 | 2017-06-20 | Ford Global Technologies, Llc | System for determining hitch angle |
| US10196088B2 (en) | 2011-04-19 | 2019-02-05 | Ford Global Technologies, Llc | Target monitoring system and method |
| US9681062B2 (en) | 2011-09-26 | 2017-06-13 | Magna Electronics Inc. | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| DE102011084762A1 (en) * | 2011-10-19 | 2013-04-25 | Robert Bosch Gmbh | Method and device for determining a position of an object in an environment of a vehicle |
| US9264673B2 (en) | 2011-11-20 | 2016-02-16 | Magna Electronics, Inc. | Vehicle vision system with enhanced functionality |
| JP6107079B2 (en) * | 2012-11-21 | 2017-04-05 | 富士通株式会社 | Notification control method, notification control device, and notification control program |
| US9478054B1 (en) * | 2013-11-09 | 2016-10-25 | Google Inc. | Image overlay compositing |
| US9464886B2 (en) | 2013-11-21 | 2016-10-11 | Ford Global Technologies, Llc | Luminescent hitch angle detection component |
| US9464887B2 (en) | 2013-11-21 | 2016-10-11 | Ford Global Technologies, Llc | Illuminated hitch angle detection component |
| US10017114B2 (en) | 2014-02-19 | 2018-07-10 | Magna Electronics Inc. | Vehicle vision system with display |
| US9296421B2 (en) | 2014-03-06 | 2016-03-29 | Ford Global Technologies, Llc | Vehicle target identification using human gesture recognition |
| JP6284408B2 (en) * | 2014-04-03 | 2018-02-28 | オリンパス株式会社 | Image processing apparatus, imaging apparatus, determination method, driving method, imaging method, and program |
| WO2015183889A1 (en) * | 2014-05-27 | 2015-12-03 | Robert Bosch Gmbh | Detection, identification, and mitigation of lens contamination for vehicle mounted camera systems |
| US9344638B2 (en) * | 2014-05-30 | 2016-05-17 | Apple Inc. | Constant bracket high dynamic range (cHDR) operations |
| US9380218B2 (en) | 2014-05-30 | 2016-06-28 | Apple Inc. | Highlight exposure metric and its applications |
| US10112537B2 (en) | 2014-09-03 | 2018-10-30 | Ford Global Technologies, Llc | Trailer angle detection target fade warning |
| US9607242B2 (en) | 2015-01-16 | 2017-03-28 | Ford Global Technologies, Llc | Target monitoring system with lens cleaning device |
| JP6393653B2 (en) * | 2015-04-09 | 2018-09-19 | 株式会社東海理化電機製作所 | Vehicle visual recognition device |
| IL239129A0 (en) | 2015-06-01 | 2015-11-30 | Brightway Vision Ltd | Image enhancements for vehicle imaging systems |
| US9836060B2 (en) | 2015-10-28 | 2017-12-05 | Ford Global Technologies, Llc | Trailer backup assist system with target management |
| US9610975B1 (en) | 2015-12-17 | 2017-04-04 | Ford Global Technologies, Llc | Hitch angle detection for trailer backup assist system |
| GB201604936D0 (en) * | 2016-03-23 | 2016-05-04 | Jaguar Land Rover Ltd | Adaptive display for low visibility |
| CN105799594B (en) * | 2016-04-14 | 2019-03-12 | 京东方科技集团股份有限公司 | Image display method, vehicle display device, sun visor and automobile |
| JP6901275B2 (en) * | 2017-02-14 | 2021-07-14 | 本田技研工業株式会社 | Information processing server, client and information processing system |
| JP6680254B2 (en) * | 2017-03-22 | 2020-04-15 | トヨタ自動車株式会社 | Vehicle display control device |
| DE102017215051A1 (en) * | 2017-08-29 | 2019-02-28 | Conti Temic Microelectronic Gmbh | Apparatus and method for reducing the influence of stray light and reflections on optical image recognition |
| DE102017215050A1 (en) * | 2017-08-29 | 2019-02-28 | Conti Temic Microelectronic Gmbh | Apparatus and method for reducing the influence of shadows and low contrast on optical image recognition |
| US10710585B2 (en) | 2017-09-01 | 2020-07-14 | Ford Global Technologies, Llc | Trailer backup assist system with predictive hitch angle functionality |
| US20190126941A1 (en) * | 2017-10-31 | 2019-05-02 | Wipro Limited | Method and system of stitching frames to assist driver of a vehicle |
| US10678255B2 (en) | 2018-02-14 | 2020-06-09 | GM Global Technology Operations LLC | Systems, methods and apparatuses are provided for enhanced surface condition detection based on image scene and ambient light analysis |
| US10762611B2 (en) * | 2018-08-07 | 2020-09-01 | Sensors Unlimited, Inc. | Scaled two-band histogram process for image enhancement |
| US11089239B1 (en) * | 2020-05-19 | 2021-08-10 | GM Global Technology Operations LLC | System and method to modify undercarriage camera image feed |
| US11501452B2 (en) | 2020-08-10 | 2022-11-15 | Honeywell International Inc. | Machine learning and vision-based approach to zero velocity update object detection |
| GB202107132D0 (en) * | 2021-05-19 | 2021-06-30 | Agco Int Gmbh | Residue spread monitoring |
| CN113012079B (en) * | 2021-05-25 | 2021-08-03 | 南京索安电子有限公司 | Low-brightness vehicle bottom image enhancement method and device and storage medium |
| US11700458B2 (en) | 2021-08-06 | 2023-07-11 | Ford Global Technologies, Llc | White balance and color correction for interior vehicle camera |
| US12371046B2 (en) * | 2023-12-20 | 2025-07-29 | Fca Us Llc | Displaying objects to assist a driver in conditions of low visibility |
Family Cites Families (249)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US5170374A (en) | 1981-05-13 | 1992-12-08 | Hitachi, Ltd. | Semiconductor memory |
| US6735506B2 (en) | 1992-05-05 | 2004-05-11 | Automotive Technologies International, Inc. | Telematics system |
| US5845000A (en) | 1992-05-05 | 1998-12-01 | Automotive Technologies International, Inc. | Optical identification and monitoring system using pattern recognition for use with vehicles |
| US6442465B2 (en) | 1992-05-05 | 2002-08-27 | Automotive Technologies International, Inc. | Vehicular component control systems and methods |
| US5001558A (en) | 1985-06-11 | 1991-03-19 | General Motors Corporation | Night vision system with color video camera |
| JPH01173825A (en) | 1987-12-28 | 1989-07-10 | Aisin Aw Co Ltd | Navigation device for vehicle |
| IT1219405B (en) | 1988-06-27 | 1990-05-11 | Fiat Ricerche | PROCEDURE AND DEVICE FOR INSTRUMENTAL VISION IN POOR CONDITIONS VISIBILITY IN PARTICULAR FOR DRIVING IN THE MIST |
| US5003288A (en) | 1988-10-25 | 1991-03-26 | Nartron Corporation | Ambient light sensing method and apparatus |
| US5614885A (en) | 1988-12-05 | 1997-03-25 | Prince Corporation | Electrical control system for vehicle options |
| FR2642855B1 (en) | 1989-02-06 | 1991-05-17 | Essilor Int | OPTICAL LENS FOR THE CORRECTION OF ASTIGMATISM |
| JPH0749925B2 (en) | 1989-03-01 | 1995-05-31 | 浜松ホトニクス株式会社 | Two-dimensional incident position detector |
| US4973844A (en) | 1989-07-10 | 1990-11-27 | Donnelly Corporation | Vehicular moisture sensor and mounting apparatus therefor |
| US5097362A (en) | 1989-07-19 | 1992-03-17 | Lynas Robert M | Rearview mirror targeting and repositioning system |
| US5027001A (en) | 1989-08-29 | 1991-06-25 | Torbert William F | Moisture sensitive automatic windshield wiper and headlight control device |
| US4982287A (en) * | 1989-11-01 | 1991-01-01 | Rca Licensing Corporation | Control of brightness level adapted to control of contrast |
| US4987357A (en) | 1989-12-18 | 1991-01-22 | General Motors Corporation | Adaptive motor vehicle cruise control |
| JP2843079B2 (en) | 1989-12-22 | 1999-01-06 | 本田技研工業株式会社 | Driving path determination method |
| US5059877A (en) | 1989-12-22 | 1991-10-22 | Libbey-Owens-Ford Co. | Rain responsive windshield wiper control |
| FR2658642B1 (en) | 1990-02-20 | 1994-06-10 | Rousseau Codes | METHOD AND DEVICE FOR DRIVING DRIVING LAND VEHICLES. |
| US5303205A (en) | 1990-02-26 | 1994-04-12 | Trend Tec Inc. | Vehicular distance measuring system with integral mirror display |
| JP2920653B2 (en) | 1990-03-15 | 1999-07-19 | アイシン精機株式会社 | In-vehicle imaging device |
| DE4111993B4 (en) | 1990-04-23 | 2005-05-25 | Volkswagen Ag | Camera for an image processing system |
| US5121200A (en) | 1990-07-06 | 1992-06-09 | Choi Seung Lyul | Travelling monitoring system for motor vehicles |
| US5027200A (en) | 1990-07-10 | 1991-06-25 | Edward Petrossian | Enhanced viewing at side and rear of motor vehicles |
| US5177685A (en) | 1990-08-09 | 1993-01-05 | Massachusetts Institute Of Technology | Automobile navigation system using real time spoken driving instructions |
| US5086253A (en) | 1990-10-15 | 1992-02-04 | Lawler Louis N | Automatic headlight dimmer apparatus |
| US5309137A (en) | 1991-02-26 | 1994-05-03 | Mitsubishi Denki Kabushiki Kaisha | Motor car traveling control device |
| US5451822A (en) | 1991-03-15 | 1995-09-19 | Gentex Corporation | Electronic control system |
| KR930001987Y1 (en) | 1991-03-28 | 1993-04-19 | 홍선택 | Automotive Back Mirror Turner |
| WO1992018848A1 (en) | 1991-04-23 | 1992-10-29 | Introlab Pty. Limited | A moisture sensor |
| US5182502A (en) | 1991-05-06 | 1993-01-26 | Lectron Products, Inc. | Automatic headlamp dimmer |
| US5245422A (en) | 1991-06-28 | 1993-09-14 | Zexel Corporation | System and method for automatically steering a vehicle within a lane in a road |
| JP2782990B2 (en) | 1991-07-11 | 1998-08-06 | 日産自動車株式会社 | Vehicle approach determination device |
| US5469298A (en) | 1991-08-14 | 1995-11-21 | Prince Corporation | Reflective display at infinity |
| JPH0554276A (en) | 1991-08-23 | 1993-03-05 | Matsushita Electric Ind Co Ltd | Obstacle detection device |
| US5193000A (en) | 1991-08-28 | 1993-03-09 | Stereographics Corporation | Multiplexing technique for stereoscopic video system |
| US5416318A (en) | 1991-10-03 | 1995-05-16 | Hegyi; Dennis J. | Combined headlamp and climate control sensor having a light diffuser and a light modulator |
| FR2682792B1 (en) | 1991-10-16 | 1995-10-20 | Ii Bc Sys | DEVICE FOR AVOIDING CARAMBOLAGES IN CHAIN. |
| JP3167752B2 (en) | 1991-10-22 | 2001-05-21 | 富士重工業株式会社 | Vehicle distance detection device |
| US5535314A (en) | 1991-11-04 | 1996-07-09 | Hughes Aircraft Company | Video image processor and method for detecting vehicles |
| JP3031013B2 (en) | 1991-11-15 | 2000-04-10 | 日産自動車株式会社 | Visual information providing device |
| US5336980A (en) | 1992-12-10 | 1994-08-09 | Leopold Kostal Gmbh & Co. | Apparatus and method for controlling a windshield wiping system |
| US5276389A (en) | 1991-12-14 | 1994-01-04 | Leopold Kostal Gmbh & Co. Kg | Method of controlling a windshield wiper system |
| US5394333A (en) | 1991-12-23 | 1995-02-28 | Zexel Usa Corp. | Correcting GPS position in a hybrid naviation system |
| US5208701A (en) | 1991-12-24 | 1993-05-04 | Xerox Corporation | Wobble correction lens with binary diffractive optic surface and refractive cylindrical surface |
| US5461357A (en) | 1992-01-29 | 1995-10-24 | Mazda Motor Corporation | Obstacle detection device for vehicle |
| JP2800531B2 (en) | 1992-02-28 | 1998-09-21 | 三菱電機株式会社 | Obstacle detection device for vehicles |
| JP2973695B2 (en) | 1992-03-12 | 1999-11-08 | 船井電機株式会社 | In-vehicle navigation system |
| JPH05265547A (en) | 1992-03-23 | 1993-10-15 | Fuji Heavy Ind Ltd | On-vehicle outside monitoring device |
| US5204778A (en) | 1992-04-06 | 1993-04-20 | Gentex Corporation | Control system for automatic rearview mirrors |
| US5325386A (en) | 1992-04-21 | 1994-06-28 | Bandgap Technology Corporation | Vertical-cavity surface emitting laser assay display system |
| EP0567660B2 (en) | 1992-04-21 | 2000-09-06 | IBP Pietzsch GmbH | Device for the guiding of vehicles |
| GB2267341B (en) | 1992-05-27 | 1996-02-21 | Koito Mfg Co Ltd | Glare sensor for a vehicle |
| US5515448A (en) | 1992-07-28 | 1996-05-07 | Yazaki Corporation | Distance measuring apparatus of a target tracking type |
| JPH0785280B2 (en) | 1992-08-04 | 1995-09-13 | タカタ株式会社 | Collision prediction judgment system by neural network |
| US5351044A (en) | 1992-08-12 | 1994-09-27 | Rockwell International Corporation | Vehicle lane position detection system |
| BR9306901A (en) | 1992-08-14 | 1998-12-08 | Vorad Safety Systems Inc | Recording of operational events in an automotive vehicle |
| ATE181602T1 (en) | 1992-08-14 | 1999-07-15 | Vorad Safety Systems Inc | INTELLIGENT BLIND SPOT DETECTION SENSOR |
| JP2783079B2 (en) | 1992-08-28 | 1998-08-06 | トヨタ自動車株式会社 | Light distribution control device for headlamp |
| US5448319A (en) | 1992-09-22 | 1995-09-05 | Olympus Optical Co., Ltd. | Optical system for monitor cameras to be mounted on vehicles |
| DE4332612C2 (en) | 1992-09-25 | 1996-02-22 | Yazaki Corp | Exterior view monitoring method for motor vehicles |
| JP3462227B2 (en) | 1992-11-13 | 2003-11-05 | 矢崎総業株式会社 | Display device for vehicles |
| JP3418985B2 (en) | 1992-12-14 | 2003-06-23 | 株式会社デンソー | Image display device |
| US5285060A (en) | 1992-12-15 | 1994-02-08 | Donnelly Corporation | Display for automatic rearview mirror |
| JP3263699B2 (en) | 1992-12-22 | 2002-03-04 | 三菱電機株式会社 | Driving environment monitoring device |
| KR940017747A (en) | 1992-12-29 | 1994-07-27 | 에프. 제이. 스미트 | Image processing device |
| JPH06213660A (en) | 1993-01-19 | 1994-08-05 | Aisin Seiki Co Ltd | Detecting method for approximate straight line of image |
| US5529138A (en) | 1993-01-22 | 1996-06-25 | Shaw; David C. H. | Vehicle collision avoidance system |
| US5289321A (en) | 1993-02-12 | 1994-02-22 | Secor James O | Consolidated rear view camera and display system for motor vehicle |
| US5313072A (en) | 1993-02-16 | 1994-05-17 | Rockwell International Corporation | Optical detector for windshield wiper control |
| US5670935A (en) | 1993-02-26 | 1997-09-23 | Donnelly Corporation | Rearview vision system for vehicle including panoramic view |
| US6396397B1 (en) | 1993-02-26 | 2002-05-28 | Donnelly Corporation | Vehicle imaging system with stereo imaging |
| US5796094A (en) | 1993-02-26 | 1998-08-18 | Donnelly Corporation | Vehicle headlight control using imaging sensor |
| US5877897A (en) | 1993-02-26 | 1999-03-02 | Donnelly Corporation | Automatic rearview mirror, vehicle lighting control and vehicle interior monitoring system using a photosensor array |
| US6498620B2 (en) | 1993-02-26 | 2002-12-24 | Donnelly Corporation | Vision system for a vehicle including an image capture device and a display system having a long focal length |
| US5550677A (en) | 1993-02-26 | 1996-08-27 | Donnelly Corporation | Automatic rearview mirror system using a photosensor array |
| JP3468428B2 (en) | 1993-03-24 | 2003-11-17 | 富士重工業株式会社 | Vehicle distance detection device |
| JP2887039B2 (en) | 1993-03-26 | 1999-04-26 | 三菱電機株式会社 | Vehicle periphery monitoring device |
| DE4408745C2 (en) | 1993-03-26 | 1997-02-27 | Honda Motor Co Ltd | Driving control device for vehicles |
| US6430303B1 (en) | 1993-03-31 | 2002-08-06 | Fujitsu Limited | Image processing apparatus |
| DE4492128T1 (en) | 1993-03-31 | 1996-06-27 | Automotive Tech Int | Position and speed sensor for vehicle occupants |
| US6084519A (en) | 1993-05-07 | 2000-07-04 | Control Devices, Inc. | Multi-function light sensor for vehicle |
| DE4318114C2 (en) | 1993-06-01 | 1998-07-16 | Kostal Leopold Gmbh & Co Kg | Sensor device |
| US6553130B1 (en) | 1993-08-11 | 2003-04-22 | Jerome H. Lemelson | Motor vehicle warning and control system and method |
| US5434407A (en) | 1993-08-23 | 1995-07-18 | Gentex Corporation | Automatic rearview mirror incorporating light pipe |
| GB9317983D0 (en) | 1993-08-28 | 1993-10-13 | Lucas Ind Plc | A driver assistance system for a vehicle |
| US5586063A (en) | 1993-09-01 | 1996-12-17 | Hardin; Larry C. | Optical range and speed detection system |
| US5638116A (en) | 1993-09-08 | 1997-06-10 | Sumitomo Electric Industries, Ltd. | Object recognition apparatus and method |
| US5374852A (en) | 1993-09-17 | 1994-12-20 | Parkes; Walter B. | Motor vehicle headlight activation apparatus for inclement weather conditions |
| US5440428A (en) | 1993-09-30 | 1995-08-08 | Hughes Aircraft Company | Automotive instrument 3-D virtual image display |
| US5883739A (en) | 1993-10-04 | 1999-03-16 | Honda Giken Kogyo Kabushiki Kaisha | Information display device for vehicle |
| US5406395A (en) | 1993-11-01 | 1995-04-11 | Hughes Aircraft Company | Holographic parking assistance device |
| JP3522317B2 (en) | 1993-12-27 | 2004-04-26 | 富士重工業株式会社 | Travel guide device for vehicles |
| US5430431A (en) | 1994-01-19 | 1995-07-04 | Nelson; Louis J. | Vehicle protection system and method |
| US5471515A (en) | 1994-01-28 | 1995-11-28 | California Institute Of Technology | Active pixel sensor with intra-pixel charge transfer |
| JP3358099B2 (en) | 1994-03-25 | 2002-12-16 | オムロン株式会社 | Optical sensor device |
| US5666028A (en) | 1994-04-06 | 1997-09-09 | Gentex Corporation | Automobile headlamp and running light control system |
| US5537003A (en) | 1994-04-08 | 1996-07-16 | Gentex Corporation | Control system for automotive vehicle headlamps and other vehicle equipment |
| FR2718874B1 (en) | 1994-04-15 | 1996-05-15 | Thomson Csf | Traffic monitoring method for automatic detection of vehicle incidents. |
| US5963247A (en) | 1994-05-31 | 1999-10-05 | Banitt; Shmuel | Visual display systems and a system for producing recordings for visualization thereon and methods therefor |
| ES1028357Y (en) | 1994-06-03 | 1995-06-16 | Cortes Luis Leon Lamata | RECEIVING DEVICE FOR REAR VIEW SCREEN. |
| US5574443A (en) | 1994-06-22 | 1996-11-12 | Hsieh; Chi-Sheng | Vehicle monitoring apparatus with broadly and reliably rearward viewing |
| JP3287117B2 (en) | 1994-07-05 | 2002-05-27 | 株式会社日立製作所 | Environment recognition device for vehicles using imaging device |
| JP3357749B2 (en) | 1994-07-12 | 2002-12-16 | 本田技研工業株式会社 | Vehicle road image processing device |
| US5793420A (en) | 1994-10-28 | 1998-08-11 | Schmidt; William P. | Video recording system for vehicle |
| US5732379A (en) | 1994-11-25 | 1998-03-24 | Itt Automotive Europe Gmbh | Brake system for a motor vehicle with yaw moment control |
| US5677851A (en) | 1994-12-15 | 1997-10-14 | Novell, Inc. | Method and apparatus to secure digital directory object changes |
| JPH08175263A (en) | 1994-12-27 | 1996-07-09 | Murakami Kaimeidou:Kk | Interior mirror with built-in display device |
| US5614788A (en) | 1995-01-31 | 1997-03-25 | Autosmart Light Switches, Inc. | Automated ambient condition responsive daytime running light system |
| US5528698A (en) | 1995-03-27 | 1996-06-18 | Rockwell International Corporation | Automotive occupant sensing device |
| JP2885125B2 (en) | 1995-03-30 | 1999-04-19 | トヨタ自動車株式会社 | Estimation method of motion state quantity changing with turning of vehicle |
| JP3539788B2 (en) | 1995-04-21 | 2004-07-07 | パナソニック モバイルコミュニケーションズ株式会社 | Image matching method |
| US5500766A (en) | 1995-05-04 | 1996-03-19 | Stonecypher; Bob | Blind spot side mirror |
| US5568027A (en) | 1995-05-19 | 1996-10-22 | Libbey-Owens-Ford Co. | Smooth rain-responsive wiper control |
| US5737226A (en) | 1995-06-05 | 1998-04-07 | Prince Corporation | Vehicle compass system with automatic calibration |
| US7202776B2 (en) | 1997-10-22 | 2007-04-10 | Intelligent Technologies International, Inc. | Method and system for detecting objects external to a vehicle |
| US7085637B2 (en) | 1997-10-22 | 2006-08-01 | Intelligent Technologies International, Inc. | Method and system for controlling a vehicle |
| US5915800A (en) | 1995-06-19 | 1999-06-29 | Fuji Jukogyo Kabushiki Kaisha | System for controlling braking of an automotive vehicle |
| JP3546600B2 (en) | 1995-09-07 | 2004-07-28 | トヨタ自動車株式会社 | Light distribution control device for headlamp |
| US5724316A (en) | 1995-09-26 | 1998-03-03 | Delco Electronics Corporation | GPS based time determining system and method |
| US5878370A (en) | 1995-12-01 | 1999-03-02 | Prince Corporation | Vehicle compass system with variable resolution |
| US6266082B1 (en) | 1995-12-19 | 2001-07-24 | Canon Kabushiki Kaisha | Communication apparatus image processing apparatus communication method and image processing method |
| US5790973A (en) | 1995-12-19 | 1998-08-04 | Prince Corporation | Last exit warning system |
| US5761094A (en) | 1996-01-18 | 1998-06-02 | Prince Corporation | Vehicle compass system |
| US5786772A (en) | 1996-03-22 | 1998-07-28 | Donnelly Corporation | Vehicle blind spot detection display system |
| US5661303A (en) | 1996-05-24 | 1997-08-26 | Libbey-Owens-Ford Co. | Compact moisture sensor with collimator lenses and prismatic coupler |
| US6550949B1 (en) | 1996-06-13 | 2003-04-22 | Gentex Corporation | Systems and components for enhancing rear vision from a vehicle |
| DE19624046A1 (en) | 1996-06-17 | 1997-12-18 | Bayerische Motoren Werke Ag | Method and device for indicating the braking strength or deceleration in a vehicle |
| JP3805832B2 (en) | 1996-07-10 | 2006-08-09 | 富士重工業株式会社 | Vehicle driving support device |
| JPH1059068A (en) | 1996-08-23 | 1998-03-03 | Yoshihisa Furuta | Dead angle confirmation device for vehicle |
| US5878357A (en) | 1996-09-03 | 1999-03-02 | Ford Global Technologies, Inc. | Method and apparatus for vehicle yaw rate estimation |
| US5924212A (en) | 1996-10-09 | 1999-07-20 | Donnelly Corporation | Electronic compass |
| JPH10161013A (en) | 1996-12-05 | 1998-06-19 | Canon Inc | Environment recognition device and camera equipped with environment recognition device |
| NZ331543A (en) | 1996-12-10 | 2000-05-26 | Touchsensor Technologies L | Differential proximity sensor and output circuit |
| US5877707A (en) | 1997-01-17 | 1999-03-02 | Kowalick; Thomas M. | GPS based seat belt monitoring system & method for using same |
| US5844505A (en) | 1997-04-01 | 1998-12-01 | Sony Corporation | Automobile navigation system |
| US5837994C1 (en) | 1997-04-02 | 2001-10-16 | Gentex Corp | Control system to automatically dim vehicle head lamps |
| US5990469A (en) | 1997-04-02 | 1999-11-23 | Gentex Corporation | Control circuit for image array sensors |
| US6049171A (en) | 1998-09-18 | 2000-04-11 | Gentex Corporation | Continuously variable headlamp control |
| US6587573B1 (en) | 2000-03-20 | 2003-07-01 | Gentex Corporation | System for controlling exterior vehicle lights |
| US6611610B1 (en) | 1997-04-02 | 2003-08-26 | Gentex Corporation | Vehicle lamp control |
| US6631316B2 (en) | 2001-03-05 | 2003-10-07 | Gentex Corporation | Image processing system to control vehicle headlamps or other vehicle equipment |
| US5923027A (en) | 1997-09-16 | 1999-07-13 | Gentex Corporation | Moisture sensor and windshield fog detector using an image sensor |
| JP3508909B2 (en) | 1997-07-01 | 2004-03-22 | 株式会社村上開明堂 | Rearview mirror quick deflection controller |
| US6353392B1 (en) | 1997-10-30 | 2002-03-05 | Donnelly Corporation | Rain sensor with fog discrimination |
| US6313454B1 (en) | 1999-07-02 | 2001-11-06 | Donnelly Corporation | Rain sensor |
| US6020704A (en) | 1997-12-02 | 2000-02-01 | Valeo Electrical Systems, Inc. | Windscreen sensing and wiper control system |
| US6124647A (en) | 1998-12-16 | 2000-09-26 | Donnelly Corporation | Information display in a rearview mirror |
| US6294989B1 (en) | 1998-12-16 | 2001-09-25 | Donnelly Corporation | Tire inflation assistance monitoring system |
| DE19812237C1 (en) | 1998-03-20 | 1999-09-23 | Daimler Chrysler Ag | Method for driving dynamics control on a road vehicle |
| US5899956A (en) | 1998-03-31 | 1999-05-04 | Advanced Future Technologies, Inc. | Vehicle mounted navigation device |
| US6477464B2 (en) | 2000-03-09 | 2002-11-05 | Donnelly Corporation | Complete mirror-based global-positioning system (GPS) navigation solution |
| US6175300B1 (en) | 1998-09-03 | 2001-01-16 | Byron K. Kendrick | Blind spot viewing system |
| US6066933A (en) | 1998-10-02 | 2000-05-23 | Ponziana; Richard L. | Rain sensing system and method having automatically registered and oriented rain sensor |
| US6266442B1 (en) | 1998-10-23 | 2001-07-24 | Facet Technology Corp. | Method and apparatus for identifying objects depicted in a videostream |
| US6201642B1 (en) | 1999-07-27 | 2001-03-13 | Donnelly Corporation | Vehicular vision system with a wide angle lens including a diffractive element |
| US6320282B1 (en) | 1999-01-19 | 2001-11-20 | Touchsensor Technologies, Llc | Touch switch with integral control circuit |
| DE19902081A1 (en) | 1999-01-20 | 2000-07-27 | Zeiss Carl Fa | Stabilized camera |
| US6166698A (en) | 1999-02-16 | 2000-12-26 | Gentex Corporation | Rearview mirror with integrated microwave receiver |
| US6144022A (en) | 1999-03-15 | 2000-11-07 | Valeo Electrical Systems, Inc. | Rain sensor using statistical analysis |
| US6333759B1 (en) | 1999-03-16 | 2001-12-25 | Joseph J. Mazzilli | 360 ° automobile video camera system |
| US6392315B1 (en) | 1999-04-05 | 2002-05-21 | Delphi Technologies, Inc. | Compensation circuit for an automotive ignition sensing system |
| EP2410742A1 (en) | 1999-04-16 | 2012-01-25 | Panasonic Corporation | Image processing apparatus and monitoring system |
| US6795221B1 (en) | 1999-08-05 | 2004-09-21 | Microvision, Inc. | Scanned display with switched feeds and distortion correction |
| US6411204B1 (en) | 1999-11-15 | 2002-06-25 | Donnelly Corporation | Deceleration based anti-collision safety light control for vehicle |
| US6704621B1 (en) | 1999-11-26 | 2004-03-09 | Gideon P. Stein | System and method for estimating ego-motion of a moving vehicle using successive images recorded along the vehicle's path of motion |
| SE520360C2 (en) | 1999-12-15 | 2003-07-01 | Goeran Sjoenell | Warning device for vehicles |
| US6526335B1 (en) | 2000-01-24 | 2003-02-25 | G. Victor Treyz | Automobile personal computer systems |
| JP2001213254A (en) | 2000-01-31 | 2001-08-07 | Yazaki Corp | Vehicle side monitoring system |
| AU2001243285A1 (en) | 2000-03-02 | 2001-09-12 | Donnelly Corporation | Video mirror systems incorporating an accessory module |
| US7167796B2 (en) | 2000-03-09 | 2007-01-23 | Donnelly Corporation | Vehicle navigation system for use with a telematics system |
| KR100373002B1 (en) | 2000-04-03 | 2003-02-25 | 현대자동차주식회사 | Method for judgment out of lane of vehicle |
| US7365769B1 (en) | 2000-07-06 | 2008-04-29 | Donald Mager | Activating a vehicle's own brake lights and/or brakes when brake lights are sensed in front of the vehicle, including responsively to the proximity of, and/or rate of closure with, a forward vehicle |
| GB2369452B (en) | 2000-07-27 | 2002-07-17 | Michael John Downs | Beam splitting blocks |
| JP3521860B2 (en) | 2000-10-02 | 2004-04-26 | 日産自動車株式会社 | Vehicle travel path recognition device |
| US7062300B1 (en) | 2000-11-09 | 2006-06-13 | Ki Il Kim | Cellular phone holder with charger mounted to vehicle dashboard |
| US6672731B2 (en) | 2000-11-20 | 2004-01-06 | Donnelly Corporation | Vehicular rearview mirror with blind spot viewing system |
| AU2002251807A1 (en) | 2001-01-23 | 2002-08-19 | Donnelly Corporation | Improved vehicular lighting system for a mirror assembly |
| US20020113873A1 (en) | 2001-02-20 | 2002-08-22 | Williams Michael R. | Rear vision system for large vehicles |
| US6424273B1 (en) | 2001-03-30 | 2002-07-23 | Koninklijke Philips Electronics N.V. | System to aid a driver to determine whether to change lanes |
| DE10118265A1 (en) | 2001-04-12 | 2002-10-17 | Bosch Gmbh Robert | Detecting vehicle lane change, involves forming track change indicating signal by comparing measured angular rate of preceding vehicle(s) with vehicle's own yaw rate |
| DE20106977U1 (en) | 2001-04-23 | 2002-08-29 | Mekra Lang Gmbh & Co Kg | Warning device in motor vehicles |
| US6539306B2 (en) | 2001-06-15 | 2003-03-25 | Gentex Corporation | Automotive mirror with integrated Loran components |
| US6497503B1 (en) | 2001-06-21 | 2002-12-24 | Ford Global Technologies, Inc. | Headlamp system with selectable beam pattern |
| WO2003029046A1 (en) | 2001-10-03 | 2003-04-10 | Maryann Winter | Apparatus and method for sensing the occupancy status of parking spaces in a parking lot |
| US6636258B2 (en) | 2001-10-19 | 2003-10-21 | Ford Global Technologies, Llc | 360° vision system for a vehicle |
| US6909753B2 (en) | 2001-12-05 | 2005-06-21 | Koninklijke Philips Electronics, N.V. | Combined MPEG-4 FGS and modulation algorithm for wireless video transmission |
| US7543946B2 (en) * | 2002-01-10 | 2009-06-09 | Gentex Corporation | Dimmable rearview assembly having a glare sensor |
| US20030137586A1 (en) | 2002-01-22 | 2003-07-24 | Infinite Innovations, Inc. | Vehicle video switching system and method |
| WO2003065084A1 (en) | 2002-01-31 | 2003-08-07 | Donnelly Corporation | Vehicle accessory module |
| EP1332923B1 (en) | 2002-02-05 | 2007-07-11 | Donnelly Hohe GmbH & Co. KG | Manoeuvring and/or parking aid device for a vehicle |
| US6730913B2 (en) | 2002-02-21 | 2004-05-04 | Ford Global Technologies, Llc | Active night vision system for vehicles employing short-pulse laser illumination and a gated camera for image capture |
| US6975775B2 (en) | 2002-03-06 | 2005-12-13 | Radiant Imaging, Inc. | Stray light correction method for imaging light and color measurement system |
| US20030222982A1 (en) | 2002-03-28 | 2003-12-04 | Hamdan Majil M. | Integrated video/data information system and method for application to commercial vehicles to enhance driver awareness |
| US7145519B2 (en) | 2002-04-18 | 2006-12-05 | Nissan Motor Co., Ltd. | Image display apparatus, method, and program for automotive vehicle |
| US7004606B2 (en) | 2002-04-23 | 2006-02-28 | Donnelly Corporation | Automatic headlamp control |
| US6946978B2 (en) | 2002-04-25 | 2005-09-20 | Donnelly Corporation | Imaging system for vehicle |
| US7123168B2 (en) | 2002-04-25 | 2006-10-17 | Donnelly Corporation | Driving separation distance indicator |
| ES2391556T3 (en) | 2002-05-03 | 2012-11-27 | Donnelly Corporation | Object detection system for vehicles |
| EP1540373B1 (en) * | 2002-08-05 | 2008-02-20 | Elbit Systems Ltd. | Vehicle mounted night vision imaging system and method |
| DE20214892U1 (en) | 2002-09-25 | 2002-11-21 | Donnelly Hohe GmbH & Co. KG, 97903 Collenberg | Monitoring device for a motor vehicle |
| WO2004047421A2 (en) * | 2002-11-14 | 2004-06-03 | Donnelly Corporation | Imaging system for vehicle |
| US7136753B2 (en) | 2002-12-05 | 2006-11-14 | Denso Corporation | Object recognition apparatus for vehicle, inter-vehicle control apparatus, and distance measurement apparatus |
| US7541743B2 (en) | 2002-12-13 | 2009-06-02 | Ford Global Technologies, Llc | Adaptive vehicle communication controlled lighting system |
| DE10346508B4 (en) | 2003-10-02 | 2007-10-11 | Daimlerchrysler Ag | Device for improving the visibility in a motor vehicle |
| CN100408398C (en) | 2003-10-28 | 2008-08-06 | 大陆-特韦斯贸易合伙股份公司及两合公司 | Method and system for improving driving performance of vehicle |
| US7526103B2 (en) | 2004-04-15 | 2009-04-28 | Donnelly Corporation | Imaging system for vehicle |
| US7576767B2 (en) * | 2004-07-26 | 2009-08-18 | Geo Semiconductors Inc. | Panoramic vision system and method |
| US7227611B2 (en) | 2004-08-23 | 2007-06-05 | The Boeing Company | Adaptive and interactive scene illumination |
| US7881496B2 (en) | 2004-09-30 | 2011-02-01 | Donnelly Corporation | Vision system for vehicle |
| US20060103727A1 (en) | 2004-11-17 | 2006-05-18 | Huan-Chin Tseng | Vehicle back up camera |
| US7720580B2 (en) | 2004-12-23 | 2010-05-18 | Donnelly Corporation | Object detection system for vehicle |
| US20060164221A1 (en) | 2005-01-18 | 2006-07-27 | Jensen John M | Sensor-activated controlled safety or warning light mounted on or facing toward rear of vehicle |
| US7952490B2 (en) | 2005-02-22 | 2011-05-31 | Continental Temic Microelectronic GmbH | Method for identifying the activation of the brake lights of preceding vehicles |
| US20060250501A1 (en) | 2005-05-06 | 2006-11-09 | Widmann Glenn R | Vehicle security monitor system and method |
| JP2006341641A (en) | 2005-06-07 | 2006-12-21 | Nissan Motor Co Ltd | Video display device and video display method |
| JP4580288B2 (en) | 2005-06-28 | 2010-11-10 | 本田技研工業株式会社 | Driving assistance car |
| US7492962B2 (en) * | 2005-08-25 | 2009-02-17 | Delphi Technologies, Inc. | System or method for enhancing an image |
| US7460951B2 (en) | 2005-09-26 | 2008-12-02 | Gm Global Technology Operations, Inc. | System and method of target tracking using sensor fusion |
| CN101816008A (en) | 2005-10-28 | 2010-08-25 | 马格纳电子系统公司 | Camera module for vehicle vision system |
| EP1949666B1 (en) | 2005-11-01 | 2013-07-17 | Magna Mirrors of America, Inc. | Interior rearview mirror with display |
| JP2007129525A (en) | 2005-11-04 | 2007-05-24 | Konica Minolta Photo Imaging Inc | Camera system and controller |
| DE602006019156D1 (en) | 2005-12-27 | 2011-02-03 | Honda Motor Co Ltd | Vehicle and steering control device for a vehicle |
| JP4462231B2 (en) | 2006-05-09 | 2010-05-12 | 株式会社デンソー | Auto light device for vehicle |
| US7724962B2 (en) | 2006-07-07 | 2010-05-25 | Siemens Corporation | Context adaptive approach in vehicle detection under various visibility conditions |
| EP2122599B1 (en) | 2007-01-25 | 2019-11-13 | Magna Electronics Inc. | Radar sensing system for vehicle |
| US7804421B2 (en) * | 2007-02-21 | 2010-09-28 | Audiovox Corporation | Vehicle safety system |
| JP4497231B2 (en) | 2007-10-09 | 2010-07-07 | 株式会社デンソー | Vehicle speed control device |
| TWI372564B (en) | 2007-10-30 | 2012-09-11 | Av Tech Corp | Video system, image emission apparatus, video receiver apparatus and control method |
| US8027029B2 (en) | 2007-11-07 | 2011-09-27 | Magna Electronics Inc. | Object detection and tracking system |
| DE102008003194A1 (en) | 2008-01-04 | 2009-07-09 | Wabco Gmbh | Driver assistance system |
| US8154418B2 (en) | 2008-03-31 | 2012-04-10 | Magna Mirrors Of America, Inc. | Interior rearview mirror system |
| US20090265069A1 (en) | 2008-04-17 | 2009-10-22 | Herman Desbrunes | Land vehicle braking system |
| US20100020170A1 (en) | 2008-07-24 | 2010-01-28 | Higgins-Luthman Michael J | Vehicle Imaging System |
| WO2010088465A1 (en) * | 2009-02-02 | 2010-08-05 | Gentex Corporation | Improved digital image processing and systems incorporating the same |
| WO2010099416A1 (en) | 2009-02-27 | 2010-09-02 | Magna Electronics | Alert system for vehicle |
| US9036026B2 (en) | 2009-06-12 | 2015-05-19 | Magna Electronics | Scalable integrated electronic control unit for vehicle |
| EP2423063B1 (en) | 2010-08-23 | 2013-03-06 | Harman Becker Automotive Systems GmbH | Method of detecting the braking of a vehicle |
| US9194943B2 (en) | 2011-04-12 | 2015-11-24 | Magna Electronics Inc. | Step filter for estimating distance in a time-of-flight ranging system |
| US9681062B2 (en) | 2011-09-26 | 2017-06-13 | Magna Electronics Inc. | Vehicle camera image quality improvement in poor visibility conditions by contrast amplification |
| DE102011118157A1 (en) | 2011-11-10 | 2013-05-16 | GM Global Technology Operations LLC (n. d. Gesetzen des Staates Delaware) | Method for operating an information and entertainment system of a motor vehicle and information and entertainment system |
| DE102011118149A1 (en) | 2011-11-10 | 2013-05-16 | Gm Global Technology Operations, Llc | Method for operating a safety system of a motor vehicle and safety system for a motor vehicle |
| JP5499011B2 (en) | 2011-11-17 | 2014-05-21 | 富士重工業株式会社 | Outside environment recognition device and outside environment recognition method |
| US10099614B2 (en) | 2011-11-28 | 2018-10-16 | Magna Electronics Inc. | Vision system for vehicle |
| US8694224B2 (en) | 2012-03-01 | 2014-04-08 | Magna Electronics Inc. | Vehicle yaw rate correction |
| DE102013217430A1 (en) | 2012-09-04 | 2014-03-06 | Magna Electronics, Inc. | Driver assistance system for a motor vehicle |
| US9090234B2 (en) | 2012-11-19 | 2015-07-28 | Magna Electronics Inc. | Braking control system for vehicle |
| US9092986B2 (en) | 2013-02-04 | 2015-07-28 | Magna Electronics Inc. | Vehicular vision system |
| US9260095B2 (en) | 2013-06-19 | 2016-02-16 | Magna Electronics Inc. | Vehicle vision system with collision mitigation |
-
2012
- 2012-09-25 US US14/343,937 patent/US9681062B2/en not_active Expired - Fee Related
- 2012-09-25 WO PCT/US2012/057007 patent/WO2013048994A1/en active Application Filing
-
2017
- 2017-06-12 US US15/619,630 patent/US9774790B1/en active Active
- 2017-09-25 US US15/713,814 patent/US10257432B2/en active Active
Cited By (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20180048867A1 (en) * | 2011-12-06 | 2018-02-15 | Mobileye Vision Technologies Ltd. | Road vertical contour detection |
| US10084992B2 (en) * | 2011-12-06 | 2018-09-25 | Mobileye Vision Technologies Ltd. | Road vertical contour detection |
| US10506200B2 (en) | 2011-12-06 | 2019-12-10 | Mobileye Vision Technologies Ltd. | Road vertical contour detection |
| US10863140B2 (en) | 2011-12-06 | 2020-12-08 | Mobileeye Vision Technologies Ltd. | Road vertical contour detection |
| US11240471B2 (en) | 2011-12-06 | 2022-02-01 | Mobileye Vision Technologies Ltd. | Road vertical contour detection |
Also Published As
| Publication number | Publication date |
|---|---|
| US9774790B1 (en) | 2017-09-26 |
| WO2013048994A1 (en) | 2013-04-04 |
| US20140232872A1 (en) | 2014-08-21 |
| US10257432B2 (en) | 2019-04-09 |
| US20180027188A1 (en) | 2018-01-25 |
| US9681062B2 (en) | 2017-06-13 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US10257432B2 (en) | Method for enhancing vehicle camera image quality | |
| US11787338B2 (en) | Vehicular vision system | |
| US11393217B2 (en) | Vehicular vision system with detection and tracking of objects at the side of a vehicle | |
| US11563919B2 (en) | Vehicular vision system with dual processor control | |
| US11572015B2 (en) | Multi-camera vehicular vision system with graphic overlay | |
| US10104298B2 (en) | Vehicle vision system with enhanced display functions | |
| US10095935B2 (en) | Vehicle vision system with enhanced pedestrian detection | |
| US11532233B2 (en) | Vehicle vision system with cross traffic detection | |
| WO2013081984A1 (en) | Vision system for vehicle | |
| US20140350834A1 (en) | Vehicle vision system using kinematic model of vehicle motion | |
| US10682966B2 (en) | Vehicle light/display control system using camera |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
| MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
| FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY |