[go: up one dir, main page]

WO2020178035A1 - Method and system for detecting an object on a surface of a vehicle - Google Patents

Method and system for detecting an object on a surface of a vehicle Download PDF

Info

Publication number
WO2020178035A1
WO2020178035A1 PCT/EP2020/054481 EP2020054481W WO2020178035A1 WO 2020178035 A1 WO2020178035 A1 WO 2020178035A1 EP 2020054481 W EP2020054481 W EP 2020054481W WO 2020178035 A1 WO2020178035 A1 WO 2020178035A1
Authority
WO
WIPO (PCT)
Prior art keywords
vehicle
interest
processor
image
region
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Ceased
Application number
PCT/EP2020/054481
Other languages
French (fr)
Inventor
Ciaran Hughes
Jonathan Horgan
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Connaught Electronics Ltd
Original Assignee
Connaught Electronics Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Connaught Electronics Ltd filed Critical Connaught Electronics Ltd
Priority to JP2021551613A priority Critical patent/JP7236555B2/en
Priority to CN202080017919.5A priority patent/CN113508394B/en
Priority to KR1020217027803A priority patent/KR102645934B1/en
Publication of WO2020178035A1 publication Critical patent/WO2020178035A1/en
Anticipated expiration legal-status Critical
Ceased legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/56Context or environment of the image exterior to a vehicle by using sensors mounted on the vehicle
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/25Determination of region of interest [ROI] or a volume of interest [VOI]
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/82Arrangements for image or video recognition or understanding using pattern recognition or machine learning using neural networks
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/50Context or environment of the image
    • G06V20/52Surveillance or monitoring of activities, e.g. for recognising suspicious objects

Definitions

  • the present disclosure relates to the detection of an object inadvertently left on a surface of a vehicle.
  • United States patent document US6163250A discloses a system which includes a sensor which detects an object located on the outside of the vehicle.
  • the sensor can be formed of known and commercially available components such as a force or pressure gauge attached to the inside of the vehicle roof. Many types of sensors may be employed including any one or more of capacitive and/or force/pressure sensors and are known in the art.
  • Embodiments of the present invention are directed to a method and system for detecting an object on a surface of a vehicle, and as set out in the appended claims.
  • the method for detecting an object on a surface of a vehicle comprises of receiving, by a processor, one or more images from at least an image capture device. A predetermined region of interest is identified from said one or more images. The region of interest of the one or more images is traced to detect at least one object on the surface of the vehicle based on a difference in pixel intensity in the region of interest. The location of the detected object in the region of interest is mapped to a position on a representative image of the surface of the vehicle. Further, the representative image of the surface of the vehicle is displayed along with an alert to a user of the vehicle. In addition, an alert symbol is overlaid at the mapped position on the representative image of the surface of the vehicle.
  • the system for detecting an object on a surface of a vehicle comprises an image capture device, a display device and a processor operatively coupled to the image capture device and the display device.
  • the processor is configured to receive one or more images from the image capture device and identify a predetermined region of interest from said one or more images.
  • the processor is further configured to trace the region of interest and detect at least one object on the surface of the vehicle based on a difference in pixel intensity.
  • the processor is further configured to map the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle. Further the processor is configured to display the representative image of the surface of the vehicle and overlay a symbol at the position on the representative image of the surface of the vehicle.
  • the method and system of the present invention provides a user of the vehicle with an alert on the display device of the vehicle including the location of the object on the roof of the vehicle.
  • a computer program comprising program instructions for causing a computer program to carry out the above method which may be embodied on a record medium, carrier signal or read-only memory.
  • FIG. 1 A exemplarily illustrates a side view of a vehicle in accordance with some embodiments of the present invention
  • FIG. 1 B exemplarily illustrates a top view of the vehicle in accordance with some embodiments of the present invention
  • FIG. 2 is a flow chart illustrating an embodiment of the inventive method or process for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention
  • FIG. 3 is a functional block diagram illustrating the primary components of a system/apparatus for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention.
  • FIG. 4 exemplarily illustrates a display of the vehicle, in accordance with some embodiments of the present invention.
  • FIG. 1 A exemplarily illustrates a side view of a vehicle in accordance with some embodiments of the present invention which shows the vehicle 101 , an image capture device 102 arranged on the roof of the vehicle and an object 103 left over on the roof of the vehicle 101 .
  • the dotted lines in FIG. 1 A shows the field of view of the image capture device which also captures the roof of the vehicle.
  • FIG. 1 B exemplarily illustrates a top view of a vehicle in accordance with some embodiments of the present invention which shows the vehicle 101 , an image capture device 102 arranged on the roof of the vehicle and an object 103 left over on the roof of the vehicle 101 .
  • FIG. 2 is a flow chart illustrating an embodiment of the inventive method or process for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention.
  • the method for detecting an object on a surface of a vehicle comprises receiving 201 one or more images from at least an image capture device.
  • a predetermined background is subtracted 202 from a predetermined region of interest of said one or more images. It will be appreciated that other methods can be used to identify an object as distinct from the background.
  • the region of interest of the one or more images is traced 203 to detect at least one object on the surface of the vehicle based on a difference in pixel intensity in the region of interest after the background is subtracted.
  • the location of the detected object in the region of interest is mapped 204 to a position on a representative image of the surface of the vehicle.
  • the representative image of the surface of the vehicle is displayed 205 along with alert to a user of the vehicle.
  • an alert symbol is overlaid at the mapped position on the representative image of the surface of the vehicle.
  • FIG. 4 exemplarily illustrates a display 401 showing the representative image 402 of the top surface or the vehicle 101 and an alert symbol 403 is overlaid on the representative image. Further the alert symbol 403 substantially indicates the location or position of the detected object on the roof of the vehicle (see FIG. 1 B).
  • the generation of the predetermined region of interest is at least based on the position of the image capture device with respect to the surface of the vehicle. For example, the position of the image capture device with respect to the surface of the vehicle determines the region of interest. In other words, the region of interest is a portion of an image which captures the surface of the vehicle.
  • generating the predetermined background is based on at least one image of the surface of the vehicle.
  • an image of the surface of the vehicle may be captured using the image capture device while there is no object placed on said surface.
  • the image capture device may capture a plurality of images at different times when the vehicle is in motion. The portion of the image which remains substantially static and does not change across images is considered the background of the image.
  • the contour of the static portion of said images may define the region of interest as mentioned above.
  • the method further comprises extracting the image of the detected object and classifying the object using a machine learning algorithm or a neural network based model and providing an audio alert to a user of the vehicle based on the classification of the object. For example, if an apple is left on the surface of the vehicle, an alert may be sounded“Alert! An apple is on the roof of the vehicle”.
  • the method further comprises sending a signal to the electronic control unit (ECU) of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle. For example, if an object is detected on the surface of the vehicle, then the ECU is signalled to limit the speed of the vehicle or prevent the vehicle from moving, such that the object inadvertently left on the vehicle does not fall off.
  • ECU electronice control unit
  • the method while tracing the region of interest ignores objects attached to the vehicle, where the objects are a part of the vehicle.
  • objects attached to the vehicle where the objects are a part of the vehicle.
  • a luggage holder mounted on the roof of the vehicle.
  • FIG. 3 is a functional block diagram illustrating the primary components of a system/apparatus 300 for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention.
  • the system comprises a processor 301 , a memory 302, a transceiver 303, an image capture device 304 and a display 305.
  • the processor 301 is operably coupled to the memory 302, the transceiver 303, the image capture device 304 and the display 305.
  • an electronic control unit (ECU) 31 1 is in communication with the processor 301 via said transceiver 303.
  • a speed governor 312 of the vehicle is operably coupled to the ECU 31 1.
  • the image capture device 304 is a fish eye camera.
  • the processor 301 is configured to receive one or more images from the image capture device 304. Further, the processor 301 subtracts a predetermined background from a predetermined region of interest from said one or more images. The processor 301 is further configured to trace the region of interest and detect at least one object on the surface of the vehicle based on a difference in pixel intensity within the region of interest. The processor 301 is further configured to map the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle. Further, the processor 301 causes the display 305 to display said representative image of the surface of the vehicle and overlay a symbol at the position on the representative image of the surface of the vehicle. For example, FIG.
  • the generation of the predetermined region of interest by the processor 301 is at least based on the position of the image capture device 304 with respect to the surface of the vehicle. For example, the position of the image capture device 304 with respect to the surface of the vehicle determines the region of interest. In other words, the region of interest is a portion of an image which captures the surface of the vehicle.
  • the processor is configured to generate the predetermined background, which is based on at least one image of the surface of the vehicle. For example, an image of the surface of the vehicle may be captured using the image capture device 304 while there is no object placed on said surface and said image stored in the memory 302 as a background image.
  • the processor 301 may cause the image capture device 304 to capture a plurality of images at different times when the vehicle is in motion.
  • the processor may determine a portion of the image which remains substantially static/unchanged through said captured plurality of images, where said determined portion of image is stored in the memory 302 as the background image.
  • the contour of the static portion of said images may define the region of interest as mentioned above and the configuration of said contour stored in memory 302 as the region of interest.
  • the processor 301 is further configured to extract the image of detected object and classify the object using a machine learning algorithm or a neural network based model.
  • the system further comprising an audio output device operably coupled to the processor 301 and the processor 301 configured to provide an audio alert to a user of the vehicle based on the classification of the object. For example, if an apple is left on the surface of the vehicle, an alert may be sounded “Alert! An apple is on the roof of the vehicle”.
  • the processor 301 is further configured to send a signal to the electronic control unit (ECU) 31 1 of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle. For example, if an object is detected on the surface of the vehicle, then the ECU is signalled to limit the speed of the vehicle or prevent the vehicle from moving using the speed governor 312, such that the object inadvertently left on the vehicle does not fall off.
  • ECU electronice control unit
  • the process described in the present disclosure may be implemented using various means.
  • the process described in the present disclosure may be implemented in hardware, firmware, software, or any combination thereof.
  • the processing units, or processors(s) may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, electronic devices, other electronic units designed to perform the functions described herein, or a combination thereof.
  • ASICs application specific integrated circuits
  • DSPs digital signal processors
  • DSPDs digital signal processing devices
  • PLDs programmable logic devices
  • FPGAs field programmable gate arrays
  • processors controllers, micro-controllers, microprocessors, electronic devices, other electronic units designed to perform the functions described herein, or a combination thereof.
  • firmware and/or software implementation software codes may be stored in a memory and executed
  • the embodiments in the invention described with reference to the drawings comprise a computer apparatus and/or processes performed in a computer apparatus.
  • the invention also extends to computer programs, particularly computer programs stored on or in a carrier adapted to bring the invention into practice.
  • the program may be in the form of source code, object code, or a code intermediate source and object code, such as in partially compiled form or in any other form suitable for use in the implementation of the method according to the invention.
  • the carrier may comprise a storage medium such as ROM, e.g. a memory stick or hard disk.
  • the carrier may be an electrical or optical signal which may be transmited via an electrical or an optical cable or by radio or other means.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Auxiliary Drives, Propulsion Controls, And Safety Devices (AREA)
  • Image Processing (AREA)

Abstract

The present invention provides a method and system for detecting an object on a surface of a vehicle. A processor receives one or more images from an image capture device. A predetermined background is subtracted from a predetermined region of interest of said one or more images. The region of interest of the one or more images is traced to detect at least one object on the surface of the vehicle based on a difference in pixel intensity in the region of interest. The location of the detected object in the region of interest is mapped to a position on a representative image of the surface of the vehicle. Further, the representative image of the surface of the vehicle is displayed along with an alert to the user of the vehicle. Further, an alert symbol is overlaid at the mapped position on the representative image of the surface of the vehicle.

Description

Title
Method and system for detecting an object on a surface of a vehicle.
Field
The present disclosure relates to the detection of an object inadvertently left on a surface of a vehicle.
Background
It is very common for a user of a vehicle to inadvertently leave personal objects (e.g. purse, wallet, phone, laptop, etc.) on the top of a vehicle. If the user forgets that the personal object is on the top of the vehicle, the user may enter the vehicle and drive away. This may cause the personal object to fall and may get damaged. Therefore, in such a situation, before the user drives away it is desirable have a system that locates the personal object and warns the user that the personal item has been left on the top of the vehicle.
United States patent document US6163250A discloses a system which includes a sensor which detects an object located on the outside of the vehicle. The sensor can be formed of known and commercially available components such as a force or pressure gauge attached to the inside of the vehicle roof. Many types of sensors may be employed including any one or more of capacitive and/or force/pressure sensors and are known in the art.
However, the prior art fails to detect and display the location of the personal object on the roof. Therefore, there is a need for system that displays the location of the personal item on the top of the vehicle along with an alert to the user of the vehicle.
Summary
Embodiments of the present invention are directed to a method and system for detecting an object on a surface of a vehicle, and as set out in the appended claims. The method for detecting an object on a surface of a vehicle comprises of receiving, by a processor, one or more images from at least an image capture device. A predetermined region of interest is identified from said one or more images. The region of interest of the one or more images is traced to detect at least one object on the surface of the vehicle based on a difference in pixel intensity in the region of interest. The location of the detected object in the region of interest is mapped to a position on a representative image of the surface of the vehicle. Further, the representative image of the surface of the vehicle is displayed along with an alert to a user of the vehicle. In addition, an alert symbol is overlaid at the mapped position on the representative image of the surface of the vehicle.
The system for detecting an object on a surface of a vehicle comprises an image capture device, a display device and a processor operatively coupled to the image capture device and the display device. The processor is configured to receive one or more images from the image capture device and identify a predetermined region of interest from said one or more images. The processor is further configured to trace the region of interest and detect at least one object on the surface of the vehicle based on a difference in pixel intensity. The processor is further configured to map the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle. Further the processor is configured to display the representative image of the surface of the vehicle and overlay a symbol at the position on the representative image of the surface of the vehicle.
Thus the method and system of the present invention provides a user of the vehicle with an alert on the display device of the vehicle including the location of the object on the roof of the vehicle. There is also provided a computer program comprising program instructions for causing a computer program to carry out the above method which may be embodied on a record medium, carrier signal or read-only memory. Brief Description of the Drawings
The invention will be more clearly understood from the following description of an embodiment thereof, given by way of example only, with reference to the accompanying drawings, in which: -
FIG. 1 A exemplarily illustrates a side view of a vehicle in accordance with some embodiments of the present invention;
FIG. 1 B exemplarily illustrates a top view of the vehicle in accordance with some embodiments of the present invention
FIG. 2 is a flow chart illustrating an embodiment of the inventive method or process for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention;
FIG. 3 is a functional block diagram illustrating the primary components of a system/apparatus for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention; and
FIG. 4 exemplarily illustrates a display of the vehicle, in accordance with some embodiments of the present invention.
Detailed Description of the Drawings
FIG. 1 A exemplarily illustrates a side view of a vehicle in accordance with some embodiments of the present invention which shows the vehicle 101 , an image capture device 102 arranged on the roof of the vehicle and an object 103 left over on the roof of the vehicle 101 . The dotted lines in FIG. 1 A shows the field of view of the image capture device which also captures the roof of the vehicle.
FIG. 1 B exemplarily illustrates a top view of a vehicle in accordance with some embodiments of the present invention which shows the vehicle 101 , an image capture device 102 arranged on the roof of the vehicle and an object 103 left over on the roof of the vehicle 101 . FIG. 2 is a flow chart illustrating an embodiment of the inventive method or process for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention. The method for detecting an object on a surface of a vehicle comprises receiving 201 one or more images from at least an image capture device. A predetermined background is subtracted 202 from a predetermined region of interest of said one or more images. It will be appreciated that other methods can be used to identify an object as distinct from the background. The region of interest of the one or more images is traced 203 to detect at least one object on the surface of the vehicle based on a difference in pixel intensity in the region of interest after the background is subtracted. The location of the detected object in the region of interest is mapped 204 to a position on a representative image of the surface of the vehicle. Further, the representative image of the surface of the vehicle is displayed 205 along with alert to a user of the vehicle. Further, an alert symbol is overlaid at the mapped position on the representative image of the surface of the vehicle. For example, FIG. 4 exemplarily illustrates a display 401 showing the representative image 402 of the top surface or the vehicle 101 and an alert symbol 403 is overlaid on the representative image. Further the alert symbol 403 substantially indicates the location or position of the detected object on the roof of the vehicle (see FIG. 1 B).
The generation of the predetermined region of interest is at least based on the position of the image capture device with respect to the surface of the vehicle. For example, the position of the image capture device with respect to the surface of the vehicle determines the region of interest. In other words, the region of interest is a portion of an image which captures the surface of the vehicle.
Further as mentioned above, generating the predetermined background is based on at least one image of the surface of the vehicle. For example, an image of the surface of the vehicle may be captured using the image capture device while there is no object placed on said surface. In another embodiment, the image capture device may capture a plurality of images at different times when the vehicle is in motion. The portion of the image which remains substantially static and does not change across images is considered the background of the image. Also, the contour of the static portion of said images may define the region of interest as mentioned above. The method further comprises extracting the image of the detected object and classifying the object using a machine learning algorithm or a neural network based model and providing an audio alert to a user of the vehicle based on the classification of the object. For example, if an apple is left on the surface of the vehicle, an alert may be sounded“Alert! An apple is on the roof of the vehicle”.
The method further comprises sending a signal to the electronic control unit (ECU) of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle. For example, if an object is detected on the surface of the vehicle, then the ECU is signalled to limit the speed of the vehicle or prevent the vehicle from moving, such that the object inadvertently left on the vehicle does not fall off.
The method while tracing the region of interest ignores objects attached to the vehicle, where the objects are a part of the vehicle. For example, a luggage holder mounted on the roof of the vehicle.
The above method could be applied to any surface of the vehicle e.g. a roof of the vehicle, a bonnet of the vehicle, or a car boot cover of the vehicle. FIG. 3 is a functional block diagram illustrating the primary components of a system/apparatus 300 for detecting an object on a surface of a vehicle, in accordance with some embodiments of the present invention. The system comprises a processor 301 , a memory 302, a transceiver 303, an image capture device 304 and a display 305. The processor 301 is operably coupled to the memory 302, the transceiver 303, the image capture device 304 and the display 305. Further, an electronic control unit (ECU) 31 1 is in communication with the processor 301 via said transceiver 303. Also, a speed governor 312 of the vehicle is operably coupled to the ECU 31 1. In a preferable embodiment the image capture device 304 is a fish eye camera.
The processor 301 is configured to receive one or more images from the image capture device 304. Further, the processor 301 subtracts a predetermined background from a predetermined region of interest from said one or more images. The processor 301 is further configured to trace the region of interest and detect at least one object on the surface of the vehicle based on a difference in pixel intensity within the region of interest. The processor 301 is further configured to map the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle. Further, the processor 301 causes the display 305 to display said representative image of the surface of the vehicle and overlay a symbol at the position on the representative image of the surface of the vehicle. For example, FIG. 4 exemplarily illustrates a display 401 showing the representative image 402 of the top surface or the vehicle 101 and an alert symbol 403 is overlaid on the representative image. Further the alert symbol 403 substantially indicates the location or position of the detected object on the roof of the vehicle (see FIG. 1 B). The generation of the predetermined region of interest by the processor 301 is at least based on the position of the image capture device 304 with respect to the surface of the vehicle. For example, the position of the image capture device 304 with respect to the surface of the vehicle determines the region of interest. In other words, the region of interest is a portion of an image which captures the surface of the vehicle.
Further the processor is configured to generate the predetermined background, which is based on at least one image of the surface of the vehicle. For example, an image of the surface of the vehicle may be captured using the image capture device 304 while there is no object placed on said surface and said image stored in the memory 302 as a background image. In another embodiment, the processor 301 may cause the image capture device 304 to capture a plurality of images at different times when the vehicle is in motion. Further, the processor may determine a portion of the image which remains substantially static/unchanged through said captured plurality of images, where said determined portion of image is stored in the memory 302 as the background image. Also, the contour of the static portion of said images may define the region of interest as mentioned above and the configuration of said contour stored in memory 302 as the region of interest.
The processor 301 is further configured to extract the image of detected object and classify the object using a machine learning algorithm or a neural network based model. The system further comprising an audio output device operably coupled to the processor 301 and the processor 301 configured to provide an audio alert to a user of the vehicle based on the classification of the object. For example, if an apple is left on the surface of the vehicle, an alert may be sounded “Alert! An apple is on the roof of the vehicle”.
The processor 301 is further configured to send a signal to the electronic control unit (ECU) 31 1 of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle. For example, if an object is detected on the surface of the vehicle, then the ECU is signalled to limit the speed of the vehicle or prevent the vehicle from moving using the speed governor 312, such that the object inadvertently left on the vehicle does not fall off.
Further, a person ordinarily skilled in the art will appreciate that the various illustrative logical/functional blocks, modules, circuits, techniques/algorithms and process steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or a combination of hardware and software. To clearly illustrate this interchangeability of hardware and a combination of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or a combination of hardware and software depends upon the design choice of a person ordinarily skilled in the art. Such skilled artisans may implement the described functionality in varying ways for each particular application, but such obvious design choices should not be interpreted as causing a departure from the scope of the present invention.
The process described in the present disclosure may be implemented using various means. For example, the process described in the present disclosure may be implemented in hardware, firmware, software, or any combination thereof. For a hardware implementation, the processing units, or processors(s) may be implemented within one or more application specific integrated circuits (ASICs), digital signal processors (DSPs), digital signal processing devices (DSPDs), programmable logic devices (PLDs), field programmable gate arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, electronic devices, other electronic units designed to perform the functions described herein, or a combination thereof. For a firmware and/or software implementation, software codes may be stored in a memory and executed by a processor. Memory may be implemented within the processor unit or external to the processor unit. As used herein the term“memory” refers to any type of volatile memory or nonvolatile memory. Furthermore, although elements of the invention may be described or claimed in the singular, the plural is contemplated unless limitation to the singular is explicitly stated.
The embodiments in the invention described with reference to the drawings comprise a computer apparatus and/or processes performed in a computer apparatus. However, the invention also extends to computer programs, particularly computer programs stored on or in a carrier adapted to bring the invention into practice. The program may be in the form of source code, object code, or a code intermediate source and object code, such as in partially compiled form or in any other form suitable for use in the implementation of the method according to the invention. The carrier may comprise a storage medium such as ROM, e.g. a memory stick or hard disk. The carrier may be an electrical or optical signal which may be transmited via an electrical or an optical cable or by radio or other means.
In the specification the terms "comprise, comprises, comprised and comprising" or any variation thereof and the terms include, includes, included and including" or any variation thereof are considered to be totally interchangeable and they should all be afforded the widest possible interpretation and vice versa.
The invention is not limited to the embodiments hereinbefore described but may be varied in both construction and detail.

Claims

Claims
1. A method for detecting an object on a surface of a vehicle comprising:
receiving by a processor one or more images from at least an image capture device;
identifying a predetermined region of interest from said one or more images;
tracing the region of interest and detecting at least one object on the surface of the vehicle based on a difference in pixel intensity;
mapping the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle; and
displaying said representative image of the surface of the vehicle and overlaying a symbol at the position on the representative image of the surface of the vehicle.
2. The method of claim 1 , further comprising generating the predetermined region of interest at least based on the position of the image capture device with respect to the surface of the vehicle.
3. The method of claim 1 , further comprising subtracting a predetermined background from the predetermined region of interest from said one or more images
4. The method of claim 3, further comprising generating the predetermined background based on at least one image of the surface of the vehicle.
5. The method of claim 1 , further comprising extracting the image of detected object and classifying the object using a machine learning algorithm or a neural network based model.
6. The method of claim 5, further comprising providing an audio alert to a user of the vehicle based on the classification of the object.
7. The method of claim 1 further comprising sending a signal to the electronic control unit of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle.
8. The method of claim 1 wherein tracing the region of interest comprises ignoring objects attached to the vehicle, wherein said objects attached to the vehicle being a part of the vehicle.
9. The method claimed in any one of the preceding claims, wherein the surface of the vehicle is one or more of a roof of the vehicle, a bonnet of the vehicle, or a car boot cover of the vehicle.
10. A system for detecting an object on a surface of a vehicle comprising:
at least one image capture device;
a display device;
a processor operatively coupled to the at least one image capture device and the display device, wherein said processor configured to:
receive one or more images from the at least one image capture device;
identify a predetermined region of interest from said one or more images;
trace the region of interest and detecting at least one object on the surface of the vehicle based on a difference in pixel intensity;
map the position of the detected object detected in the region of interest to a position on a representative image of the surface of the vehicle; and
display using the display said representative image of the surface of the vehicle and overlaying a symbol at the position on the representative image of the surface of the vehicle.
1 1. The system of claim 10, wherein the processor is further configured to generate the predetermined region of interest at least based on the position of the image capture device with respect to the surface of the vehicle.
12. The system of claim 10, wherein the processor is further configured to subtract a predetermined background from the predetermined region of interest from said one or more images
13. The system of claim 12, wherein the processor is further configured to generate the predetermined background based on at least one image of the surface of the vehicle while there is no object or objects placed on the surface of the vehicle.
14. The system of claim 10, wherein the processor is further configured to extract the image of detected object and classifying the object using a machine learning algorithm or a neural network based model.
15. The system of claim 14, further comprising an audio output device operably coupled to the processor, wherein said processor provides an audio alert to a user of the vehicle based on the classification of the object using the audio output device.
16. The system of claim 10, further comprising an electronic control unit of the vehicle wherein the processor is configured to send a signal to the electronic control unit of the vehicle for preventing motion of the vehicle or limiting the speed of the vehicle.
17. The system of claim 10, wherein the image capture device is a fish eye camera.
18. A computer program comprising instructions which, when the program is executed by a data processing device/computer, cause the data processing device/computer to carry out the steps of the method of any of claims 1 to 9.
19. A computer readable medium comprising instructions which, when executed by a computer, cause the computer to carry out the method of any of claims 1 to 9.
PCT/EP2020/054481 2019-03-01 2020-02-20 Method and system for detecting an object on a surface of a vehicle Ceased WO2020178035A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
JP2021551613A JP7236555B2 (en) 2019-03-01 2020-02-20 Method and system for detecting objects on vehicle surfaces
CN202080017919.5A CN113508394B (en) 2019-03-01 2020-02-20 Method and system for detecting objects on a vehicle surface
KR1020217027803A KR102645934B1 (en) 2019-03-01 2020-02-20 Method and system for detecting objects on the surface of a vehicle

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
DE102019105275.0 2019-03-01
DE102019105275.0A DE102019105275A1 (en) 2019-03-01 2019-03-01 Method and system for detecting an object on a surface of a vehicle

Publications (1)

Publication Number Publication Date
WO2020178035A1 true WO2020178035A1 (en) 2020-09-10

Family

ID=69844773

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2020/054481 Ceased WO2020178035A1 (en) 2019-03-01 2020-02-20 Method and system for detecting an object on a surface of a vehicle

Country Status (5)

Country Link
JP (1) JP7236555B2 (en)
KR (1) KR102645934B1 (en)
CN (1) CN113508394B (en)
DE (1) DE102019105275A1 (en)
WO (1) WO2020178035A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12479362B2 (en) 2023-12-15 2025-11-25 Continental Autonomous Mobility US, LLC Method for alerting a driver to objects left on a vehicle roof and alert system

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6163250A (en) 1999-08-31 2000-12-19 International Business Machines Corporation System and method for sensing objects on surface of vehicle
US20080247599A1 (en) * 2007-04-05 2008-10-09 Porikli Fatih M Method for Detecting Objects Left-Behind in a Scene
US20170364072A1 (en) * 2016-06-15 2017-12-21 Ford Global Technologies, Llc Vehicle exterior surface object detection

Family Cites Families (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS6454196U (en) * 1987-09-30 1989-04-04
JP3889908B2 (en) * 1999-10-25 2007-03-07 矢崎総業株式会社 Warning display device for vehicles
DE102004043756B4 (en) * 2004-09-10 2019-05-29 Volkswagen Ag Arrangement for detecting objects in a vehicle
DE102006003193A1 (en) * 2006-01-24 2007-07-26 Daimlerchrysler Ag Instrument cluster testing method, involves filing representation of camera to messages, so that camera detected image is indicated to each data bus signal to display actual state of cluster in the event of error
US9204047B2 (en) * 2011-04-08 2015-12-01 Nokia Technologies Oy Imaging
US9530310B2 (en) * 2013-11-01 2016-12-27 Xerox Corporation Method and system for detecting and tracking a vehicle of interest utilizing a network of traffic image-capturing units
DE102013226104A1 (en) * 2013-12-16 2015-06-18 Application Solutions (Electronics and Vision) Ltd. METHOD AND DEVICE FOR MONITORING AN EXTERNAL DIMENSION OF A VEHICLE
EP3089449B1 (en) * 2015-04-30 2020-03-04 InterDigital CE Patent Holdings Method for obtaining light-field data using a non-light-field imaging device, corresponding device, computer program product and non-transitory computer-readable carrier medium
ITUB20151802A1 (en) * 2015-07-01 2017-01-01 Magneti Marelli Spa On-board vehicle system and improved method for detecting objects in a vehicle's surroundings.
US10150448B2 (en) * 2015-09-18 2018-12-11 Ford Global Technologies. Llc Autonomous vehicle unauthorized passenger or object detection
DE102016215524B4 (en) * 2016-08-18 2019-03-28 Volkswagen Aktiengesellschaft System for object recognition and object recall in a vehicle
DE102017205093A1 (en) * 2017-03-27 2018-09-27 Conti Temic Microelectronic Gmbh Method and system for predicting sensor signals of a vehicle
KR102400017B1 (en) * 2017-05-17 2022-05-19 삼성전자주식회사 Method and device for identifying an object
GB2566672B (en) * 2017-07-19 2020-05-27 Jaguar Land Rover Ltd Method and apparatus for outputting a control signal
DE102018211299A1 (en) * 2018-07-09 2020-01-09 Robert Bosch Gmbh Method for checking an area on a vehicle for foreign objects
JP2020052639A (en) * 2018-09-26 2020-04-02 三菱電機株式会社 Vehicle monitoring device and vehicle monitoring system

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6163250A (en) 1999-08-31 2000-12-19 International Business Machines Corporation System and method for sensing objects on surface of vehicle
US20080247599A1 (en) * 2007-04-05 2008-10-09 Porikli Fatih M Method for Detecting Objects Left-Behind in a Scene
US20170364072A1 (en) * 2016-06-15 2017-12-21 Ford Global Technologies, Llc Vehicle exterior surface object detection

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US12479362B2 (en) 2023-12-15 2025-11-25 Continental Autonomous Mobility US, LLC Method for alerting a driver to objects left on a vehicle roof and alert system

Also Published As

Publication number Publication date
CN113508394A (en) 2021-10-15
KR20210115045A (en) 2021-09-24
JP2022522765A (en) 2022-04-20
CN113508394B (en) 2025-01-28
JP7236555B2 (en) 2023-03-09
DE102019105275A1 (en) 2020-09-03
KR102645934B1 (en) 2024-03-12

Similar Documents

Publication Publication Date Title
JP7238217B2 (en) A system for identifying defined objects
CN113379734B (en) Quality detection method, quality detection device, quality detection equipment and computer-readable storage medium
CN109670383B (en) Video shielding area selection method and device, electronic equipment and system
JP6764481B2 (en) Monitoring device
CN113034354B (en) Image processing method and device, electronic equipment and readable storage medium
US20150169980A1 (en) Object recognition device
CN106803083B (en) Pedestrian detection method and device
KR20180011794A (en) Object visualization method
KR102071410B1 (en) Smart mirror
CN113658141A (en) Transparent packaging bag sealing identification method and device, storage medium and electronic equipment
US20140375821A1 (en) Detection system
CN107341379A (en) Screen unlocking method and device, electronic device and computer-readable storage medium
KR101243294B1 (en) Method and apparatus for extracting and tracking moving objects
KR101236234B1 (en) Detection system of road line using both laser sensor and camera
WO2020178035A1 (en) Method and system for detecting an object on a surface of a vehicle
KR20150107581A (en) Image processing apparatus and image processing method
US11961284B2 (en) Task appropriateness determination apparatus
KR20220152590A (en) Apparatus for preventing accidents caught in vehicle door and method thereof
JPH07101265A (en) Operating state detector
KR102820740B1 (en) Apparatus for Tracking Detection Object and Driving Method Thereof
Radojcic et al. TRANSFORMING TRAFFIC SAFETY: DETECTION OF CAR-PEDESTRIAN CONTACT USING COMPUTER VISION TECHNOLOGIES.
KR101300279B1 (en) Detection of vehicles in different illumination environments
US20220215692A1 (en) Operation-correctness determining apparatus and operation-correctness determining method
Sharran et al. Drowsy Driver Detection System
CN120047886A (en) Dangerous state monitoring method, system and storage medium

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20711504

Country of ref document: EP

Kind code of ref document: A1

ENP Entry into the national phase

Ref document number: 2021551613

Country of ref document: JP

Kind code of ref document: A

ENP Entry into the national phase

Ref document number: 20217027803

Country of ref document: KR

Kind code of ref document: A

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20711504

Country of ref document: EP

Kind code of ref document: A1

WWG Wipo information: grant in national office

Ref document number: 202080017919.5

Country of ref document: CN