[go: up one dir, main page]

US20220092829A1 - Augmented reality detection for locating autonomous vehicles - Google Patents

Augmented reality detection for locating autonomous vehicles Download PDF

Info

Publication number
US20220092829A1
US20220092829A1 US17/540,586 US202117540586A US2022092829A1 US 20220092829 A1 US20220092829 A1 US 20220092829A1 US 202117540586 A US202117540586 A US 202117540586A US 2022092829 A1 US2022092829 A1 US 2022092829A1
Authority
US
United States
Prior art keywords
electronic device
autonomous vehicle
identification signal
indication
signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US17/540,586
Inventor
Matthew Hortop
Patrick Hunt
Lacey Trelfa
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Rivian IP Holdings LLC
Rivian Automotive LLC
Original Assignee
Rivian IP Holdings LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Rivian IP Holdings LLC filed Critical Rivian IP Holdings LLC
Priority to US17/540,586 priority Critical patent/US20220092829A1/en
Assigned to RIVIAN IP HOLDINGS, LLC reassignment RIVIAN IP HOLDINGS, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Rivian Automotive, LLC
Assigned to Rivian Automotive, LLC reassignment Rivian Automotive, LLC ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: Hortop, Matthew, HUNT, PATRICK, TRELFA, LACEY
Publication of US20220092829A1 publication Critical patent/US20220092829A1/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01DMEASURING NOT SPECIALLY ADAPTED FOR A SPECIFIC VARIABLE; ARRANGEMENTS FOR MEASURING TWO OR MORE VARIABLES NOT COVERED IN A SINGLE OTHER SUBCLASS; TARIFF METERING APPARATUS; MEASURING OR TESTING NOT OTHERWISE PROVIDED FOR
    • G01D21/00Measuring or testing not otherwise provided for
    • G01D21/02Measuring two or more variables by means not covered by a single other subclass
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0487Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
    • G06F3/0488Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
    • G06K9/00671
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/20Scenes; Scene-specific elements in augmented reality scenes
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/12Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/63Control of cameras or camera modules by using electronic viewfinders
    • H04N23/631Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N23/00Cameras or camera modules comprising electronic image sensors; Control thereof
    • H04N23/60Control of cameras or camera modules
    • H04N23/64Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
    • H04N5/23222
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04WWIRELESS COMMUNICATION NETWORKS
    • H04W4/00Services specially adapted for wireless communication networks; Facilities therefor
    • H04W4/30Services specially adapted for particular environments, situations or purposes
    • H04W4/40Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P]
    • H04W4/48Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P] for in-vehicle communication
    • G06K2209/23
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/08Detecting or categorising vehicles

Definitions

  • the present disclosure relates to autonomous vehicles.
  • Autonomous vehicles are of great interest for transportation applications and can provide benefits to the ride-sharing industry.
  • the intended passenger may have to contact the driver or have human interactions with the driver of the vehicle to identify his/her particular vehicle, especially when in a crowded area such as a city or an airport.
  • Such interaction may be challenging, if even possible, when attempting to locate an autonomous vehicle within a pool of autonomous vehicles or a crowded area with obstructions.
  • Disclosed embodiments include methods, systems, and non-transitory computer readable medium for augmented reality detection for locating a particular autonomous vehicle.
  • an illustrative method includes providing, by an electronic device to a remote server, a request for an autonomous vehicle.
  • a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger is received by an electronic device.
  • the passenger is prompted by the electronic device to orient an image sensor of the electronic device towards a designated area of the accessible range.
  • the designated area is imaged by the image sensor for the autonomous vehicle.
  • a visual representation of the imaged designated area is generated on a graphical user interface of the electronic device.
  • the autonomous vehicle is identified within the generated visual representation.
  • an illustrative system includes a computer processing system and a computer memory.
  • the computer processing system is configured to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
  • non-transitory computer readable medium includes program instructions for locating autonomous vehicles.
  • the program instructions when executed cause a computer processing system to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
  • FIG. 1 is a block diagram in partial schematic form of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIG. 2A is a block diagram in partial schematic form of a partial top view of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIG. 2B is a graph of signal strength of detected location signals as a function of image sensor direction of an electronic device for locating a particular autonomous vehicle.
  • FIG. 3 is a block diagram in partial schematic form of another illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIGS. 4A-4D are top views of illustrative identifications of an autonomous vehicle.
  • FIG. 5 is a flow chart of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle.
  • first digit of three-digit reference numbers and the first two digits of four-digit reference numbers correspond to the first digit of one-digit figure numbers and the first two digits of two-digit figure numbers, respectively, in which the element first appears.
  • FIG. 1 illustrates an illustrative framework 100 for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure.
  • a passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle, e.g., from a pool of autonomous vehicles and then use the application on the electronic device 102 to select the proper intended autonomous vehicle from a group of nearby vehicles 104 .
  • An autonomous vehicle can be configured with an array of sensors 105 , including, e.g., LIDAR (light detection and ranging), camera systems for acquiring real-time video imagery of obstacles and other vehicles, and GPS (global positioning system).
  • LIDAR light detection and ranging
  • camera systems for acquiring real-time video imagery of obstacles and other vehicles
  • GPS global positioning system
  • An autonomous vehicle can also be configured with wireless communication systems for sending and receiving communications regarding road information and traffic in real-time, as well as a computer for applying driving rules and making driving decisions based on data and information acquired from the foregoing, in a suitable manner such as conventionally known in the art.
  • Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102 .
  • Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet.
  • Remote monitoring backend system(s) 108 may select a particular autonomous vehicle, e.g., from a pool of autonomous vehicles based on one or more criteria which may include, but are not limited to, distance to the intended passenger's pick-up location, amount of time it would take to arrive at the intended passenger's pick-up location, destination of the intended passenger, and/or desired vehicle accommodations of the intended passenger.
  • Remote monitoring backend system(s) 108 may provide the request to the selected autonomous vehicle via wireless and/or wired communication network 110 .
  • Remote monitoring backend system(s) 108 may also communicate with on-board computers of the pool of autonomous vehicles 104 to control various aspects of vehicle features, as will be described further herein.
  • Remote monitoring backend system(s) 108 may include a computer processing system 108 , which may include one or more computer processing units (CPUs) 112 located in one or more computer systems, and one or more memories 114 , which may include RAM, ROM, and any suitable nonvolatile memory. Remote monitoring backend system(s) 108 may also include one or more databases 116 to store data for both vehicles and drivers, as will be explained further herein, and one or more interfaces 118 to facilitate communication via networks including the Internet and for facilitating input/output of data. Remote monitoring backend system(s) 108 carries out monitoring of vehicles, vehicle systems, vehicle usage, and systems usage for individual vehicles and for individual drivers who use those vehicles, and carries out data analytics on the data gathered from the vehicles and the vehicle systems.
  • CPUs computer processing units
  • memories 114 which may include RAM, ROM, and any suitable nonvolatile memory.
  • Remote monitoring backend system(s) 108 may also include one or more databases 116 to store data for both vehicles and drivers, as will be explained
  • the pool of autonomous vehicles may include various detectors and sensors for monitoring vehicle dynamics, usage, and demand placed on vehicle systems during the driving of the pool of autonomous vehicles.
  • the autonomous vehicles may include GPS units for monitoring location, speed, and direction, battery monitoring sensors in the form of electrical circuits, for monitoring battery usage, battery discharge as a function of time, and battery charging rates, and other suitable sensors for monitoring the health and status of other major subsystems of the vehicle, such as removable modules, and for detecting warning or fault conditions.
  • Such data may be stored in an onboard computer system with suitable memory and may be uploaded to the remote monitoring backend system(s) 108 through wireless and/or wired communication network 110 via a vehicle onboard wireless transceiver and/or via wired communication.
  • Each autonomous vehicle of the pool of autonomous vehicles can emit one or more location signals.
  • the one or more location signals of each autonomous vehicle may include rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection.
  • the pulses while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger.
  • the one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle.
  • the one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths.
  • Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102 , e.g., via text messaging or other suitable messaging.
  • Electronic device 102 can detect the one or more location signals of the autonomous vehicle once it is in an accessible range of the autonomous vehicle.
  • the intended passenger can be prompted, via a visual instruction displayed at graphical user interface 106 or via an audio instruction, or both, to orient an image sensor (not shown in FIG. 1 ) of the electronic device 102 toward a designated area once the autonomous vehicle is within an accessible range. After the passenger orients the image sensor to the designated area, the image sensor of electronic device 102 may capture a scene of the pool of autonomous vehicles 104 , as will be described further herein.
  • a visual representation of the imaged area may be generated by electronic device 102 using one or more internal processing components.
  • the visual representation of the imaged area may be displayed on graphical user interface 106 with the autonomous vehicle identified to the passenger via one or more indications 120 , as will be described further herein.
  • the autonomous vehicle may be partially obstructed by one or more obstructions 122 and can still be identified to the intended passenger.
  • Access to the autonomous vehicle may be provided to the intended passenger via an access button (field) 122 on graphical user interface 106 , e.g., an unlock graphic field on the touch-sensitive display screen displaying graphical user interface 106 that the intended passenger may press.
  • FIG. 2A illustrates a block diagram of a partial top view 200 of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure.
  • Electronic device 102 may include an image sensor 202 .
  • the image sensor 202 may be sensitive to visible light, non-visible light (e.g., infrared light), or both.
  • the image sensor 202 may be a conventional image sensor of the type typically found in smart phones.
  • a user having already communicated a request for a vehicle using the app running on the electronic device 102 , may be alerted by the electronic device 102 that an intended vehicle is with detectable range of the user and may direct the use via a visual instruction, an audio instruction, or both, to point the electronic device toward a field of view encompassing one or more vehicles and to scan the image sensor of the electronic device 102 across the field of view by tilting the electronic device 102 laterally as illustrated in FIG. 2A .
  • Electronic device 102 may be tilted in one or more axes of a three-dimensional system, e.g., x,y,z-system as illustrated in FIG. 2A . Tilting of electronic device 102 allows image sensor 202 to capture images of the designated area with different orientations whereby different vehicles in the field of view are aligned with different lines of sight (e.g., dotted lines 1 , 2 and 3 ). For example, electronic device 102 may be tilted slightly back and forth to scan and capture the signal strength of the one or more location signals.
  • the one or more location signals can be recorded as a function of orientation of the electronic device, e.g., the vector direction of the image sensor 202 of the electronic device 102 and the orientation with the highest line of sight signal strength may be the one that points to the intended autonomous vehicle. That information can then be correlated to the captured image associated with the various orientations to flag the intended autonomous vehicle using an indicator 120 , as will be described further herein.
  • Processing of the image information by electronic device 102 may determine an orientation of electronic device 102 which provides a strongest line of sight signal to capture an image of the pool of autonomous vehicles.
  • electronic device 102 may have an internal inertial sensor that provides data to a processor of the electronic device 102 .
  • the electronic device 102 may prompt the user to slowly scan the image sensor 202 across a field of view containing several vehicles, may evaluate the signal strength of detected location signals as a function of orientation based on correlated inertial data simultaneously captured by the electronic device, and then identify which location in the captured imagery corresponds to the greatest signal strength of the detected location signals, and then display on the electronic device that location superimposed with the imagery of the scene so as to identify to the user the intended autonomous vehicle.
  • the electronic device 102 may prompt a passenger, via graphical user interface 106 , to modify the orientation of the electronic device 102 to different orientations to facilitate gathering data for identification of the autonomous vehicle intended for the passenger within view of the image sensor 202 .
  • the computer processor of the electronic device 102 is configured to discriminate the detected location signals from the background of conventional imagery signals and to evaluate which orientation (line of sight) of the image sensor of the electronic device 102 correlates to the greatest intensity of the detected location signals.
  • existing image sensors typically found in smart-phone cameras may be used to capture the location signals.
  • Such smart-phone cameras can detect pulsating light at rates faster than rates the human eye can detect.
  • smart-phone cameras can detect pulsating light at rates lower than or greater than 1,000 frames per second, such as 300 frames per second, 600 frames per second, up to 1,000 frames per second or more.
  • Electronic device 102 can additionally include suitable filters such as one or more infrared (IR) filters or other color filters.
  • the filter(s) may be supported by an actuator to permit movement of the filters relative to the image sensor 202 .
  • the filter(s) may be applied to improve image quality captured by image sensor 202 , tune image quality captured by image sensor 202 , and/or select one or more desired wavelengths for selection of location signals of specified wavelengths for capture by the image sensor 202 and processing at the electronic device 102 .
  • the electronic device 102 may remove an infrared filter from the light path for image sensor 202 , using for an example, an actuator internally coupled to and controlled by a processor of electronic device 102 .
  • the electronic device may insert such filters, e.g., in sequence, into the light path of the image sensor 202 .
  • FIG. 2B illustrates a plot 210 of position versus signal strength for tilting of an electronic device for locating a particular autonomous vehicle according to an example of the disclosure.
  • plot line 212 illustrates an example curve of signal strength of the detected location signals relative to position.
  • directional lines 1 , 2 , 3 of FIG. 2A correlate to positions 1 , 2 , 3 of plot 210 illustrated in FIG. 2B associated with tilting of the electronic device 102 .
  • a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b, which is the vehicle emitting the location signals from among the group 104 of nearby autonomous vehicles.
  • electronic device 102 is oriented such that the image sensor 202 follows a line of sight along directional line 3 toward vehicle 104 c, a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b emitting the location signals.
  • the passenger may be prompted by a spoken or graphical message at the electronic device 102 to point the image sensor of electronic device 102 in different directions to facilitate determination of a stronger line of sight signal strength of detected location signals.
  • the stronger line of sight signal in the example of FIG. 2A follows along directional line 2 toward vehicle 104 b.
  • Orienting electronic device 102 to have image sensor 202 pointing toward directional line 2 points image sensor 202 at the autonomous vehicle 104 b intended to be paired with the intended passenger by remote monitoring backend system(s) 106 .
  • plot line 212 of FIG. 2B the line of sight signal strength following direction line 2 peaks as a result of image sensor being directly oriented to view the center most autonomous vehicle 104 b, in this example.
  • the application running on electronic device 102 can capture the signal strength of the location signals via the image sensor and can integrate that signal strength within suitable time intervals.
  • image sensor 202 can capture radiation, which includes the location signal, as a function of time, and the processing system of electronic device 102 can record that signal as a function of time and simultaneously record the orientation of electronic device 102 as a function of time.
  • the location signal can be of a specified pulse pattern, wavelength pattern, or both, for example, and the processing system of the electronic device can analyze the captured radiation to detect the expected signal pattern, e.g., through use of cross-correlation, for instance, that involves mathematical comparison of the detected radiation with the functional form of the expected signal, or using other mathematical pattern detection techniques known to those of skill in the art.
  • That detection can be correlated to the orientation that was recorded as a function of time to select the orientation that corresponds to the strongest amplitude of signal detection.
  • the signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle, e.g., 104 b, aligned as the proper source of the expected location signal, such as described in further detail in connection with FIGS. 4A-4D .
  • the application running on electronic device 102 can detect the intensity of the location signal (e.g., plot line 212 ) through use of signals collected under the control of one or more suitable wavelength filters controllably placed into the path of light impinging on the image sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of the electronic device 102 is simultaneously captured as a function of time.
  • the intensity of the location signal e.g., plot line 212
  • suitable wavelength filters controllably placed into the path of light impinging on the image sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of the electronic device 102 is simultaneously captured as a function of time.
  • an image of the field of view may be captured under normal imaging, and the application running on electronic device 102 may instruct the user to scan the field of view after actuating one or more desired color filters into the field of view to record the signal strength of one or more given wavelengths as a function of time while the orientation of the electronic device 102 is also recorded as a function in order to correlate the signal strength with orientation of the electronic device 102 .
  • the signal strength of the detected signal could also be directly correlated with the orientation during the scanning process.
  • That correlation can then be used by the processing system of the electronic device 102 to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle aligned as the proper source of the expected location signal, e.g., vehicle 104 b, such as described in further detail in connection with FIGS. 4A-4D .
  • FIG. 3 illustrates an illustrative framework 300 for an approach of augmented reality detection for locating a particular autonomous vehicle according to another example of the disclosure.
  • This example is similar to the example of FIG. 1 , except that in the example of FIG. 3 , accurate position information for both the autonomous vehicle and the user based at least upon, e.g., global positioning system (GPS) data and map information, provides the primary data on which identification of the intended autonomous vehicle is based.
  • GPS global positioning system
  • the autonomous vehicle 104 b determines its own position accurately, e.g., within about 1 meter or less, based on one or more of GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle.
  • the position of the electronic device 102 of the user is likewise determined accurately, e.g., to within about 1 meter or less, using GPS data and high definition map information accessed by the electronic device 102 , e.g., using conventional approaches known to those of skill in the art.
  • the electronic device 102 may determine its absolute orientation with a combination of inertial sensor data and compass technology as conventionally known.
  • the electronic device can prompt a user via a visual instruction, audio instruction, or both, to point the electronic device toward a pinpoint location that identifies the intended autonomous vehicle.
  • FIG. 3 is similar to FIG. 1 , and the prior discussion relating to FIGS. 1-2 is equally applicable to the example of FIG. 3 . Thus, where applicable, the same reference numerals as in FIG. 1 have been retained for FIG. 3 .
  • a passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle from a pool of autonomous vehicles. Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102 .
  • Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet.
  • Remote monitoring backend system(s) 108 is like that described for FIG. 1 .
  • the passenger can be alerted to that fact via a wireless communication to the electronic device 102 , e.g., from either the autonomous vehicle itself or from the backend system 108 .
  • a passenger can then be prompted by either a visual instruction, audio instruction, or both, at the electronic device 102 to point the electronic device 102 having image sensor 202 toward a designated area where the autonomous vehicle is located.
  • the location and orientation of the electronic device can be communicated to the backend system 108 or the autonomous vehicle 104 b, and either of those systems can instruct the user to reorient the electronic device 102 to either the left or the right until the proper line of sight orientation with the pinpoint location of the autonomous vehicle 104 b is achieved.
  • electronic device 102 may have functionality such that when it receives the pinpoint location of the intended vehicle via wireless and/or wired communication network 110 , the pinpoint location is correlated to the captured image displayed to the passenger on the graphical user interface 106 of electronic device 102 . With the intended autonomous vehicle correlation to the image on graphical user interface 106 , the autonomous vehicle may be identified to the passenger via an indication, as will be described further herein.
  • FIGS. 4A-4D illustrate illustrative indications of an autonomous vehicle according to examples of the disclosure.
  • An indication to the intended passenger of the respectively paired autonomous vehicle can be readily customized by the passenger via graphical user interface 106 .
  • the indication may be one or more geometric annotation(s) 402 , as illustrated in FIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.”
  • the indication may be shaded and/or highlighted with color or gray-scale, e.g., color or gray-scale annotation 404 , as illustrated in FIG. 4B .
  • the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C .
  • FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle.
  • Superimposed image 408 can be the intended passenger's face or a cartoon character.
  • FIG. 5 is a flow chart 500 of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle according to examples of the disclosure.
  • a request for an autonomous vehicle is provided by electronic device 102 to a remote monitoring backend system 108 , for example, via user input into an application installed on electronic device 102 .
  • the electronic device 102 may receive, at 504 , a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger.
  • the signal can include a location signal from the autonomous vehicle, which can include, for example, rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection.
  • the pulses while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger.
  • the one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle.
  • the one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths.
  • Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102 , e.g., via text messaging or other suitable messaging.
  • the intended passenger may be prompted, at 506 , by electronic device 102 to orient an image sensor 202 towards a designated area of the accessible range.
  • the intended passenger may be prompted by receiving a text message, a phone call, a pop-up notification, a video call, and/or an audible alarm or voice message on electronic device 102 .
  • Image sensor 202 images, at 508 , the designated area for the autonomous vehicle.
  • the user may be prompted by the electronic device via any suitable audible or graphical message or indication such as previously described to point the electronic device 102 toward a direction that corresponds to the designated area at which the intended vehicle 104 b is expected, e.g., based on known GPS location information of both the electronic device 102 and the intended vehicle 104 b, and based on directional (orientation) information known at the electronic device 102 via its inertial sensors and conventional compass functionality.
  • the processing system can process collected location signal data as a function of orientation such as described in the examples above, to determine the orientation of the electronic device that corresponds to the strongest location signal, and can determine that vehicle 104 b is the source of the location signal and is the intended vehicle 104 b.
  • a visual representation of the imaged designated area on the graphical user interface 106 is generated and displayed on the display screen of the electronic device 102 .
  • the autonomous vehicle is identified, at 512 , by the electronic device 102 within the generated visual representation, e.g., via a suitable graphical indicator such as shown in FIGS. 4A-4D .
  • the indicator may be one or more geometric annotation(s) 402 , as illustrated in FIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.”
  • the indication may be shaded and/or highlighted with a color, e.g., color annotation 404 , as illustrated in FIG. 4B .
  • the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C .
  • FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle.
  • Superimposed image 408 can be the intended passenger's face or a cartoon character.
  • the methods and systems described herein may be implemented using any suitable computer processing system with any suitable combination of hardware, software and/or firmware.
  • entities that control the pool of autonomous vehicles 104 can interact with remote monitoring backend system 106 hosted on one or more computer servers through a wireless and/or wired communication network 110 .
  • the computer processing system 112 may execute software operations, program instructions or routines to implement calculations and analyses described above.
  • Such program instructions, accumulated data, and processed data may be stored on one or more non-transitory computer-readable memories 114 and/or one or more data stores for use in databases 116 .
  • Communications may be carried out according to a client server architecture whereby passengers and autonomous vehicles 104 may access the remote monitoring backend system 108 via one or more servers via one or more wireless and/or wired communication networks 110 .
  • the systems may include element managers, real-time data buffers, conveyors, file input processors, database indices, data buffers and data managers for managing data and processing.
  • the remote monitoring backend system(s) 108 may also include multiple displays, display interfaces, input/output devices such as keyboards, microphones, mice, touch screens and the like for permitting administrators or support personnel to manage the remote monitoring backend system(s) 108 .
  • the systems and methods may include and utilize data signals conveyed via networks (e.g., local area network, wide area network, internet, combinations thereof, etc.), fiber optic medium, carrier waves, wireless networks, etc. for communication with one or more data processing devices.
  • the data signals can carry any or all of the data disclosed herein that is provided to or from a device.
  • the methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing system.
  • the software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein. Any suitable computer languages may be used such as C, C++, Java, etc., as will be appreciated by those skilled in the art. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to carry out the methods and systems described herein.
  • the systems' and methods' data may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.).
  • storage devices and programming constructs e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.
  • data structures describe formats for use in organizing and storing data in databases, programs, memory, or other non-transitory computer-readable media for use by a computer program.
  • a module or processor includes but is not limited to a unit of code that performs a software operation, and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code.
  • the software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Human Computer Interaction (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Medical Informatics (AREA)
  • Traffic Control Systems (AREA)
  • Navigation (AREA)

Abstract

Disclosed embodiments include methods, systems, and non-transitory computer readable medium for augmented reality detection for locating a particular autonomous vehicle. In a non-limiting embodiment, an illustrative method includes providing, by an electronic device to a remote server, a request for an autonomous vehicle. A first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger is received by an electronic device. The passenger is prompted by the electronic device to orient an image sensor of the electronic device towards a designated area of the accessible range. The designated area is imaged by the image sensor for the autonomous vehicle. A visual representation of the imaged designated area is generated on a graphical user interface of the electronic device. The autonomous vehicle is identified within the generated visual representation.

Description

    CROSS-REFERENCE TO RELATED APPLICATION
  • The present disclosure is a continuation (CON) of co-pending U.S. patent application Ser. No. 16/810,652, filed on Mar. 5, 2020, and entitled “AUGMENTED REALITY DETECTION FOR LOCATING AUTONOMOUS VEHICLES,” the contents of which are incorporated in full by reference herein for all purposes.
  • TECHNICAL FIELD
  • The present disclosure relates to autonomous vehicles.
  • BACKGROUND
  • The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.
  • Autonomous vehicles are of great interest for transportation applications and can provide benefits to the ride-sharing industry. Conventionally, when an intended passenger makes a request for a ride-sharing vehicle, the intended passenger may have to contact the driver or have human interactions with the driver of the vehicle to identify his/her particular vehicle, especially when in a crowded area such as a city or an airport. Such interaction, however, may be challenging, if even possible, when attempting to locate an autonomous vehicle within a pool of autonomous vehicles or a crowded area with obstructions.
  • SUMMARY
  • Disclosed embodiments include methods, systems, and non-transitory computer readable medium for augmented reality detection for locating a particular autonomous vehicle.
  • In a non-limiting embodiment, an illustrative method includes providing, by an electronic device to a remote server, a request for an autonomous vehicle. A first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger is received by an electronic device. The passenger is prompted by the electronic device to orient an image sensor of the electronic device towards a designated area of the accessible range. The designated area is imaged by the image sensor for the autonomous vehicle. A visual representation of the imaged designated area is generated on a graphical user interface of the electronic device. The autonomous vehicle is identified within the generated visual representation.
  • In another non-limiting embodiment, an illustrative system includes a computer processing system and a computer memory. The computer processing system is configured to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
  • In another non-limiting embodiment, non-transitory computer readable medium includes program instructions for locating autonomous vehicles. The program instructions when executed cause a computer processing system to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
  • Further features, advantages, and areas of applicability will become apparent from the description provided herein. It should be understood that the description and specific examples are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present disclosure in any way. The components in the figures are not necessarily to scale, with emphasis instead being placed upon illustrating the principles of the disclosed embodiments. In the drawings:
  • FIG. 1 is a block diagram in partial schematic form of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIG. 2A is a block diagram in partial schematic form of a partial top view of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIG. 2B is a graph of signal strength of detected location signals as a function of image sensor direction of an electronic device for locating a particular autonomous vehicle.
  • FIG. 3 is a block diagram in partial schematic form of another illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
  • FIGS. 4A-4D are top views of illustrative identifications of an autonomous vehicle.
  • FIG. 5 is a flow chart of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle.
  • Like reference symbols in the various drawings generally indicate like elements.
  • DETAILED DESCRIPTION
  • The following description is merely illustrative in nature and is not intended to limit the present disclosure, application, or uses. It will be noted that the first digit of three-digit reference numbers and the first two digits of four-digit reference numbers correspond to the first digit of one-digit figure numbers and the first two digits of two-digit figure numbers, respectively, in which the element first appears.
  • FIG. 1 illustrates an illustrative framework 100 for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure. A passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle, e.g., from a pool of autonomous vehicles and then use the application on the electronic device 102 to select the proper intended autonomous vehicle from a group of nearby vehicles 104. An autonomous vehicle can be configured with an array of sensors 105, including, e.g., LIDAR (light detection and ranging), camera systems for acquiring real-time video imagery of obstacles and other vehicles, and GPS (global positioning system). An autonomous vehicle can also be configured with wireless communication systems for sending and receiving communications regarding road information and traffic in real-time, as well as a computer for applying driving rules and making driving decisions based on data and information acquired from the foregoing, in a suitable manner such as conventionally known in the art. Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102. Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet. Remote monitoring backend system(s) 108 may select a particular autonomous vehicle, e.g., from a pool of autonomous vehicles based on one or more criteria which may include, but are not limited to, distance to the intended passenger's pick-up location, amount of time it would take to arrive at the intended passenger's pick-up location, destination of the intended passenger, and/or desired vehicle accommodations of the intended passenger. Remote monitoring backend system(s) 108 may provide the request to the selected autonomous vehicle via wireless and/or wired communication network 110. Remote monitoring backend system(s) 108 may also communicate with on-board computers of the pool of autonomous vehicles 104 to control various aspects of vehicle features, as will be described further herein.
  • Remote monitoring backend system(s) 108 may include a computer processing system 108, which may include one or more computer processing units (CPUs) 112 located in one or more computer systems, and one or more memories 114, which may include RAM, ROM, and any suitable nonvolatile memory. Remote monitoring backend system(s) 108 may also include one or more databases 116 to store data for both vehicles and drivers, as will be explained further herein, and one or more interfaces 118 to facilitate communication via networks including the Internet and for facilitating input/output of data. Remote monitoring backend system(s) 108 carries out monitoring of vehicles, vehicle systems, vehicle usage, and systems usage for individual vehicles and for individual drivers who use those vehicles, and carries out data analytics on the data gathered from the vehicles and the vehicle systems. In examples, the pool of autonomous vehicles may include various detectors and sensors for monitoring vehicle dynamics, usage, and demand placed on vehicle systems during the driving of the pool of autonomous vehicles. For example, the autonomous vehicles may include GPS units for monitoring location, speed, and direction, battery monitoring sensors in the form of electrical circuits, for monitoring battery usage, battery discharge as a function of time, and battery charging rates, and other suitable sensors for monitoring the health and status of other major subsystems of the vehicle, such as removable modules, and for detecting warning or fault conditions. Such data may be stored in an onboard computer system with suitable memory and may be uploaded to the remote monitoring backend system(s) 108 through wireless and/or wired communication network 110 via a vehicle onboard wireless transceiver and/or via wired communication.
  • Each autonomous vehicle of the pool of autonomous vehicles can emit one or more location signals. For example, the one or more location signals of each autonomous vehicle may include rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection. The pulses, while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger. The one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle. The one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths. Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102, e.g., via text messaging or other suitable messaging. Electronic device 102 can detect the one or more location signals of the autonomous vehicle once it is in an accessible range of the autonomous vehicle. The intended passenger can be prompted, via a visual instruction displayed at graphical user interface 106 or via an audio instruction, or both, to orient an image sensor (not shown in FIG. 1) of the electronic device 102 toward a designated area once the autonomous vehicle is within an accessible range. After the passenger orients the image sensor to the designated area, the image sensor of electronic device 102 may capture a scene of the pool of autonomous vehicles 104, as will be described further herein. A visual representation of the imaged area may be generated by electronic device 102 using one or more internal processing components. The visual representation of the imaged area may be displayed on graphical user interface 106 with the autonomous vehicle identified to the passenger via one or more indications 120, as will be described further herein. In some cases, the autonomous vehicle may be partially obstructed by one or more obstructions 122 and can still be identified to the intended passenger. Access to the autonomous vehicle may be provided to the intended passenger via an access button (field) 122 on graphical user interface 106, e.g., an unlock graphic field on the touch-sensitive display screen displaying graphical user interface 106 that the intended passenger may press.
  • FIG. 2A illustrates a block diagram of a partial top view 200 of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure. Electronic device 102 may include an image sensor 202. The image sensor 202 may be sensitive to visible light, non-visible light (e.g., infrared light), or both. The image sensor 202 may be a conventional image sensor of the type typically found in smart phones. In an example, a user, having already communicated a request for a vehicle using the app running on the electronic device 102, may be alerted by the electronic device 102 that an intended vehicle is with detectable range of the user and may direct the use via a visual instruction, an audio instruction, or both, to point the electronic device toward a field of view encompassing one or more vehicles and to scan the image sensor of the electronic device 102 across the field of view by tilting the electronic device 102 laterally as illustrated in FIG. 2A.
  • Electronic device 102 may be tilted in one or more axes of a three-dimensional system, e.g., x,y,z-system as illustrated in FIG. 2A. Tilting of electronic device 102 allows image sensor 202 to capture images of the designated area with different orientations whereby different vehicles in the field of view are aligned with different lines of sight (e.g., dotted lines 1, 2 and 3). For example, electronic device 102 may be tilted slightly back and forth to scan and capture the signal strength of the one or more location signals. The one or more location signals can be recorded as a function of orientation of the electronic device, e.g., the vector direction of the image sensor 202 of the electronic device 102 and the orientation with the highest line of sight signal strength may be the one that points to the intended autonomous vehicle. That information can then be correlated to the captured image associated with the various orientations to flag the intended autonomous vehicle using an indicator 120, as will be described further herein.
  • Processing of the image information by electronic device 102 may determine an orientation of electronic device 102 which provides a strongest line of sight signal to capture an image of the pool of autonomous vehicles. For example, electronic device 102 may have an internal inertial sensor that provides data to a processor of the electronic device 102. As noted above, the electronic device 102 may prompt the user to slowly scan the image sensor 202 across a field of view containing several vehicles, may evaluate the signal strength of detected location signals as a function of orientation based on correlated inertial data simultaneously captured by the electronic device, and then identify which location in the captured imagery corresponds to the greatest signal strength of the detected location signals, and then display on the electronic device that location superimposed with the imagery of the scene so as to identify to the user the intended autonomous vehicle. In another example, the electronic device 102 may prompt a passenger, via graphical user interface 106, to modify the orientation of the electronic device 102 to different orientations to facilitate gathering data for identification of the autonomous vehicle intended for the passenger within view of the image sensor 202. Although the image sensor 202 of the electronic device 102 is exposed to conventional image signals during this process, the computer processor of the electronic device 102 is configured to discriminate the detected location signals from the background of conventional imagery signals and to evaluate which orientation (line of sight) of the image sensor of the electronic device 102 correlates to the greatest intensity of the detected location signals.
  • In this regard, existing image sensors typically found in smart-phone cameras may be used to capture the location signals. Such smart-phone cameras can detect pulsating light at rates faster than rates the human eye can detect. For example, smart-phone cameras can detect pulsating light at rates lower than or greater than 1,000 frames per second, such as 300 frames per second, 600 frames per second, up to 1,000 frames per second or more. Electronic device 102 can additionally include suitable filters such as one or more infrared (IR) filters or other color filters. The filter(s) may be supported by an actuator to permit movement of the filters relative to the image sensor 202. The filter(s) may be applied to improve image quality captured by image sensor 202, tune image quality captured by image sensor 202, and/or select one or more desired wavelengths for selection of location signals of specified wavelengths for capture by the image sensor 202 and processing at the electronic device 102. When the location signals are within an infrared light spectrum, for example, the electronic device 102 may remove an infrared filter from the light path for image sensor 202, using for an example, an actuator internally coupled to and controlled by a processor of electronic device 102. When the location signals comprise specified color wavelengths, e.g., a progression of signal pulses of one or more wavelengths, e.g., in sequence, the electronic device may insert such filters, e.g., in sequence, into the light path of the image sensor 202.
  • FIG. 2B illustrates a plot 210 of position versus signal strength for tilting of an electronic device for locating a particular autonomous vehicle according to an example of the disclosure. For illustration purposes, plot line 212 illustrates an example curve of signal strength of the detected location signals relative to position. In an example, directional lines 1, 2, 3 of FIG. 2A correlate to positions 1, 2, 3 of plot 210 illustrated in FIG. 2B associated with tilting of the electronic device 102. When electronic device 102 is oriented such that the image sensor 202 follows a line of sight along directional line 1 toward vehicle 104 a of FIG. 2A, a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b, which is the vehicle emitting the location signals from among the group 104 of nearby autonomous vehicles. Similarly, when electronic device 102 is oriented such that the image sensor 202 follows a line of sight along directional line 3 toward vehicle 104 c, a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b emitting the location signals. In an example, the passenger may be prompted by a spoken or graphical message at the electronic device 102 to point the image sensor of electronic device 102 in different directions to facilitate determination of a stronger line of sight signal strength of detected location signals. The stronger line of sight signal in the example of FIG. 2A, follows along directional line 2 toward vehicle 104 b. Orienting electronic device 102 to have image sensor 202 pointing toward directional line 2 points image sensor 202 at the autonomous vehicle 104 b intended to be paired with the intended passenger by remote monitoring backend system(s) 106. As illustrated by plot line 212 of FIG. 2B, the line of sight signal strength following direction line 2 peaks as a result of image sensor being directly oriented to view the center most autonomous vehicle 104 b, in this example. In order to generate plot line 212, the application running on electronic device 102 can capture the signal strength of the location signals via the image sensor and can integrate that signal strength within suitable time intervals. For example, image sensor 202 can capture radiation, which includes the location signal, as a function of time, and the processing system of electronic device 102 can record that signal as a function of time and simultaneously record the orientation of electronic device 102 as a function of time. As noted previously, the location signal can be of a specified pulse pattern, wavelength pattern, or both, for example, and the processing system of the electronic device can analyze the captured radiation to detect the expected signal pattern, e.g., through use of cross-correlation, for instance, that involves mathematical comparison of the detected radiation with the functional form of the expected signal, or using other mathematical pattern detection techniques known to those of skill in the art. With the timing of the location signal having been detected, that detection can be correlated to the orientation that was recorded as a function of time to select the orientation that corresponds to the strongest amplitude of signal detection. The signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle, e.g., 104 b, aligned as the proper source of the expected location signal, such as described in further detail in connection with FIGS. 4A-4D.
  • In another example, the application running on electronic device 102 can detect the intensity of the location signal (e.g., plot line 212) through use of signals collected under the control of one or more suitable wavelength filters controllably placed into the path of light impinging on the image sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of the electronic device 102 is simultaneously captured as a function of time. In this approach, for example, an image of the field of view may be captured under normal imaging, and the application running on electronic device 102 may instruct the user to scan the field of view after actuating one or more desired color filters into the field of view to record the signal strength of one or more given wavelengths as a function of time while the orientation of the electronic device 102 is also recorded as a function in order to correlate the signal strength with orientation of the electronic device 102. The signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used by the processing system of the electronic device 102 to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle aligned as the proper source of the expected location signal, e.g., vehicle 104 b, such as described in further detail in connection with FIGS. 4A-4D.
  • FIG. 3 illustrates an illustrative framework 300 for an approach of augmented reality detection for locating a particular autonomous vehicle according to another example of the disclosure. This example is similar to the example of FIG. 1, except that in the example of FIG. 3, accurate position information for both the autonomous vehicle and the user based at least upon, e.g., global positioning system (GPS) data and map information, provides the primary data on which identification of the intended autonomous vehicle is based. For example, the autonomous vehicle 104 b determines its own position accurately, e.g., within about 1 meter or less, based on one or more of GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle. Conventional approaches known in the art for identifying accurate positions of autonomous vehicles utilizing GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle may be utilized in this regard. Similarly, the position of the electronic device 102 of the user is likewise determined accurately, e.g., to within about 1 meter or less, using GPS data and high definition map information accessed by the electronic device 102, e.g., using conventional approaches known to those of skill in the art. Moreover, the electronic device 102 may determine its absolute orientation with a combination of inertial sensor data and compass technology as conventionally known. With this information, as explained below, the electronic device can prompt a user via a visual instruction, audio instruction, or both, to point the electronic device toward a pinpoint location that identifies the intended autonomous vehicle. In other respects, FIG. 3 is similar to FIG. 1, and the prior discussion relating to FIGS. 1-2 is equally applicable to the example of FIG. 3. Thus, where applicable, the same reference numerals as in FIG. 1 have been retained for FIG. 3.
  • As shown in FIG. 3, similar to FIG. 1, a passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle from a pool of autonomous vehicles. Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102. Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet. Remote monitoring backend system(s) 108 is like that described for FIG. 1.
  • When the autonomous vehicle 104 b has arrived within a desired proximity of the location of the intended passenger, the passenger can be alerted to that fact via a wireless communication to the electronic device 102, e.g., from either the autonomous vehicle itself or from the backend system 108. A passenger can then be prompted by either a visual instruction, audio instruction, or both, at the electronic device 102 to point the electronic device 102 having image sensor 202 toward a designated area where the autonomous vehicle is located. For instance, the location and orientation of the electronic device can be communicated to the backend system 108 or the autonomous vehicle 104 b, and either of those systems can instruct the user to reorient the electronic device 102 to either the left or the right until the proper line of sight orientation with the pinpoint location of the autonomous vehicle 104 b is achieved. Additionally, electronic device 102 may have functionality such that when it receives the pinpoint location of the intended vehicle via wireless and/or wired communication network 110, the pinpoint location is correlated to the captured image displayed to the passenger on the graphical user interface 106 of electronic device 102. With the intended autonomous vehicle correlation to the image on graphical user interface 106, the autonomous vehicle may be identified to the passenger via an indication, as will be described further herein.
  • FIGS. 4A-4D illustrate illustrative indications of an autonomous vehicle according to examples of the disclosure. An indication to the intended passenger of the respectively paired autonomous vehicle can be readily customized by the passenger via graphical user interface 106. For example, the indication may be one or more geometric annotation(s) 402, as illustrated in FIG. 4A, such as an arrow accompanied by text stating: “YOUR RIDE.” In another example, the indication may be shaded and/or highlighted with color or gray-scale, e.g., color or gray-scale annotation 404, as illustrated in FIG. 4B. In yet another example, the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C. FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle. Superimposed image 408 can be the intended passenger's face or a cartoon character.
  • FIG. 5 is a flow chart 500 of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle according to examples of the disclosure. As shown at 502, a request for an autonomous vehicle is provided by electronic device 102 to a remote monitoring backend system 108, for example, via user input into an application installed on electronic device 102. The electronic device 102 may receive, at 504, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger. The signal can include a location signal from the autonomous vehicle, which can include, for example, rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection. The pulses, while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger. The one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle. The one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths. Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102, e.g., via text messaging or other suitable messaging. The intended passenger may be prompted, at 506, by electronic device 102 to orient an image sensor 202 towards a designated area of the accessible range. For example, the intended passenger may be prompted by receiving a text message, a phone call, a pop-up notification, a video call, and/or an audible alarm or voice message on electronic device 102. Image sensor 202 images, at 508, the designated area for the autonomous vehicle. For example, the user may be prompted by the electronic device via any suitable audible or graphical message or indication such as previously described to point the electronic device 102 toward a direction that corresponds to the designated area at which the intended vehicle 104 b is expected, e.g., based on known GPS location information of both the electronic device 102 and the intended vehicle 104 b, and based on directional (orientation) information known at the electronic device 102 via its inertial sensors and conventional compass functionality. The processing system can process collected location signal data as a function of orientation such as described in the examples above, to determine the orientation of the electronic device that corresponds to the strongest location signal, and can determine that vehicle 104 b is the source of the location signal and is the intended vehicle 104 b. At 510, a visual representation of the imaged designated area on the graphical user interface 106 is generated and displayed on the display screen of the electronic device 102. The autonomous vehicle is identified, at 512, by the electronic device 102 within the generated visual representation, e.g., via a suitable graphical indicator such as shown in FIGS. 4A-4D. For example, the indicator may be one or more geometric annotation(s) 402, as illustrated in FIG. 4A, such as an arrow accompanied by text stating: “YOUR RIDE.” In another example, the indication may be shaded and/or highlighted with a color, e.g., color annotation 404, as illustrated in FIG. 4B. In yet another example, the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C. FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle. Superimposed image 408 can be the intended passenger's face or a cartoon character.
  • The methods and systems described herein may be implemented using any suitable computer processing system with any suitable combination of hardware, software and/or firmware. As shown in FIGS. 1 and 3, entities that control the pool of autonomous vehicles 104 can interact with remote monitoring backend system 106 hosted on one or more computer servers through a wireless and/or wired communication network 110. The computer processing system 112 may execute software operations, program instructions or routines to implement calculations and analyses described above. Such program instructions, accumulated data, and processed data may be stored on one or more non-transitory computer-readable memories 114 and/or one or more data stores for use in databases 116. Communications may be carried out according to a client server architecture whereby passengers and autonomous vehicles 104 may access the remote monitoring backend system 108 via one or more servers via one or more wireless and/or wired communication networks 110.
  • The systems may include element managers, real-time data buffers, conveyors, file input processors, database indices, data buffers and data managers for managing data and processing. The remote monitoring backend system(s) 108 may also include multiple displays, display interfaces, input/output devices such as keyboards, microphones, mice, touch screens and the like for permitting administrators or support personnel to manage the remote monitoring backend system(s) 108.
  • This written description describes illustrative embodiments, but other variations fall within the scope of the disclosure. For example, the systems and methods may include and utilize data signals conveyed via networks (e.g., local area network, wide area network, internet, combinations thereof, etc.), fiber optic medium, carrier waves, wireless networks, etc. for communication with one or more data processing devices. The data signals can carry any or all of the data disclosed herein that is provided to or from a device.
  • The methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing system. The software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein. Any suitable computer languages may be used such as C, C++, Java, etc., as will be appreciated by those skilled in the art. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to carry out the methods and systems described herein.
  • The systems' and methods' data (e.g., associations, mappings, data input, data output, intermediate data results, final data results, etc.) may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.). It is noted that data structures describe formats for use in organizing and storing data in databases, programs, memory, or other non-transitory computer-readable media for use by a computer program.
  • The computer components, software modules, functions, data stores and data structures described herein may be connected directly or indirectly to each other in order to allow the flow of data needed for their operations. It is also noted that a module or processor includes but is not limited to a unit of code that performs a software operation, and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code. The software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.
  • It should be understood that as used in the description herein and throughout the claims that follow, the meaning of “a,” “an,” and “the” includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise. In addition, as used in the description herein and throughout the claims that follow, the meanings of “and” and “or” include both the conjunctive and disjunctive and may be used interchangeably unless the context expressly dictates otherwise; the phrase “exclusive or” may be used to indicate situation where only the disjunctive meaning may apply. In addition, as used in the description herein and throughout the claims that follow, the meaning of “about” and/or “approximately” refers to ±10% of the quantity indicated, unless otherwise indicated.
  • While the disclosed subject matter has been described in terms of illustrative embodiments, it will be understood by those skilled in the art that various modifications can be made thereto without departing from the scope of the claimed subject matter as set forth in the claims.

Claims (20)

What is claimed is:
1. A method for locating a particular autonomous vehicle, the method comprising:
receiving, at a server, a request for an autonomous vehicle from an application executing on an electronic device;
dispatching the autonomous vehicle to a location of the electronic device; and
in response to receiving an indication that the autonomous vehicle is within an accessible range of the electronic device, causing, by the server, an identification signal to be transmitted from the autonomous vehicle, the identification signal being transmitted by the autonomous vehicle for capture by a sensor of the electronic device.
2. The method of claim 1, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on a communication signal received at the electronic device from the autonomous vehicle, the communication signal chosen from a Wi-Fi signal and a Bluetooth signal.
3. The method of claim 1, further comprising, in response to receiving, at the server, an indication that the identification signal has been captured by the sensor of the electronic device, causing, by the server, a presentation of the autonomous vehicle to be identified within a generated visual representation on a graphical user interface of the application displayed on the electronic device.
4. The method of claim 1, further comprising obtaining, at the server, a vector between the electronic device and the autonomous vehicle, the vector based on an intensity of the identification signal as a function of scanning performed by the electronic device for the identification signal.
5. The method of claim 1, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on location information of the autonomous vehicle and of the electronic device.
6. The method of claim 1, further comprising:
receiving, by the server and from the autonomous vehicle, a designated area within the accessible range of the electronic device; and
providing, by the server, the designated area to the electronic device for the electronic device to indicate to a user of the electronic device where to orient the sensor of the electronic device for the capture of the identification signal.
7. The method of claim 1, further comprising:
selecting, by the server, the autonomous vehicle from a pool of autonomous vehicles; and
providing, by the server, the request to the autonomous vehicle.
8. A system for an autonomous vehicle, comprising:
a processor; and
memory storing instructions that, when executed, cause the processor to:
receive a request for an autonomous vehicle from an application executing on an electronic device;
dispatch the autonomous vehicle to a location of the electronic device; and
in response to receiving an indication that the autonomous vehicle is within an accessible range of the electronic device, cause an identification signal to be transmitted from the autonomous vehicle, the identification signal being transmitted by the autonomous vehicle for capture by a sensor of the electronic device.
9. The system of claim 8, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on a communication signal received at the electronic device from the autonomous vehicle, the communication signal chosen from a Wi-Fi signal and a Bluetooth signal.
10. The system of claim 8, wherein the memory storing instructions that, when executed, further cause the processor to, in response to receiving an indication that the identification signal has been captured by the sensor of the electronic device, cause a presentation of the autonomous vehicle to be identified within a generated visual representation on a graphical user interface of the application displayed on the electronic device.
11. The system of claim 8, wherein the memory storing instructions that, when executed, further cause the processor to obtain a vector between the electronic device and the autonomous vehicle, the vector based on an intensity of the identification signal as a function of scanning performed by the electronic device for the identification signal.
12. The system of claim 8, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on location information of the autonomous vehicle and of the electronic device.
13. The system of claim 8, wherein the memory storing instructions that, when executed, further cause the processor to:
receive, from the autonomous vehicle, a designated area within the accessible range of the electronic device; and
provide the designated area to the electronic device for the electronic device to indicate to a user of the electronic device where to orient the sensor of the electronic device for the capture of the identification signal.
14. The system of claim 8, wherein the memory storing instructions that, when executed, further cause the processor to:
select the autonomous vehicle from a pool of autonomous vehicles; and
provide the request to the autonomous vehicle.
15. A non-transitory computer-readable medium having instructions stored thereon for programming a processing device to perform steps of:
receiving a request for an autonomous vehicle from an application executing on an electronic device;
dispatching the autonomous vehicle to a location of the electronic device; and
in response to receiving an indication that the autonomous vehicle is within an accessible range of the electronic device, causing an identification signal to be transmitted from the autonomous vehicle, the identification signal being transmitted by the autonomous vehicle for capture by a sensor of the electronic device.
16. The non-transitory computer readable medium of claim 15, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on a communication signal received at the electronic device from the autonomous vehicle, the communication signal chosen from a Wi-Fi signal and a Bluetooth signal.
17. The non-transitory computer readable medium of claim 15, wherein the steps further include, in response to receiving an indication that the identification signal has been captured by the sensor of the electronic device, causing a presentation of the autonomous vehicle to be identified within a generated visual representation on a graphical user interface of the application displayed on the electronic device.
18. The non-transitory computer readable medium of claim 15, wherein the steps further include obtaining a vector between the electronic device and the autonomous vehicle, the vector based on an intensity of the identification signal as a function of scanning performed by the electronic device for the identification signal.
19. The non-transitory computer readable medium of claim 15, wherein the indication that the autonomous vehicle is within an accessible range of the electronic device is based on location information of the autonomous vehicle and of the electronic device.
20. The non-transitory computer readable medium of claim 15, wherein the steps further include:
receiving, from the autonomous vehicle, a designated area within the accessible range of the electronic device; and
providing the designated area to the electronic device for the electronic device to indicate to a user of the electronic device where to orient the sensor of the electronic device for the capture of the identification signal.
US17/540,586 2020-03-05 2021-12-02 Augmented reality detection for locating autonomous vehicles Abandoned US20220092829A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US17/540,586 US20220092829A1 (en) 2020-03-05 2021-12-02 Augmented reality detection for locating autonomous vehicles

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/810,652 US11263787B2 (en) 2020-03-05 2020-03-05 Augmented reality detection for locating autonomous vehicles
US17/540,586 US20220092829A1 (en) 2020-03-05 2021-12-02 Augmented reality detection for locating autonomous vehicles

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US16/810,652 Continuation US11263787B2 (en) 2020-03-05 2020-03-05 Augmented reality detection for locating autonomous vehicles

Publications (1)

Publication Number Publication Date
US20220092829A1 true US20220092829A1 (en) 2022-03-24

Family

ID=77388804

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/810,652 Active US11263787B2 (en) 2020-03-05 2020-03-05 Augmented reality detection for locating autonomous vehicles
US17/540,586 Abandoned US20220092829A1 (en) 2020-03-05 2021-12-02 Augmented reality detection for locating autonomous vehicles

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US16/810,652 Active US11263787B2 (en) 2020-03-05 2020-03-05 Augmented reality detection for locating autonomous vehicles

Country Status (3)

Country Link
US (2) US11263787B2 (en)
CN (1) CN113361316A (en)
DE (1) DE102021201774A1 (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10471611B2 (en) 2016-01-15 2019-11-12 Irobot Corporation Autonomous monitoring robot systems
US20240061129A1 (en) * 2022-08-18 2024-02-22 Textron Innovations Inc. Updating geofencing data in a utility vehicle control device

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100159833A1 (en) * 2008-12-19 2010-06-24 At&T Mobility Ii Llc Headset Locator Device
US20130010103A1 (en) * 2010-03-14 2013-01-10 Ns Solutions Corporation Information processing system, information processing method and program, information processing apparatus, vacant space guidance system, vacant space guidance method and program, image display system, image display method and program
US20150363915A1 (en) * 2014-06-16 2015-12-17 Huawei Technologies Co., Ltd. Method and Apparatus for Presenting Panoramic Photo in Mobile Terminal, and Mobile Terminal
US20170345209A1 (en) * 2016-05-31 2017-11-30 Matthew Adam Simari Visualization alignment for three-dimensional scanning
US20190073111A1 (en) * 2017-09-05 2019-03-07 Future Mobility Corporation Limited Methods and systems using a rotatable three-dimensional object providing visual user information on a display
US10809081B1 (en) * 2018-05-03 2020-10-20 Zoox, Inc. User interface and augmented reality for identifying vehicles and persons

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7720436B2 (en) * 2006-01-09 2010-05-18 Nokia Corporation Displaying network objects in mobile devices based on geolocation
DE102009037835B4 (en) * 2009-08-18 2012-12-06 Metaio Gmbh Method for displaying virtual information in a real environment
JP6065370B2 (en) * 2012-02-03 2017-01-25 ソニー株式会社 Information processing apparatus, information processing method, and program
US9852381B2 (en) * 2012-12-20 2017-12-26 Nokia Technologies Oy Method and apparatus for providing behavioral pattern generation for mixed reality objects
US9959674B2 (en) * 2013-02-26 2018-05-01 Qualcomm Incorporated Directional and X-ray view techniques for navigation using a mobile device
US9041741B2 (en) * 2013-03-14 2015-05-26 Qualcomm Incorporated User interface for a head mounted display
EP2983138A4 (en) * 2013-04-04 2017-02-22 Sony Corporation Display control device, display control method and program
US10217284B2 (en) * 2013-09-30 2019-02-26 Qualcomm Incorporated Augmented virtuality
US10802582B1 (en) * 2014-04-22 2020-10-13 sigmund lindsay clements Eye tracker in an augmented reality glasses for eye gaze to input displayed input icons
US20160140868A1 (en) * 2014-11-13 2016-05-19 Netapp, Inc. Techniques for using augmented reality for computer systems maintenance
US20160358383A1 (en) * 2015-06-05 2016-12-08 Steffen Gauglitz Systems and methods for augmented reality-based remote collaboration
KR102355455B1 (en) * 2016-06-20 2022-01-24 매직 립, 인코포레이티드 Augmented Reality Display System for Evaluation and Modification of Neurological Conditions Including Visual Processing and Perceptual States
US10078916B2 (en) * 2016-07-01 2018-09-18 Invia Robotics, Inc. Pick to augmented reality
EP3321892A1 (en) * 2016-11-10 2018-05-16 Gemalto Sa A physical key for provisioning a communication device with data allowing it to access a vehicle resource
US9841814B1 (en) * 2017-01-27 2017-12-12 Emergent AR Platforms Corp. Intentional user experience
CA3054282A1 (en) * 2017-03-02 2018-09-07 Motorola Solutions, Inc. Method and apparatus for gathering visual data using an augmented-reality application
US11120264B2 (en) 2017-06-02 2021-09-14 Apple Inc. Augmented reality interface for facilitating identification of arriving vehicle
US11173798B2 (en) * 2018-02-12 2021-11-16 Ford Global Technologies, Llc Systems and methods for vehicle wireless charging
US10546169B2 (en) * 2018-03-22 2020-01-28 Hall Labs Llc Augmented reality navigation system
US10318811B1 (en) * 2018-04-22 2019-06-11 Bubbler International Llc Methods and systems for detecting objects by non-visible radio frequencies and displaying associated augmented reality effects
KR102627612B1 (en) * 2019-02-19 2024-01-22 삼성전자주식회사 Method for displaying nerby information using augmented reality and electonic device therof
US11598643B2 (en) * 2019-03-26 2023-03-07 Snap Inc. Augmented reality guidance interface
US10726631B1 (en) * 2019-08-03 2020-07-28 VIRNECT inc. Augmented reality system and method with frame region recording and reproduction technology based on object tracking

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20100159833A1 (en) * 2008-12-19 2010-06-24 At&T Mobility Ii Llc Headset Locator Device
US20130010103A1 (en) * 2010-03-14 2013-01-10 Ns Solutions Corporation Information processing system, information processing method and program, information processing apparatus, vacant space guidance system, vacant space guidance method and program, image display system, image display method and program
US20150363915A1 (en) * 2014-06-16 2015-12-17 Huawei Technologies Co., Ltd. Method and Apparatus for Presenting Panoramic Photo in Mobile Terminal, and Mobile Terminal
US20170345209A1 (en) * 2016-05-31 2017-11-30 Matthew Adam Simari Visualization alignment for three-dimensional scanning
US20190073111A1 (en) * 2017-09-05 2019-03-07 Future Mobility Corporation Limited Methods and systems using a rotatable three-dimensional object providing visual user information on a display
US10809081B1 (en) * 2018-05-03 2020-10-20 Zoox, Inc. User interface and augmented reality for identifying vehicles and persons

Also Published As

Publication number Publication date
DE102021201774A1 (en) 2021-09-09
CN113361316A (en) 2021-09-07
US20210279913A1 (en) 2021-09-09
US11263787B2 (en) 2022-03-01

Similar Documents

Publication Publication Date Title
US20250225690A1 (en) Ar display apparatus and ar display method
US9761139B2 (en) Location based parking management system
US20210001841A1 (en) Obstacle Avoidance Method and Apparatus for Autonomous Driving Vehicle
US20210019854A1 (en) Location Signaling with Respect to an Autonomous Vehicle and a Rider
CA2767312C (en) Automatic video surveillance system and method
US11900672B2 (en) Integrated internal and external camera system in vehicles
US20180196417A1 (en) Location Signaling with Respect to an Autonomous Vehicle and a Rider
US6243482B1 (en) Obstacle detection system for low-flying airborne craft
US20120320212A1 (en) Surrounding area monitoring apparatus for vehicle
US11520033B2 (en) Techniques for determining a location of a mobile object
EP3070698B1 (en) Driving-support-image generation device, driving-support-image display device, driving-support-image display system, and driving-support-image generation program
US10996070B2 (en) Route guidance apparatus and method
EP3566022B1 (en) Location signaling with respect to an autonomous vehicle and a rider
US20220092829A1 (en) Augmented reality detection for locating autonomous vehicles
EP4389550A1 (en) Electronic device mounted on vehicle and operation method therefor
US10996469B2 (en) Method and apparatus for providing driving information of vehicle, and recording medium
KR20200084777A (en) System and method for training and operating an autonomous vehicle
FR3079924A1 (en) FOLLOW-UP BY GEOLOCATION
US12238459B2 (en) Rendezvous assistance apparatus, rendezvous assistance system, and rendezvous assistance method
WO2006098311A1 (en) Road scene map creating device, method, and program
WO2019072461A1 (en) Parking assistance method, control device for carrying out the parking assistance method, and vehicle comprising the control device
JP6810723B2 (en) Information processing equipment, information processing methods, and programs
US12151686B2 (en) Information provision system, information terminal, and information provision method
US20240101133A1 (en) Notification device and notification method
US12420822B2 (en) Data transmitter and method for data transmission

Legal Events

Date Code Title Description
STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: RIVIAN AUTOMOTIVE, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HORTOP, MATTHEW;HUNT, PATRICK;TRELFA, LACEY;SIGNING DATES FROM 20211216 TO 20211217;REEL/FRAME:058460/0166

Owner name: RIVIAN IP HOLDINGS, LLC, MICHIGAN

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RIVIAN AUTOMOTIVE, LLC;REEL/FRAME:058461/0028

Effective date: 20211217

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STCB Information on status: application discontinuation

Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION