US20220092829A1 - Augmented reality detection for locating autonomous vehicles - Google Patents
Augmented reality detection for locating autonomous vehicles Download PDFInfo
- Publication number
- US20220092829A1 US20220092829A1 US17/540,586 US202117540586A US2022092829A1 US 20220092829 A1 US20220092829 A1 US 20220092829A1 US 202117540586 A US202117540586 A US 202117540586A US 2022092829 A1 US2022092829 A1 US 2022092829A1
- Authority
- US
- United States
- Prior art keywords
- electronic device
- autonomous vehicle
- identification signal
- indication
- signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01D—MEASURING NOT SPECIALLY ADAPTED FOR A SPECIFIC VARIABLE; ARRANGEMENTS FOR MEASURING TWO OR MORE VARIABLES NOT COVERED IN A SINGLE OTHER SUBCLASS; TARIFF METERING APPARATUS; MEASURING OR TESTING NOT OTHERWISE PROVIDED FOR
- G01D21/00—Measuring or testing not otherwise provided for
- G01D21/02—Measuring two or more variables by means not covered by a single other subclass
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G06K9/00671—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/20—Scenes; Scene-specific elements in augmented reality scenes
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/12—Protocols specially adapted for proprietary or special-purpose networking environments, e.g. medical networks, sensor networks, networks in vehicles or remote metering networks
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/63—Control of cameras or camera modules by using electronic viewfinders
- H04N23/631—Graphical user interfaces [GUI] specially adapted for controlling image capture or setting capture parameters
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N23/00—Cameras or camera modules comprising electronic image sensors; Control thereof
- H04N23/60—Control of cameras or camera modules
- H04N23/64—Computer-aided capture of images, e.g. transfer from script file into camera, check of taken image quality, advice or proposal for image composition or decision on when to take image
-
- H04N5/23222—
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04W—WIRELESS COMMUNICATION NETWORKS
- H04W4/00—Services specially adapted for wireless communication networks; Facilities therefor
- H04W4/30—Services specially adapted for particular environments, situations or purposes
- H04W4/40—Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P]
- H04W4/48—Services specially adapted for particular environments, situations or purposes for vehicles, e.g. vehicle-to-pedestrians [V2P] for in-vehicle communication
-
- G06K2209/23—
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V2201/00—Indexing scheme relating to image or video recognition or understanding
- G06V2201/08—Detecting or categorising vehicles
Definitions
- the present disclosure relates to autonomous vehicles.
- Autonomous vehicles are of great interest for transportation applications and can provide benefits to the ride-sharing industry.
- the intended passenger may have to contact the driver or have human interactions with the driver of the vehicle to identify his/her particular vehicle, especially when in a crowded area such as a city or an airport.
- Such interaction may be challenging, if even possible, when attempting to locate an autonomous vehicle within a pool of autonomous vehicles or a crowded area with obstructions.
- Disclosed embodiments include methods, systems, and non-transitory computer readable medium for augmented reality detection for locating a particular autonomous vehicle.
- an illustrative method includes providing, by an electronic device to a remote server, a request for an autonomous vehicle.
- a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger is received by an electronic device.
- the passenger is prompted by the electronic device to orient an image sensor of the electronic device towards a designated area of the accessible range.
- the designated area is imaged by the image sensor for the autonomous vehicle.
- a visual representation of the imaged designated area is generated on a graphical user interface of the electronic device.
- the autonomous vehicle is identified within the generated visual representation.
- an illustrative system includes a computer processing system and a computer memory.
- the computer processing system is configured to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
- non-transitory computer readable medium includes program instructions for locating autonomous vehicles.
- the program instructions when executed cause a computer processing system to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
- FIG. 1 is a block diagram in partial schematic form of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
- FIG. 2A is a block diagram in partial schematic form of a partial top view of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
- FIG. 2B is a graph of signal strength of detected location signals as a function of image sensor direction of an electronic device for locating a particular autonomous vehicle.
- FIG. 3 is a block diagram in partial schematic form of another illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle.
- FIGS. 4A-4D are top views of illustrative identifications of an autonomous vehicle.
- FIG. 5 is a flow chart of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle.
- first digit of three-digit reference numbers and the first two digits of four-digit reference numbers correspond to the first digit of one-digit figure numbers and the first two digits of two-digit figure numbers, respectively, in which the element first appears.
- FIG. 1 illustrates an illustrative framework 100 for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure.
- a passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle, e.g., from a pool of autonomous vehicles and then use the application on the electronic device 102 to select the proper intended autonomous vehicle from a group of nearby vehicles 104 .
- An autonomous vehicle can be configured with an array of sensors 105 , including, e.g., LIDAR (light detection and ranging), camera systems for acquiring real-time video imagery of obstacles and other vehicles, and GPS (global positioning system).
- LIDAR light detection and ranging
- camera systems for acquiring real-time video imagery of obstacles and other vehicles
- GPS global positioning system
- An autonomous vehicle can also be configured with wireless communication systems for sending and receiving communications regarding road information and traffic in real-time, as well as a computer for applying driving rules and making driving decisions based on data and information acquired from the foregoing, in a suitable manner such as conventionally known in the art.
- Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102 .
- Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet.
- Remote monitoring backend system(s) 108 may select a particular autonomous vehicle, e.g., from a pool of autonomous vehicles based on one or more criteria which may include, but are not limited to, distance to the intended passenger's pick-up location, amount of time it would take to arrive at the intended passenger's pick-up location, destination of the intended passenger, and/or desired vehicle accommodations of the intended passenger.
- Remote monitoring backend system(s) 108 may provide the request to the selected autonomous vehicle via wireless and/or wired communication network 110 .
- Remote monitoring backend system(s) 108 may also communicate with on-board computers of the pool of autonomous vehicles 104 to control various aspects of vehicle features, as will be described further herein.
- Remote monitoring backend system(s) 108 may include a computer processing system 108 , which may include one or more computer processing units (CPUs) 112 located in one or more computer systems, and one or more memories 114 , which may include RAM, ROM, and any suitable nonvolatile memory. Remote monitoring backend system(s) 108 may also include one or more databases 116 to store data for both vehicles and drivers, as will be explained further herein, and one or more interfaces 118 to facilitate communication via networks including the Internet and for facilitating input/output of data. Remote monitoring backend system(s) 108 carries out monitoring of vehicles, vehicle systems, vehicle usage, and systems usage for individual vehicles and for individual drivers who use those vehicles, and carries out data analytics on the data gathered from the vehicles and the vehicle systems.
- CPUs computer processing units
- memories 114 which may include RAM, ROM, and any suitable nonvolatile memory.
- Remote monitoring backend system(s) 108 may also include one or more databases 116 to store data for both vehicles and drivers, as will be explained
- the pool of autonomous vehicles may include various detectors and sensors for monitoring vehicle dynamics, usage, and demand placed on vehicle systems during the driving of the pool of autonomous vehicles.
- the autonomous vehicles may include GPS units for monitoring location, speed, and direction, battery monitoring sensors in the form of electrical circuits, for monitoring battery usage, battery discharge as a function of time, and battery charging rates, and other suitable sensors for monitoring the health and status of other major subsystems of the vehicle, such as removable modules, and for detecting warning or fault conditions.
- Such data may be stored in an onboard computer system with suitable memory and may be uploaded to the remote monitoring backend system(s) 108 through wireless and/or wired communication network 110 via a vehicle onboard wireless transceiver and/or via wired communication.
- Each autonomous vehicle of the pool of autonomous vehicles can emit one or more location signals.
- the one or more location signals of each autonomous vehicle may include rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection.
- the pulses while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger.
- the one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle.
- the one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths.
- Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102 , e.g., via text messaging or other suitable messaging.
- Electronic device 102 can detect the one or more location signals of the autonomous vehicle once it is in an accessible range of the autonomous vehicle.
- the intended passenger can be prompted, via a visual instruction displayed at graphical user interface 106 or via an audio instruction, or both, to orient an image sensor (not shown in FIG. 1 ) of the electronic device 102 toward a designated area once the autonomous vehicle is within an accessible range. After the passenger orients the image sensor to the designated area, the image sensor of electronic device 102 may capture a scene of the pool of autonomous vehicles 104 , as will be described further herein.
- a visual representation of the imaged area may be generated by electronic device 102 using one or more internal processing components.
- the visual representation of the imaged area may be displayed on graphical user interface 106 with the autonomous vehicle identified to the passenger via one or more indications 120 , as will be described further herein.
- the autonomous vehicle may be partially obstructed by one or more obstructions 122 and can still be identified to the intended passenger.
- Access to the autonomous vehicle may be provided to the intended passenger via an access button (field) 122 on graphical user interface 106 , e.g., an unlock graphic field on the touch-sensitive display screen displaying graphical user interface 106 that the intended passenger may press.
- FIG. 2A illustrates a block diagram of a partial top view 200 of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure.
- Electronic device 102 may include an image sensor 202 .
- the image sensor 202 may be sensitive to visible light, non-visible light (e.g., infrared light), or both.
- the image sensor 202 may be a conventional image sensor of the type typically found in smart phones.
- a user having already communicated a request for a vehicle using the app running on the electronic device 102 , may be alerted by the electronic device 102 that an intended vehicle is with detectable range of the user and may direct the use via a visual instruction, an audio instruction, or both, to point the electronic device toward a field of view encompassing one or more vehicles and to scan the image sensor of the electronic device 102 across the field of view by tilting the electronic device 102 laterally as illustrated in FIG. 2A .
- Electronic device 102 may be tilted in one or more axes of a three-dimensional system, e.g., x,y,z-system as illustrated in FIG. 2A . Tilting of electronic device 102 allows image sensor 202 to capture images of the designated area with different orientations whereby different vehicles in the field of view are aligned with different lines of sight (e.g., dotted lines 1 , 2 and 3 ). For example, electronic device 102 may be tilted slightly back and forth to scan and capture the signal strength of the one or more location signals.
- the one or more location signals can be recorded as a function of orientation of the electronic device, e.g., the vector direction of the image sensor 202 of the electronic device 102 and the orientation with the highest line of sight signal strength may be the one that points to the intended autonomous vehicle. That information can then be correlated to the captured image associated with the various orientations to flag the intended autonomous vehicle using an indicator 120 , as will be described further herein.
- Processing of the image information by electronic device 102 may determine an orientation of electronic device 102 which provides a strongest line of sight signal to capture an image of the pool of autonomous vehicles.
- electronic device 102 may have an internal inertial sensor that provides data to a processor of the electronic device 102 .
- the electronic device 102 may prompt the user to slowly scan the image sensor 202 across a field of view containing several vehicles, may evaluate the signal strength of detected location signals as a function of orientation based on correlated inertial data simultaneously captured by the electronic device, and then identify which location in the captured imagery corresponds to the greatest signal strength of the detected location signals, and then display on the electronic device that location superimposed with the imagery of the scene so as to identify to the user the intended autonomous vehicle.
- the electronic device 102 may prompt a passenger, via graphical user interface 106 , to modify the orientation of the electronic device 102 to different orientations to facilitate gathering data for identification of the autonomous vehicle intended for the passenger within view of the image sensor 202 .
- the computer processor of the electronic device 102 is configured to discriminate the detected location signals from the background of conventional imagery signals and to evaluate which orientation (line of sight) of the image sensor of the electronic device 102 correlates to the greatest intensity of the detected location signals.
- existing image sensors typically found in smart-phone cameras may be used to capture the location signals.
- Such smart-phone cameras can detect pulsating light at rates faster than rates the human eye can detect.
- smart-phone cameras can detect pulsating light at rates lower than or greater than 1,000 frames per second, such as 300 frames per second, 600 frames per second, up to 1,000 frames per second or more.
- Electronic device 102 can additionally include suitable filters such as one or more infrared (IR) filters or other color filters.
- the filter(s) may be supported by an actuator to permit movement of the filters relative to the image sensor 202 .
- the filter(s) may be applied to improve image quality captured by image sensor 202 , tune image quality captured by image sensor 202 , and/or select one or more desired wavelengths for selection of location signals of specified wavelengths for capture by the image sensor 202 and processing at the electronic device 102 .
- the electronic device 102 may remove an infrared filter from the light path for image sensor 202 , using for an example, an actuator internally coupled to and controlled by a processor of electronic device 102 .
- the electronic device may insert such filters, e.g., in sequence, into the light path of the image sensor 202 .
- FIG. 2B illustrates a plot 210 of position versus signal strength for tilting of an electronic device for locating a particular autonomous vehicle according to an example of the disclosure.
- plot line 212 illustrates an example curve of signal strength of the detected location signals relative to position.
- directional lines 1 , 2 , 3 of FIG. 2A correlate to positions 1 , 2 , 3 of plot 210 illustrated in FIG. 2B associated with tilting of the electronic device 102 .
- a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b, which is the vehicle emitting the location signals from among the group 104 of nearby autonomous vehicles.
- electronic device 102 is oriented such that the image sensor 202 follows a line of sight along directional line 3 toward vehicle 104 c, a corresponding line of sight signal strength will be minimal as image sensor 202 is not pointing directly at the center autonomous vehicle 104 b emitting the location signals.
- the passenger may be prompted by a spoken or graphical message at the electronic device 102 to point the image sensor of electronic device 102 in different directions to facilitate determination of a stronger line of sight signal strength of detected location signals.
- the stronger line of sight signal in the example of FIG. 2A follows along directional line 2 toward vehicle 104 b.
- Orienting electronic device 102 to have image sensor 202 pointing toward directional line 2 points image sensor 202 at the autonomous vehicle 104 b intended to be paired with the intended passenger by remote monitoring backend system(s) 106 .
- plot line 212 of FIG. 2B the line of sight signal strength following direction line 2 peaks as a result of image sensor being directly oriented to view the center most autonomous vehicle 104 b, in this example.
- the application running on electronic device 102 can capture the signal strength of the location signals via the image sensor and can integrate that signal strength within suitable time intervals.
- image sensor 202 can capture radiation, which includes the location signal, as a function of time, and the processing system of electronic device 102 can record that signal as a function of time and simultaneously record the orientation of electronic device 102 as a function of time.
- the location signal can be of a specified pulse pattern, wavelength pattern, or both, for example, and the processing system of the electronic device can analyze the captured radiation to detect the expected signal pattern, e.g., through use of cross-correlation, for instance, that involves mathematical comparison of the detected radiation with the functional form of the expected signal, or using other mathematical pattern detection techniques known to those of skill in the art.
- That detection can be correlated to the orientation that was recorded as a function of time to select the orientation that corresponds to the strongest amplitude of signal detection.
- the signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle, e.g., 104 b, aligned as the proper source of the expected location signal, such as described in further detail in connection with FIGS. 4A-4D .
- the application running on electronic device 102 can detect the intensity of the location signal (e.g., plot line 212 ) through use of signals collected under the control of one or more suitable wavelength filters controllably placed into the path of light impinging on the image sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of the electronic device 102 is simultaneously captured as a function of time.
- the intensity of the location signal e.g., plot line 212
- suitable wavelength filters controllably placed into the path of light impinging on the image sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of the electronic device 102 is simultaneously captured as a function of time.
- an image of the field of view may be captured under normal imaging, and the application running on electronic device 102 may instruct the user to scan the field of view after actuating one or more desired color filters into the field of view to record the signal strength of one or more given wavelengths as a function of time while the orientation of the electronic device 102 is also recorded as a function in order to correlate the signal strength with orientation of the electronic device 102 .
- the signal strength of the detected signal could also be directly correlated with the orientation during the scanning process.
- That correlation can then be used by the processing system of the electronic device 102 to overlay a graphic at the proper location on the field of view of the interface 106 of the display screen that corresponds to the vehicle aligned as the proper source of the expected location signal, e.g., vehicle 104 b, such as described in further detail in connection with FIGS. 4A-4D .
- FIG. 3 illustrates an illustrative framework 300 for an approach of augmented reality detection for locating a particular autonomous vehicle according to another example of the disclosure.
- This example is similar to the example of FIG. 1 , except that in the example of FIG. 3 , accurate position information for both the autonomous vehicle and the user based at least upon, e.g., global positioning system (GPS) data and map information, provides the primary data on which identification of the intended autonomous vehicle is based.
- GPS global positioning system
- the autonomous vehicle 104 b determines its own position accurately, e.g., within about 1 meter or less, based on one or more of GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle.
- the position of the electronic device 102 of the user is likewise determined accurately, e.g., to within about 1 meter or less, using GPS data and high definition map information accessed by the electronic device 102 , e.g., using conventional approaches known to those of skill in the art.
- the electronic device 102 may determine its absolute orientation with a combination of inertial sensor data and compass technology as conventionally known.
- the electronic device can prompt a user via a visual instruction, audio instruction, or both, to point the electronic device toward a pinpoint location that identifies the intended autonomous vehicle.
- FIG. 3 is similar to FIG. 1 , and the prior discussion relating to FIGS. 1-2 is equally applicable to the example of FIG. 3 . Thus, where applicable, the same reference numerals as in FIG. 1 have been retained for FIG. 3 .
- a passenger may utilize a suitable application installed on an electronic device 102 to request an autonomous vehicle from a pool of autonomous vehicles. Interaction between the passenger and the electronic device 102 may occur via a graphical user interface 106 of the electronic device 102 .
- Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/or wire communication network 110 such as the Internet.
- Remote monitoring backend system(s) 108 is like that described for FIG. 1 .
- the passenger can be alerted to that fact via a wireless communication to the electronic device 102 , e.g., from either the autonomous vehicle itself or from the backend system 108 .
- a passenger can then be prompted by either a visual instruction, audio instruction, or both, at the electronic device 102 to point the electronic device 102 having image sensor 202 toward a designated area where the autonomous vehicle is located.
- the location and orientation of the electronic device can be communicated to the backend system 108 or the autonomous vehicle 104 b, and either of those systems can instruct the user to reorient the electronic device 102 to either the left or the right until the proper line of sight orientation with the pinpoint location of the autonomous vehicle 104 b is achieved.
- electronic device 102 may have functionality such that when it receives the pinpoint location of the intended vehicle via wireless and/or wired communication network 110 , the pinpoint location is correlated to the captured image displayed to the passenger on the graphical user interface 106 of electronic device 102 . With the intended autonomous vehicle correlation to the image on graphical user interface 106 , the autonomous vehicle may be identified to the passenger via an indication, as will be described further herein.
- FIGS. 4A-4D illustrate illustrative indications of an autonomous vehicle according to examples of the disclosure.
- An indication to the intended passenger of the respectively paired autonomous vehicle can be readily customized by the passenger via graphical user interface 106 .
- the indication may be one or more geometric annotation(s) 402 , as illustrated in FIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.”
- the indication may be shaded and/or highlighted with color or gray-scale, e.g., color or gray-scale annotation 404 , as illustrated in FIG. 4B .
- the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C .
- FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle.
- Superimposed image 408 can be the intended passenger's face or a cartoon character.
- FIG. 5 is a flow chart 500 of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle according to examples of the disclosure.
- a request for an autonomous vehicle is provided by electronic device 102 to a remote monitoring backend system 108 , for example, via user input into an application installed on electronic device 102 .
- the electronic device 102 may receive, at 504 , a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger.
- the signal can include a location signal from the autonomous vehicle, which can include, for example, rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection.
- the pulses while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger.
- the one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle.
- the one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths.
- Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the vehicles 104 between such vehicle and the electronic device 102 , e.g., via text messaging or other suitable messaging.
- the intended passenger may be prompted, at 506 , by electronic device 102 to orient an image sensor 202 towards a designated area of the accessible range.
- the intended passenger may be prompted by receiving a text message, a phone call, a pop-up notification, a video call, and/or an audible alarm or voice message on electronic device 102 .
- Image sensor 202 images, at 508 , the designated area for the autonomous vehicle.
- the user may be prompted by the electronic device via any suitable audible or graphical message or indication such as previously described to point the electronic device 102 toward a direction that corresponds to the designated area at which the intended vehicle 104 b is expected, e.g., based on known GPS location information of both the electronic device 102 and the intended vehicle 104 b, and based on directional (orientation) information known at the electronic device 102 via its inertial sensors and conventional compass functionality.
- the processing system can process collected location signal data as a function of orientation such as described in the examples above, to determine the orientation of the electronic device that corresponds to the strongest location signal, and can determine that vehicle 104 b is the source of the location signal and is the intended vehicle 104 b.
- a visual representation of the imaged designated area on the graphical user interface 106 is generated and displayed on the display screen of the electronic device 102 .
- the autonomous vehicle is identified, at 512 , by the electronic device 102 within the generated visual representation, e.g., via a suitable graphical indicator such as shown in FIGS. 4A-4D .
- the indicator may be one or more geometric annotation(s) 402 , as illustrated in FIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.”
- the indication may be shaded and/or highlighted with a color, e.g., color annotation 404 , as illustrated in FIG. 4B .
- the indication may be a geometric annotation 406 such as a circle around the autonomous vehicle, as illustrated in FIG. 4C .
- FIG. 4D illustrates another example indication of a superimposed image 408 over the autonomous vehicle.
- Superimposed image 408 can be the intended passenger's face or a cartoon character.
- the methods and systems described herein may be implemented using any suitable computer processing system with any suitable combination of hardware, software and/or firmware.
- entities that control the pool of autonomous vehicles 104 can interact with remote monitoring backend system 106 hosted on one or more computer servers through a wireless and/or wired communication network 110 .
- the computer processing system 112 may execute software operations, program instructions or routines to implement calculations and analyses described above.
- Such program instructions, accumulated data, and processed data may be stored on one or more non-transitory computer-readable memories 114 and/or one or more data stores for use in databases 116 .
- Communications may be carried out according to a client server architecture whereby passengers and autonomous vehicles 104 may access the remote monitoring backend system 108 via one or more servers via one or more wireless and/or wired communication networks 110 .
- the systems may include element managers, real-time data buffers, conveyors, file input processors, database indices, data buffers and data managers for managing data and processing.
- the remote monitoring backend system(s) 108 may also include multiple displays, display interfaces, input/output devices such as keyboards, microphones, mice, touch screens and the like for permitting administrators or support personnel to manage the remote monitoring backend system(s) 108 .
- the systems and methods may include and utilize data signals conveyed via networks (e.g., local area network, wide area network, internet, combinations thereof, etc.), fiber optic medium, carrier waves, wireless networks, etc. for communication with one or more data processing devices.
- the data signals can carry any or all of the data disclosed herein that is provided to or from a device.
- the methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing system.
- the software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein. Any suitable computer languages may be used such as C, C++, Java, etc., as will be appreciated by those skilled in the art. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to carry out the methods and systems described herein.
- the systems' and methods' data may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.).
- storage devices and programming constructs e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.
- data structures describe formats for use in organizing and storing data in databases, programs, memory, or other non-transitory computer-readable media for use by a computer program.
- a module or processor includes but is not limited to a unit of code that performs a software operation, and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code.
- the software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Signal Processing (AREA)
- Multimedia (AREA)
- Human Computer Interaction (AREA)
- General Engineering & Computer Science (AREA)
- Computer Networks & Wireless Communication (AREA)
- Health & Medical Sciences (AREA)
- Computing Systems (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Traffic Control Systems (AREA)
- Navigation (AREA)
Abstract
Description
- The present disclosure is a continuation (CON) of co-pending U.S. patent application Ser. No. 16/810,652, filed on Mar. 5, 2020, and entitled “AUGMENTED REALITY DETECTION FOR LOCATING AUTONOMOUS VEHICLES,” the contents of which are incorporated in full by reference herein for all purposes.
- The present disclosure relates to autonomous vehicles.
- The statements in this section merely provide background information related to the present disclosure and may not constitute prior art.
- Autonomous vehicles are of great interest for transportation applications and can provide benefits to the ride-sharing industry. Conventionally, when an intended passenger makes a request for a ride-sharing vehicle, the intended passenger may have to contact the driver or have human interactions with the driver of the vehicle to identify his/her particular vehicle, especially when in a crowded area such as a city or an airport. Such interaction, however, may be challenging, if even possible, when attempting to locate an autonomous vehicle within a pool of autonomous vehicles or a crowded area with obstructions.
- Disclosed embodiments include methods, systems, and non-transitory computer readable medium for augmented reality detection for locating a particular autonomous vehicle.
- In a non-limiting embodiment, an illustrative method includes providing, by an electronic device to a remote server, a request for an autonomous vehicle. A first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger is received by an electronic device. The passenger is prompted by the electronic device to orient an image sensor of the electronic device towards a designated area of the accessible range. The designated area is imaged by the image sensor for the autonomous vehicle. A visual representation of the imaged designated area is generated on a graphical user interface of the electronic device. The autonomous vehicle is identified within the generated visual representation.
- In another non-limiting embodiment, an illustrative system includes a computer processing system and a computer memory. The computer processing system is configured to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
- In another non-limiting embodiment, non-transitory computer readable medium includes program instructions for locating autonomous vehicles. The program instructions when executed cause a computer processing system to execute steps including: providing, by an electronic device to a remote server, a request for an autonomous vehicle; receiving, by the electronic device, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger; prompting, by the electronic device, the passenger to orient an image sensor of the electronic device towards a designated area of the accessible range; imaging, by the image sensor, the designated area for the autonomous vehicle; generating, by the electronic device, a visual representation of the imaged designated area on a graphical user interface of the electronic device; and identifying, by the electronic device, the autonomous vehicle within the generated visual representation.
- Further features, advantages, and areas of applicability will become apparent from the description provided herein. It should be understood that the description and specific examples are intended for purposes of illustration only and are not intended to limit the scope of the present disclosure.
- The drawings described herein are for illustration purposes only and are not intended to limit the scope of the present disclosure in any way. The components in the figures are not necessarily to scale, with emphasis instead being placed upon illustrating the principles of the disclosed embodiments. In the drawings:
-
FIG. 1 is a block diagram in partial schematic form of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle. -
FIG. 2A is a block diagram in partial schematic form of a partial top view of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle. -
FIG. 2B is a graph of signal strength of detected location signals as a function of image sensor direction of an electronic device for locating a particular autonomous vehicle. -
FIG. 3 is a block diagram in partial schematic form of another illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle. -
FIGS. 4A-4D are top views of illustrative identifications of an autonomous vehicle. -
FIG. 5 is a flow chart of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle. - Like reference symbols in the various drawings generally indicate like elements.
- The following description is merely illustrative in nature and is not intended to limit the present disclosure, application, or uses. It will be noted that the first digit of three-digit reference numbers and the first two digits of four-digit reference numbers correspond to the first digit of one-digit figure numbers and the first two digits of two-digit figure numbers, respectively, in which the element first appears.
-
FIG. 1 illustrates anillustrative framework 100 for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure. A passenger may utilize a suitable application installed on anelectronic device 102 to request an autonomous vehicle, e.g., from a pool of autonomous vehicles and then use the application on theelectronic device 102 to select the proper intended autonomous vehicle from a group ofnearby vehicles 104. An autonomous vehicle can be configured with an array ofsensors 105, including, e.g., LIDAR (light detection and ranging), camera systems for acquiring real-time video imagery of obstacles and other vehicles, and GPS (global positioning system). An autonomous vehicle can also be configured with wireless communication systems for sending and receiving communications regarding road information and traffic in real-time, as well as a computer for applying driving rules and making driving decisions based on data and information acquired from the foregoing, in a suitable manner such as conventionally known in the art. Interaction between the passenger and theelectronic device 102 may occur via agraphical user interface 106 of theelectronic device 102.Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/orwire communication network 110 such as the Internet. Remote monitoring backend system(s) 108 may select a particular autonomous vehicle, e.g., from a pool of autonomous vehicles based on one or more criteria which may include, but are not limited to, distance to the intended passenger's pick-up location, amount of time it would take to arrive at the intended passenger's pick-up location, destination of the intended passenger, and/or desired vehicle accommodations of the intended passenger. Remote monitoring backend system(s) 108 may provide the request to the selected autonomous vehicle via wireless and/orwired communication network 110. Remote monitoring backend system(s) 108 may also communicate with on-board computers of the pool ofautonomous vehicles 104 to control various aspects of vehicle features, as will be described further herein. - Remote monitoring backend system(s) 108 may include a
computer processing system 108, which may include one or more computer processing units (CPUs) 112 located in one or more computer systems, and one ormore memories 114, which may include RAM, ROM, and any suitable nonvolatile memory. Remote monitoring backend system(s) 108 may also include one ormore databases 116 to store data for both vehicles and drivers, as will be explained further herein, and one ormore interfaces 118 to facilitate communication via networks including the Internet and for facilitating input/output of data. Remote monitoring backend system(s) 108 carries out monitoring of vehicles, vehicle systems, vehicle usage, and systems usage for individual vehicles and for individual drivers who use those vehicles, and carries out data analytics on the data gathered from the vehicles and the vehicle systems. In examples, the pool of autonomous vehicles may include various detectors and sensors for monitoring vehicle dynamics, usage, and demand placed on vehicle systems during the driving of the pool of autonomous vehicles. For example, the autonomous vehicles may include GPS units for monitoring location, speed, and direction, battery monitoring sensors in the form of electrical circuits, for monitoring battery usage, battery discharge as a function of time, and battery charging rates, and other suitable sensors for monitoring the health and status of other major subsystems of the vehicle, such as removable modules, and for detecting warning or fault conditions. Such data may be stored in an onboard computer system with suitable memory and may be uploaded to the remote monitoring backend system(s) 108 through wireless and/orwired communication network 110 via a vehicle onboard wireless transceiver and/or via wired communication. - Each autonomous vehicle of the pool of autonomous vehicles can emit one or more location signals. For example, the one or more location signals of each autonomous vehicle may include rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection. The pulses, while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger. The one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle. The one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths. Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of the
vehicles 104 between such vehicle and theelectronic device 102, e.g., via text messaging or other suitable messaging.Electronic device 102 can detect the one or more location signals of the autonomous vehicle once it is in an accessible range of the autonomous vehicle. The intended passenger can be prompted, via a visual instruction displayed atgraphical user interface 106 or via an audio instruction, or both, to orient an image sensor (not shown inFIG. 1 ) of theelectronic device 102 toward a designated area once the autonomous vehicle is within an accessible range. After the passenger orients the image sensor to the designated area, the image sensor ofelectronic device 102 may capture a scene of the pool ofautonomous vehicles 104, as will be described further herein. A visual representation of the imaged area may be generated byelectronic device 102 using one or more internal processing components. The visual representation of the imaged area may be displayed ongraphical user interface 106 with the autonomous vehicle identified to the passenger via one ormore indications 120, as will be described further herein. In some cases, the autonomous vehicle may be partially obstructed by one ormore obstructions 122 and can still be identified to the intended passenger. Access to the autonomous vehicle may be provided to the intended passenger via an access button (field) 122 ongraphical user interface 106, e.g., an unlock graphic field on the touch-sensitive display screen displayinggraphical user interface 106 that the intended passenger may press. -
FIG. 2A illustrates a block diagram of a partialtop view 200 of an illustrative framework for an approach of augmented reality detection for locating a particular autonomous vehicle according to an example of the disclosure.Electronic device 102 may include animage sensor 202. Theimage sensor 202 may be sensitive to visible light, non-visible light (e.g., infrared light), or both. Theimage sensor 202 may be a conventional image sensor of the type typically found in smart phones. In an example, a user, having already communicated a request for a vehicle using the app running on theelectronic device 102, may be alerted by theelectronic device 102 that an intended vehicle is with detectable range of the user and may direct the use via a visual instruction, an audio instruction, or both, to point the electronic device toward a field of view encompassing one or more vehicles and to scan the image sensor of theelectronic device 102 across the field of view by tilting theelectronic device 102 laterally as illustrated inFIG. 2A . -
Electronic device 102 may be tilted in one or more axes of a three-dimensional system, e.g., x,y,z-system as illustrated inFIG. 2A . Tilting ofelectronic device 102 allowsimage sensor 202 to capture images of the designated area with different orientations whereby different vehicles in the field of view are aligned with different lines of sight (e.g., dotted 1, 2 and 3). For example,lines electronic device 102 may be tilted slightly back and forth to scan and capture the signal strength of the one or more location signals. The one or more location signals can be recorded as a function of orientation of the electronic device, e.g., the vector direction of theimage sensor 202 of theelectronic device 102 and the orientation with the highest line of sight signal strength may be the one that points to the intended autonomous vehicle. That information can then be correlated to the captured image associated with the various orientations to flag the intended autonomous vehicle using anindicator 120, as will be described further herein. - Processing of the image information by
electronic device 102 may determine an orientation ofelectronic device 102 which provides a strongest line of sight signal to capture an image of the pool of autonomous vehicles. For example,electronic device 102 may have an internal inertial sensor that provides data to a processor of theelectronic device 102. As noted above, theelectronic device 102 may prompt the user to slowly scan theimage sensor 202 across a field of view containing several vehicles, may evaluate the signal strength of detected location signals as a function of orientation based on correlated inertial data simultaneously captured by the electronic device, and then identify which location in the captured imagery corresponds to the greatest signal strength of the detected location signals, and then display on the electronic device that location superimposed with the imagery of the scene so as to identify to the user the intended autonomous vehicle. In another example, theelectronic device 102 may prompt a passenger, viagraphical user interface 106, to modify the orientation of theelectronic device 102 to different orientations to facilitate gathering data for identification of the autonomous vehicle intended for the passenger within view of theimage sensor 202. Although theimage sensor 202 of theelectronic device 102 is exposed to conventional image signals during this process, the computer processor of theelectronic device 102 is configured to discriminate the detected location signals from the background of conventional imagery signals and to evaluate which orientation (line of sight) of the image sensor of theelectronic device 102 correlates to the greatest intensity of the detected location signals. - In this regard, existing image sensors typically found in smart-phone cameras may be used to capture the location signals. Such smart-phone cameras can detect pulsating light at rates faster than rates the human eye can detect. For example, smart-phone cameras can detect pulsating light at rates lower than or greater than 1,000 frames per second, such as 300 frames per second, 600 frames per second, up to 1,000 frames per second or more.
Electronic device 102 can additionally include suitable filters such as one or more infrared (IR) filters or other color filters. The filter(s) may be supported by an actuator to permit movement of the filters relative to theimage sensor 202. The filter(s) may be applied to improve image quality captured byimage sensor 202, tune image quality captured byimage sensor 202, and/or select one or more desired wavelengths for selection of location signals of specified wavelengths for capture by theimage sensor 202 and processing at theelectronic device 102. When the location signals are within an infrared light spectrum, for example, theelectronic device 102 may remove an infrared filter from the light path forimage sensor 202, using for an example, an actuator internally coupled to and controlled by a processor ofelectronic device 102. When the location signals comprise specified color wavelengths, e.g., a progression of signal pulses of one or more wavelengths, e.g., in sequence, the electronic device may insert such filters, e.g., in sequence, into the light path of theimage sensor 202. -
FIG. 2B illustrates a plot 210 of position versus signal strength for tilting of an electronic device for locating a particular autonomous vehicle according to an example of the disclosure. For illustration purposes,plot line 212 illustrates an example curve of signal strength of the detected location signals relative to position. In an example, 1, 2, 3 ofdirectional lines FIG. 2A correlate to 1, 2, 3 of plot 210 illustrated inpositions FIG. 2B associated with tilting of theelectronic device 102. Whenelectronic device 102 is oriented such that theimage sensor 202 follows a line of sight alongdirectional line 1 towardvehicle 104 a ofFIG. 2A , a corresponding line of sight signal strength will be minimal asimage sensor 202 is not pointing directly at the centerautonomous vehicle 104 b, which is the vehicle emitting the location signals from among thegroup 104 of nearby autonomous vehicles. Similarly, whenelectronic device 102 is oriented such that theimage sensor 202 follows a line of sight alongdirectional line 3 towardvehicle 104 c, a corresponding line of sight signal strength will be minimal asimage sensor 202 is not pointing directly at the centerautonomous vehicle 104 b emitting the location signals. In an example, the passenger may be prompted by a spoken or graphical message at theelectronic device 102 to point the image sensor ofelectronic device 102 in different directions to facilitate determination of a stronger line of sight signal strength of detected location signals. The stronger line of sight signal in the example ofFIG. 2A , follows alongdirectional line 2 towardvehicle 104 b. Orientingelectronic device 102 to haveimage sensor 202 pointing towarddirectional line 2points image sensor 202 at theautonomous vehicle 104 b intended to be paired with the intended passenger by remote monitoring backend system(s) 106. As illustrated byplot line 212 ofFIG. 2B , the line of sight signal strength followingdirection line 2 peaks as a result of image sensor being directly oriented to view the center mostautonomous vehicle 104 b, in this example. In order to generateplot line 212, the application running onelectronic device 102 can capture the signal strength of the location signals via the image sensor and can integrate that signal strength within suitable time intervals. For example,image sensor 202 can capture radiation, which includes the location signal, as a function of time, and the processing system ofelectronic device 102 can record that signal as a function of time and simultaneously record the orientation ofelectronic device 102 as a function of time. As noted previously, the location signal can be of a specified pulse pattern, wavelength pattern, or both, for example, and the processing system of the electronic device can analyze the captured radiation to detect the expected signal pattern, e.g., through use of cross-correlation, for instance, that involves mathematical comparison of the detected radiation with the functional form of the expected signal, or using other mathematical pattern detection techniques known to those of skill in the art. With the timing of the location signal having been detected, that detection can be correlated to the orientation that was recorded as a function of time to select the orientation that corresponds to the strongest amplitude of signal detection. The signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used to overlay a graphic at the proper location on the field of view of theinterface 106 of the display screen that corresponds to the vehicle, e.g., 104 b, aligned as the proper source of the expected location signal, such as described in further detail in connection withFIGS. 4A-4D . - In another example, the application running on
electronic device 102 can detect the intensity of the location signal (e.g., plot line 212) through use of signals collected under the control of one or more suitable wavelength filters controllably placed into the path of light impinging on theimage sensor 202 so as to permit the capture desired wavelengths that make up the location signal as a function of time while the orientation of theelectronic device 102 is simultaneously captured as a function of time. In this approach, for example, an image of the field of view may be captured under normal imaging, and the application running onelectronic device 102 may instruct the user to scan the field of view after actuating one or more desired color filters into the field of view to record the signal strength of one or more given wavelengths as a function of time while the orientation of theelectronic device 102 is also recorded as a function in order to correlate the signal strength with orientation of theelectronic device 102. The signal strength of the detected signal could also be directly correlated with the orientation during the scanning process. That correlation can then be used by the processing system of theelectronic device 102 to overlay a graphic at the proper location on the field of view of theinterface 106 of the display screen that corresponds to the vehicle aligned as the proper source of the expected location signal, e.g.,vehicle 104 b, such as described in further detail in connection withFIGS. 4A-4D . -
FIG. 3 illustrates an illustrative framework 300 for an approach of augmented reality detection for locating a particular autonomous vehicle according to another example of the disclosure. This example is similar to the example ofFIG. 1 , except that in the example ofFIG. 3 , accurate position information for both the autonomous vehicle and the user based at least upon, e.g., global positioning system (GPS) data and map information, provides the primary data on which identification of the intended autonomous vehicle is based. For example, theautonomous vehicle 104 b determines its own position accurately, e.g., within about 1 meter or less, based on one or more of GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle. Conventional approaches known in the art for identifying accurate positions of autonomous vehicles utilizing GPS data, high definition maps loaded on the autonomous vehicle, and LIDAR attached to the autonomous vehicle may be utilized in this regard. Similarly, the position of theelectronic device 102 of the user is likewise determined accurately, e.g., to within about 1 meter or less, using GPS data and high definition map information accessed by theelectronic device 102, e.g., using conventional approaches known to those of skill in the art. Moreover, theelectronic device 102 may determine its absolute orientation with a combination of inertial sensor data and compass technology as conventionally known. With this information, as explained below, the electronic device can prompt a user via a visual instruction, audio instruction, or both, to point the electronic device toward a pinpoint location that identifies the intended autonomous vehicle. In other respects,FIG. 3 is similar toFIG. 1 , and the prior discussion relating toFIGS. 1-2 is equally applicable to the example ofFIG. 3 . Thus, where applicable, the same reference numerals as inFIG. 1 have been retained forFIG. 3 . - As shown in
FIG. 3 , similar toFIG. 1 , a passenger may utilize a suitable application installed on anelectronic device 102 to request an autonomous vehicle from a pool of autonomous vehicles. Interaction between the passenger and theelectronic device 102 may occur via agraphical user interface 106 of theelectronic device 102.Electronic device 102 may transmit the request to one or more remote monitoring backend system(s) 108 via wireless and/orwire communication network 110 such as the Internet. Remote monitoring backend system(s) 108 is like that described forFIG. 1 . - When the
autonomous vehicle 104 b has arrived within a desired proximity of the location of the intended passenger, the passenger can be alerted to that fact via a wireless communication to theelectronic device 102, e.g., from either the autonomous vehicle itself or from thebackend system 108. A passenger can then be prompted by either a visual instruction, audio instruction, or both, at theelectronic device 102 to point theelectronic device 102 havingimage sensor 202 toward a designated area where the autonomous vehicle is located. For instance, the location and orientation of the electronic device can be communicated to thebackend system 108 or theautonomous vehicle 104 b, and either of those systems can instruct the user to reorient theelectronic device 102 to either the left or the right until the proper line of sight orientation with the pinpoint location of theautonomous vehicle 104 b is achieved. Additionally,electronic device 102 may have functionality such that when it receives the pinpoint location of the intended vehicle via wireless and/orwired communication network 110, the pinpoint location is correlated to the captured image displayed to the passenger on thegraphical user interface 106 ofelectronic device 102. With the intended autonomous vehicle correlation to the image ongraphical user interface 106, the autonomous vehicle may be identified to the passenger via an indication, as will be described further herein. -
FIGS. 4A-4D illustrate illustrative indications of an autonomous vehicle according to examples of the disclosure. An indication to the intended passenger of the respectively paired autonomous vehicle can be readily customized by the passenger viagraphical user interface 106. For example, the indication may be one or more geometric annotation(s) 402, as illustrated inFIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.” In another example, the indication may be shaded and/or highlighted with color or gray-scale, e.g., color or gray-scale annotation 404, as illustrated inFIG. 4B . In yet another example, the indication may be ageometric annotation 406 such as a circle around the autonomous vehicle, as illustrated inFIG. 4C .FIG. 4D illustrates another example indication of asuperimposed image 408 over the autonomous vehicle.Superimposed image 408 can be the intended passenger's face or a cartoon character. -
FIG. 5 is aflow chart 500 of an illustrative approach of augmented reality detection for locating a particular autonomous vehicle according to examples of the disclosure. As shown at 502, a request for an autonomous vehicle is provided byelectronic device 102 to a remotemonitoring backend system 108, for example, via user input into an application installed onelectronic device 102. Theelectronic device 102 may receive, at 504, a first signal indicating that the autonomous vehicle is within an accessible range to an intended passenger. The signal can include a location signal from the autonomous vehicle, which can include, for example, rapidly pulsing exterior vehicle lights at a specified frequency or in a specified pattern intended for detection. The pulses, while of the visible light spectrum, may be pulsating at a frequency too high to be visually perceived as pulsating or flickering by the passenger. The one or more location signals of each autonomous vehicle may also be pulses of nonvisible spectrum light, such as laser light emitted from a LIDAR device affixed to the autonomous vehicle. The one or more location signals of each autonomous vehicle may be made up of a spectrum shift of light colors, e.g., a pulsed pattern involving different wavelengths. Communication using Wi-Fi and/or Bluetooth signals may also be used for communicating additional identification information of the intended one of thevehicles 104 between such vehicle and theelectronic device 102, e.g., via text messaging or other suitable messaging. The intended passenger may be prompted, at 506, byelectronic device 102 to orient animage sensor 202 towards a designated area of the accessible range. For example, the intended passenger may be prompted by receiving a text message, a phone call, a pop-up notification, a video call, and/or an audible alarm or voice message onelectronic device 102.Image sensor 202 images, at 508, the designated area for the autonomous vehicle. For example, the user may be prompted by the electronic device via any suitable audible or graphical message or indication such as previously described to point theelectronic device 102 toward a direction that corresponds to the designated area at which the intendedvehicle 104 b is expected, e.g., based on known GPS location information of both theelectronic device 102 and the intendedvehicle 104 b, and based on directional (orientation) information known at theelectronic device 102 via its inertial sensors and conventional compass functionality. The processing system can process collected location signal data as a function of orientation such as described in the examples above, to determine the orientation of the electronic device that corresponds to the strongest location signal, and can determine thatvehicle 104 b is the source of the location signal and is the intendedvehicle 104 b. At 510, a visual representation of the imaged designated area on thegraphical user interface 106 is generated and displayed on the display screen of theelectronic device 102. The autonomous vehicle is identified, at 512, by theelectronic device 102 within the generated visual representation, e.g., via a suitable graphical indicator such as shown inFIGS. 4A-4D . For example, the indicator may be one or more geometric annotation(s) 402, as illustrated inFIG. 4A , such as an arrow accompanied by text stating: “YOUR RIDE.” In another example, the indication may be shaded and/or highlighted with a color, e.g.,color annotation 404, as illustrated inFIG. 4B . In yet another example, the indication may be ageometric annotation 406 such as a circle around the autonomous vehicle, as illustrated inFIG. 4C .FIG. 4D illustrates another example indication of asuperimposed image 408 over the autonomous vehicle.Superimposed image 408 can be the intended passenger's face or a cartoon character. - The methods and systems described herein may be implemented using any suitable computer processing system with any suitable combination of hardware, software and/or firmware. As shown in
FIGS. 1 and 3 , entities that control the pool ofautonomous vehicles 104 can interact with remotemonitoring backend system 106 hosted on one or more computer servers through a wireless and/orwired communication network 110. Thecomputer processing system 112 may execute software operations, program instructions or routines to implement calculations and analyses described above. Such program instructions, accumulated data, and processed data may be stored on one or more non-transitory computer-readable memories 114 and/or one or more data stores for use indatabases 116. Communications may be carried out according to a client server architecture whereby passengers andautonomous vehicles 104 may access the remotemonitoring backend system 108 via one or more servers via one or more wireless and/orwired communication networks 110. - The systems may include element managers, real-time data buffers, conveyors, file input processors, database indices, data buffers and data managers for managing data and processing. The remote monitoring backend system(s) 108 may also include multiple displays, display interfaces, input/output devices such as keyboards, microphones, mice, touch screens and the like for permitting administrators or support personnel to manage the remote monitoring backend system(s) 108.
- This written description describes illustrative embodiments, but other variations fall within the scope of the disclosure. For example, the systems and methods may include and utilize data signals conveyed via networks (e.g., local area network, wide area network, internet, combinations thereof, etc.), fiber optic medium, carrier waves, wireless networks, etc. for communication with one or more data processing devices. The data signals can carry any or all of the data disclosed herein that is provided to or from a device.
- The methods and systems described herein may be implemented on many different types of processing devices by program code comprising program instructions that are executable by the device processing system. The software program instructions may include source code, object code, machine code, or any other stored data that is operable to cause a processing system to perform the methods and operations described herein. Any suitable computer languages may be used such as C, C++, Java, etc., as will be appreciated by those skilled in the art. Other implementations may also be used, however, such as firmware or even appropriately designed hardware configured to carry out the methods and systems described herein.
- The systems' and methods' data (e.g., associations, mappings, data input, data output, intermediate data results, final data results, etc.) may be stored and implemented in one or more different types of computer-implemented data stores, such as different types of storage devices and programming constructs (e.g., RAM, ROM, Flash memory, flat files, databases, programming data structures, programming variables, IF-THEN (or similar type) statement constructs, etc.). It is noted that data structures describe formats for use in organizing and storing data in databases, programs, memory, or other non-transitory computer-readable media for use by a computer program.
- The computer components, software modules, functions, data stores and data structures described herein may be connected directly or indirectly to each other in order to allow the flow of data needed for their operations. It is also noted that a module or processor includes but is not limited to a unit of code that performs a software operation, and can be implemented for example as a subroutine unit of code, or as a software function unit of code, or as an object (as in an object-oriented paradigm), or as an applet, or in a computer script language, or as another type of computer code. The software components and/or functionality may be located on a single computer or distributed across multiple computers depending upon the situation at hand.
- It should be understood that as used in the description herein and throughout the claims that follow, the meaning of “a,” “an,” and “the” includes plural reference unless the context clearly dictates otherwise. Also, as used in the description herein and throughout the claims that follow, the meaning of “in” includes “in” and “on” unless the context clearly dictates otherwise. In addition, as used in the description herein and throughout the claims that follow, the meanings of “and” and “or” include both the conjunctive and disjunctive and may be used interchangeably unless the context expressly dictates otherwise; the phrase “exclusive or” may be used to indicate situation where only the disjunctive meaning may apply. In addition, as used in the description herein and throughout the claims that follow, the meaning of “about” and/or “approximately” refers to ±10% of the quantity indicated, unless otherwise indicated.
- While the disclosed subject matter has been described in terms of illustrative embodiments, it will be understood by those skilled in the art that various modifications can be made thereto without departing from the scope of the claimed subject matter as set forth in the claims.
Claims (20)
Priority Applications (1)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US17/540,586 US20220092829A1 (en) | 2020-03-05 | 2021-12-02 | Augmented reality detection for locating autonomous vehicles |
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US16/810,652 US11263787B2 (en) | 2020-03-05 | 2020-03-05 | Augmented reality detection for locating autonomous vehicles |
| US17/540,586 US20220092829A1 (en) | 2020-03-05 | 2021-12-02 | Augmented reality detection for locating autonomous vehicles |
Related Parent Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/810,652 Continuation US11263787B2 (en) | 2020-03-05 | 2020-03-05 | Augmented reality detection for locating autonomous vehicles |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20220092829A1 true US20220092829A1 (en) | 2022-03-24 |
Family
ID=77388804
Family Applications (2)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/810,652 Active US11263787B2 (en) | 2020-03-05 | 2020-03-05 | Augmented reality detection for locating autonomous vehicles |
| US17/540,586 Abandoned US20220092829A1 (en) | 2020-03-05 | 2021-12-02 | Augmented reality detection for locating autonomous vehicles |
Family Applications Before (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US16/810,652 Active US11263787B2 (en) | 2020-03-05 | 2020-03-05 | Augmented reality detection for locating autonomous vehicles |
Country Status (3)
| Country | Link |
|---|---|
| US (2) | US11263787B2 (en) |
| CN (1) | CN113361316A (en) |
| DE (1) | DE102021201774A1 (en) |
Families Citing this family (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US10471611B2 (en) | 2016-01-15 | 2019-11-12 | Irobot Corporation | Autonomous monitoring robot systems |
| US20240061129A1 (en) * | 2022-08-18 | 2024-02-22 | Textron Innovations Inc. | Updating geofencing data in a utility vehicle control device |
Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100159833A1 (en) * | 2008-12-19 | 2010-06-24 | At&T Mobility Ii Llc | Headset Locator Device |
| US20130010103A1 (en) * | 2010-03-14 | 2013-01-10 | Ns Solutions Corporation | Information processing system, information processing method and program, information processing apparatus, vacant space guidance system, vacant space guidance method and program, image display system, image display method and program |
| US20150363915A1 (en) * | 2014-06-16 | 2015-12-17 | Huawei Technologies Co., Ltd. | Method and Apparatus for Presenting Panoramic Photo in Mobile Terminal, and Mobile Terminal |
| US20170345209A1 (en) * | 2016-05-31 | 2017-11-30 | Matthew Adam Simari | Visualization alignment for three-dimensional scanning |
| US20190073111A1 (en) * | 2017-09-05 | 2019-03-07 | Future Mobility Corporation Limited | Methods and systems using a rotatable three-dimensional object providing visual user information on a display |
| US10809081B1 (en) * | 2018-05-03 | 2020-10-20 | Zoox, Inc. | User interface and augmented reality for identifying vehicles and persons |
Family Cites Families (23)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US7720436B2 (en) * | 2006-01-09 | 2010-05-18 | Nokia Corporation | Displaying network objects in mobile devices based on geolocation |
| DE102009037835B4 (en) * | 2009-08-18 | 2012-12-06 | Metaio Gmbh | Method for displaying virtual information in a real environment |
| JP6065370B2 (en) * | 2012-02-03 | 2017-01-25 | ソニー株式会社 | Information processing apparatus, information processing method, and program |
| US9852381B2 (en) * | 2012-12-20 | 2017-12-26 | Nokia Technologies Oy | Method and apparatus for providing behavioral pattern generation for mixed reality objects |
| US9959674B2 (en) * | 2013-02-26 | 2018-05-01 | Qualcomm Incorporated | Directional and X-ray view techniques for navigation using a mobile device |
| US9041741B2 (en) * | 2013-03-14 | 2015-05-26 | Qualcomm Incorporated | User interface for a head mounted display |
| EP2983138A4 (en) * | 2013-04-04 | 2017-02-22 | Sony Corporation | Display control device, display control method and program |
| US10217284B2 (en) * | 2013-09-30 | 2019-02-26 | Qualcomm Incorporated | Augmented virtuality |
| US10802582B1 (en) * | 2014-04-22 | 2020-10-13 | sigmund lindsay clements | Eye tracker in an augmented reality glasses for eye gaze to input displayed input icons |
| US20160140868A1 (en) * | 2014-11-13 | 2016-05-19 | Netapp, Inc. | Techniques for using augmented reality for computer systems maintenance |
| US20160358383A1 (en) * | 2015-06-05 | 2016-12-08 | Steffen Gauglitz | Systems and methods for augmented reality-based remote collaboration |
| KR102355455B1 (en) * | 2016-06-20 | 2022-01-24 | 매직 립, 인코포레이티드 | Augmented Reality Display System for Evaluation and Modification of Neurological Conditions Including Visual Processing and Perceptual States |
| US10078916B2 (en) * | 2016-07-01 | 2018-09-18 | Invia Robotics, Inc. | Pick to augmented reality |
| EP3321892A1 (en) * | 2016-11-10 | 2018-05-16 | Gemalto Sa | A physical key for provisioning a communication device with data allowing it to access a vehicle resource |
| US9841814B1 (en) * | 2017-01-27 | 2017-12-12 | Emergent AR Platforms Corp. | Intentional user experience |
| CA3054282A1 (en) * | 2017-03-02 | 2018-09-07 | Motorola Solutions, Inc. | Method and apparatus for gathering visual data using an augmented-reality application |
| US11120264B2 (en) | 2017-06-02 | 2021-09-14 | Apple Inc. | Augmented reality interface for facilitating identification of arriving vehicle |
| US11173798B2 (en) * | 2018-02-12 | 2021-11-16 | Ford Global Technologies, Llc | Systems and methods for vehicle wireless charging |
| US10546169B2 (en) * | 2018-03-22 | 2020-01-28 | Hall Labs Llc | Augmented reality navigation system |
| US10318811B1 (en) * | 2018-04-22 | 2019-06-11 | Bubbler International Llc | Methods and systems for detecting objects by non-visible radio frequencies and displaying associated augmented reality effects |
| KR102627612B1 (en) * | 2019-02-19 | 2024-01-22 | 삼성전자주식회사 | Method for displaying nerby information using augmented reality and electonic device therof |
| US11598643B2 (en) * | 2019-03-26 | 2023-03-07 | Snap Inc. | Augmented reality guidance interface |
| US10726631B1 (en) * | 2019-08-03 | 2020-07-28 | VIRNECT inc. | Augmented reality system and method with frame region recording and reproduction technology based on object tracking |
-
2020
- 2020-03-05 US US16/810,652 patent/US11263787B2/en active Active
-
2021
- 2021-02-25 DE DE102021201774.6A patent/DE102021201774A1/en active Pending
- 2021-02-26 CN CN202110216545.6A patent/CN113361316A/en active Pending
- 2021-12-02 US US17/540,586 patent/US20220092829A1/en not_active Abandoned
Patent Citations (6)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20100159833A1 (en) * | 2008-12-19 | 2010-06-24 | At&T Mobility Ii Llc | Headset Locator Device |
| US20130010103A1 (en) * | 2010-03-14 | 2013-01-10 | Ns Solutions Corporation | Information processing system, information processing method and program, information processing apparatus, vacant space guidance system, vacant space guidance method and program, image display system, image display method and program |
| US20150363915A1 (en) * | 2014-06-16 | 2015-12-17 | Huawei Technologies Co., Ltd. | Method and Apparatus for Presenting Panoramic Photo in Mobile Terminal, and Mobile Terminal |
| US20170345209A1 (en) * | 2016-05-31 | 2017-11-30 | Matthew Adam Simari | Visualization alignment for three-dimensional scanning |
| US20190073111A1 (en) * | 2017-09-05 | 2019-03-07 | Future Mobility Corporation Limited | Methods and systems using a rotatable three-dimensional object providing visual user information on a display |
| US10809081B1 (en) * | 2018-05-03 | 2020-10-20 | Zoox, Inc. | User interface and augmented reality for identifying vehicles and persons |
Also Published As
| Publication number | Publication date |
|---|---|
| DE102021201774A1 (en) | 2021-09-09 |
| CN113361316A (en) | 2021-09-07 |
| US20210279913A1 (en) | 2021-09-09 |
| US11263787B2 (en) | 2022-03-01 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20250225690A1 (en) | Ar display apparatus and ar display method | |
| US9761139B2 (en) | Location based parking management system | |
| US20210001841A1 (en) | Obstacle Avoidance Method and Apparatus for Autonomous Driving Vehicle | |
| US20210019854A1 (en) | Location Signaling with Respect to an Autonomous Vehicle and a Rider | |
| CA2767312C (en) | Automatic video surveillance system and method | |
| US11900672B2 (en) | Integrated internal and external camera system in vehicles | |
| US20180196417A1 (en) | Location Signaling with Respect to an Autonomous Vehicle and a Rider | |
| US6243482B1 (en) | Obstacle detection system for low-flying airborne craft | |
| US20120320212A1 (en) | Surrounding area monitoring apparatus for vehicle | |
| US11520033B2 (en) | Techniques for determining a location of a mobile object | |
| EP3070698B1 (en) | Driving-support-image generation device, driving-support-image display device, driving-support-image display system, and driving-support-image generation program | |
| US10996070B2 (en) | Route guidance apparatus and method | |
| EP3566022B1 (en) | Location signaling with respect to an autonomous vehicle and a rider | |
| US20220092829A1 (en) | Augmented reality detection for locating autonomous vehicles | |
| EP4389550A1 (en) | Electronic device mounted on vehicle and operation method therefor | |
| US10996469B2 (en) | Method and apparatus for providing driving information of vehicle, and recording medium | |
| KR20200084777A (en) | System and method for training and operating an autonomous vehicle | |
| FR3079924A1 (en) | FOLLOW-UP BY GEOLOCATION | |
| US12238459B2 (en) | Rendezvous assistance apparatus, rendezvous assistance system, and rendezvous assistance method | |
| WO2006098311A1 (en) | Road scene map creating device, method, and program | |
| WO2019072461A1 (en) | Parking assistance method, control device for carrying out the parking assistance method, and vehicle comprising the control device | |
| JP6810723B2 (en) | Information processing equipment, information processing methods, and programs | |
| US12151686B2 (en) | Information provision system, information terminal, and information provision method | |
| US20240101133A1 (en) | Notification device and notification method | |
| US12420822B2 (en) | Data transmitter and method for data transmission |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
| AS | Assignment |
Owner name: RIVIAN AUTOMOTIVE, LLC, MICHIGAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HORTOP, MATTHEW;HUNT, PATRICK;TRELFA, LACEY;SIGNING DATES FROM 20211216 TO 20211217;REEL/FRAME:058460/0166 Owner name: RIVIAN IP HOLDINGS, LLC, MICHIGAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:RIVIAN AUTOMOTIVE, LLC;REEL/FRAME:058461/0028 Effective date: 20211217 |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
| STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
| STCB | Information on status: application discontinuation |
Free format text: EXPRESSLY ABANDONED -- DURING EXAMINATION |