WO2025106992A1 - Operator training using artificial intelligence agent - Google Patents
Operator training using artificial intelligence agent Download PDFInfo
- Publication number
- WO2025106992A1 WO2025106992A1 PCT/US2024/056416 US2024056416W WO2025106992A1 WO 2025106992 A1 WO2025106992 A1 WO 2025106992A1 US 2024056416 W US2024056416 W US 2024056416W WO 2025106992 A1 WO2025106992 A1 WO 2025106992A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- sensor
- presentation
- aircraft
- vehicle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01C—MEASURING DISTANCES, LEVELS OR BEARINGS; SURVEYING; NAVIGATION; GYROSCOPIC INSTRUMENTS; PHOTOGRAMMETRY OR VIDEOGRAMMETRY
- G01C23/00—Combined instruments indicating more than one navigational value, e.g. for aircraft; Combined measuring devices for measuring two or more variables of movement, e.g. distance, speed or acceleration
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/20—Arrangements for acquiring, generating, sharing or displaying traffic information
- G08G5/22—Arrangements for acquiring, generating, sharing or displaying traffic information located on the ground
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/20—Arrangements for acquiring, generating, sharing or displaying traffic information
- G08G5/23—Details of user output interfaces, e.g. information presented
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/50—Navigation or guidance aids
- G08G5/53—Navigation or guidance aids for cruising
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/50—Navigation or guidance aids
- G08G5/55—Navigation or guidance aids for a single aircraft
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/16—Control of vehicles or other craft
- G09B19/165—Control of aircraft
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
- G09B9/003—Simulators for teaching or training purposes for military purposes and tactics
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B9/00—Simulators for teaching or training purposes
- G09B9/02—Simulators for teaching or training purposes for teaching control of vehicles or other craft
- G09B9/08—Simulators for teaching or training purposes for teaching control of vehicles or other craft for teaching control of aircraft, e.g. Link trainer
- G09B9/30—Simulation of view from aircraft
- G09B9/307—Simulation of view from aircraft by helmet-mounted projector or display
-
- G—PHYSICS
- G08—SIGNALLING
- G08G—TRAFFIC CONTROL SYSTEMS
- G08G5/00—Traffic control systems for aircraft
- G08G5/20—Arrangements for acquiring, generating, sharing or displaying traffic information
- G08G5/21—Arrangements for acquiring, generating, sharing or displaying traffic information located onboard the aircraft
Definitions
- the present disclosure relates to the field of data processing. More specifically, the present disclosure relates to methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience.
- Display devices are used for various types of training, such as in simulators. Such display devices may display virtual reality and augmented reality content.
- movement of a display device with respect to a user using the display device may alter a perception of the content that may be displayed. For instance, due to a movement of the display device due to external forces, such as movement of display devices in flight helmets due to acceleration of aircraft, the user’s perception of the displayed content may change, which is not desired.
- the present invention relates to a method of debriefing based on a recording of a real flight of an aircraft by an operator in which the operator encountered a situation requiring a reaction by the operator.
- the recording comprising data of the real flight, the data comprising geospatial locations of the aircraft throughout the real flight and optionally those of one or more objects associated with the situation if applicable, flight parameters of the aircraft, and environmental conditions, and, optionally, a mission objective.
- the method comprising: (a) displaying a replay of at least a portion of the real flight; and (b) displaying an alternative reaction to the reaction of the operator using an Al agent, the Al agent being initialized with at least a portion of the data.
- the invention relates to a system of debriefing based on a recording of a real flight of an aircraft by an operator in which the operator encountered a situation requiring a reaction by the operator.
- the recording comprising data of the real flight, the data comprising geospatial locations of the aircraft throughout the real flight and optionally those of one or more objects associated with the situation if applicable, flight parameters of the aircraft, and environmental conditions, and, optionally, a mission objective.
- the system comprises: (a) a processor configured with an Al agent and the recording; (b) at least one display communicatively coupled to the processor, and configured for displaying a replay of at least a portion of the real flight, and for displaying an alternative reaction to the reaction of the operator using the Al agent, the Al agent being initialized with at least a portion of the data.
- drawings may contain text or captions that may explain certain embodiments of the present disclosure. This text is included for illustrative, non-limiting, explanatory purposes of certain embodiments detailed in the present disclosure.
- FIG.l is an illustration of an online platform consistent with various embodiments of the present disclosure.
- FIG. 2 shows a wearable display device for facilitating provisioning of a virtual experience, in accordance with some embodiments.
- FIG. 3 is a block diagram of a system for facilitating provisioning of a virtual experience in accordance with some embodiments.
- FIG. 4 is a block diagram of a first head mount display for facilitating provisioning of a virtual experience in accordance with some embodiments.
- FIG. 5 is a block diagram of an apparatus for facilitating provisioning of a virtual experience in accordance with some embodiments.
- FIG. 6 is a flowchart of a method of facilitating provisioning of a virtual experience in accordance with some embodiments.
- FIG. 7 shows a system for facilitating provisioning of a virtual experience, in accordance with some embodiments.
- FIG. 8 shows a corrected augmented reality view, in accordance with some embodiments.
- FIG. 9 shows an augmented reality view shown to a real pilot while a civilian aircraft is taxiing at an airport, in accordance with an exemplary embodiment.
- FIG. 10 is a block diagram of a computing device for implementing the methods disclosed herein, in accordance with some embodiments.
- FIG. 11 illustrates a debriefing virtual environment in accordance with the principles of the present invention
- any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the disclosure and may further incorporate only one or a plurality of the above-disclosed features.
- any embodiment discussed and identified as being “preferred” is considered to be part of a best mode contemplated for carrying out the embodiments of the present disclosure.
- Other embodiments also may be discussed for additional illustrative purposes in providing a full and enabling disclosure.
- many embodiments, such as adaptations, variations, modifications, and equivalent arrangements, will be implicitly disclosed by the embodiments described herein and fall within the scope of the present disclosure.
- any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present invention. Accordingly, it is intended that the scope of patent protection is to be defined by the issued claim(s) rather than the description set forth herein.
- the present disclosure includes many aspects and features. Moreover, while many aspects and features relate to, and are described in the context of facilitating provisioning of a virtual experience, embodiments of the present disclosure are not limited to use only in this context.
- FIG. 1 is an illustration of an online platform 100 consistent with various embodiments of the present disclosure.
- the online platform 100 to facilitate provisioning of a virtual experience may be hosted on a centralized server 102, such as, for example, a cloud computing service.
- the centralized server 102 may communicate with other network entities, such as, for example, an augmented and virtual reality display device 106, a sensor system 110 of an aircraft, database 114 (such as 3D model database) over a communication network 104, such as, but not limited to, the Internet.
- users of the online platform 100 may include relevant parties such as, but not limited to, trainees, trainers, pilots, administrators, and so on.
- a user 112 may access online platform 100 through a web based software application or browser.
- the web based software application may be embodied as, for example, but not be limited to, a website, a web application, a desktop application, and a mobile application compatible with a computing device 1000.
- FIG. 2 shows a wearable display device 200 for facilitating provisioning of a virtual experience.
- the wearable display device 200 may be utilized in conjunction with and/or to effectuate and/or facilitate operation of any element described elsewhere herein or illustrated in any figure herein.
- the wearable display device 200 may include a support member 202 configured to be mounted on a user 204.
- the support member 202 may include a structure allowing the support member 202 to be easily mountable on the user 204.
- the wearable display device 200 may include a head mounted device (HMD).
- the wearable display device 200 may include a display device 206 attached to the support member 202.
- HMD head mounted device
- the wearable display device 200 may include a display device in front of one eye of the user 204, (a monocular HMD), in front of both eyes of the user 204, (a binocular HMD), an optical display device (which may reflect projected images), and so on.
- the display device 206 may be configured for displaying at least one display data.
- the display data may include virtual reality data related to a simulation, such as a training simulation.
- the training simulation may correspond to vehicular racing, such as Formula 1®, and may be used by race car drivers to train for race events.
- the training simulation may correspond to flight training, and may be used by air force pilots for flight training in fighter aircraft.
- the display data may include augmented reality data.
- the display data may include one or more augmented reality components overlaid on top of live image.
- the augmented reality data may be related to flight training including a first aircraft training simultaneously with a plurality of aircrafts in different locations.
- the augmented reality data may include augmented reality components displaying the plurality of plurality of aircrafts in different locations to a display device associated with a pilot of the first aircraft.
- the wearable display device 200 may include at least one disturbance sensor 208 configured for sensing a disturbance in a spatial relationship between the display device 206 and the user 204.
- the spatial relationship between the display device 206 and the user 204 may include at least one of a distance and an orientation.
- the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the display device 206 and the eyes of the user 204.
- the disturbance in the spatial relationship may include a change in at least one of the distance and the orientation between the display device 206 and the user 204. Further, the disturbance in the spatial relationship may lead to an alteration in how the user 204 may view the at least one display data. For instance, if the disturbance in the spatial relationship leads to a reduction in the distance between the display device 206 and the user 204, the user 204 may perceive one or more objects in the at least one display data to be closer.
- the user 204 may perceive the at least one display data to be closer by “x-y” centimeters.
- the wearable display device 200 may include a processing device 210 communicatively coupled with the display device 206. Further, the processing device 210 may be configured for receiving the at least one display data. Further, the processing device 210 may be configured for analyzing the disturbance in the spatial relationship. Further, the processing device 210 may be configured for generating a correction data based on the analyzing. Further, the processing device 210 may be configured for generating a corrected display data based on the at least one display data and the correction data. Further, the correction data may include an instruction to shift a perspective view of the at least one display data to compensate for the disturbance in the spatial relationship between the display device 206 and the user 204. Accordingly, the correction data may be generated contrary to the disturbance in the spatial relationship.
- the disturbance may include an angular disturbance
- the display device 206 may undergo an angular displacement as a result of the angular disturbance.
- the correction data may include an instruction of translation of the display data to compensate for the angular disturbance.
- the display data may be translated along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data.
- the disturbance may include a longitudinal disturbance
- the display device 206 may undergo a longitudinal displacement as a result of the longitudinal displacement.
- the correction data may include an instruction of translation of the display data to compensate for the longitudinal disturbance.
- the display data may be projected along a distance perpendicular to a line of sight of the user 204 to negate the angular displacement of the display data.
- the display data may be projected along a distance perpendicular to the line of sight of the user 204 opposite to a direction of the longitudinal disturbance to compensate for the longitudinal disturbance.
- the support member 202 may include a head gear configured to be mounted on a head of the user 204.
- the head gear may include a helmet configured to be worn over a crown of the head.
- the head gear may include a shell configured to accommodate at least a part of a head of the user 204.
- a shape of the shell may define a concavity to facilitate accommodation of at least the part of the head.
- the shell may include an interior layer 212, an exterior layer 214 and a deformable layer 216 disposed in between the interior layer 212 and the exterior layer 214.
- the deformable layer 216 may be configured to provide cushioning.
- the display device 206 may be attached to at least one of the interior layer 212 and the exterior layer 214.
- the disturbance in the spatial relationship may be based on a deformation of the deformable layer 216 due to an acceleration of the head gear.
- the spatial relationship may include at least one vector representing at least one position of at least one part of the display device 206 in relation to at least one eye of the user 204.
- a vector of the at least one vector may be characterized by an orientation and a distance.
- the spatial relationship between the display device 206 and the user 204 may include at least one of a distance and an orientation.
- the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the display device 206 and the eyes of the user 204.
- the spatial relationship may describe an optimal arrangement of the display device 206 with respect to the user 204. Further, so that the optimal arrangement of the display device 206 with respect to the user 204 may allow the user to clearly view the display data without perceived distortion.
- the at least one disturbance sensor 208 may include an accelerometer configured for sensing the acceleration. Further, in some embodiments, the at least one disturbance sensor 208 may include at least one proximity sensor configured for sensing at least one proximity between the at least one part of the display device 206 and the user 204. Further, in some embodiments, the at least one disturbance sensor 208 may include a deformation sensor configured for sensing a deformation of the deformable layer 216. [0038] Further, in some embodiments, the display device 206 may include a see-through display device 206 configured to allow the user 204 to view a physical surrounding of the wearable device.
- the at least one display data may include at least one object model associated with at least one object. Further, in some embodiments, the generating of the corrected display data may include applying at least one transformation to the at least one object model based on the correction data.
- the applying of the at least one transformation to the at least one object model based on the correction data may include translation of the display data to compensate for the angular disturbance.
- the correction data may include one or more instructions to translate the display data along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data.
- the applying of the at least one transformation to the at least one object model based on the correction data may include translation of the display data along the horizontal axis, the vertical axis, and the diagonal axis of the display data, to negate the angular displacement of the display data.
- the applying of the at least one transformation to the at least one object model based on the correction data may include translation may include projection of the display data along a distance perpendicular to a line of sight of the user 204 to negate the angular displacement of the display data.
- the applying of the at least one transform may include projection of the display data along a distance perpendicular to the line of sight of the user 204 opposite to a direction of the longitudinal disturbance to compensate for the longitudinal disturbance.
- the at least one disturbance sensor 208 may include a camera configured to capture an image of each of a face of the user 204 and at least a part of the head gear. Further, the spatial relationship may include disposition of at least the part of the head gear in relation to the face of the user 204.
- the at least one disturbance sensor 208 may include a camera disposed on the display device 206. Further, the camera may be configured to capture an image of at least a part of a face of the user 204. Further, the wearable display device 200 may include a calibration input device configured to receive a calibration input. Further, the camera may be configured to capture a reference image of at least the part of the face of the user 204 based on receiving the calibration input. Further, the calibration input may be received in an absence of the disturbance. For instance, the calibration input device may include a button configured to be pushed by the user 204 in absence of the disturbance whereupon the reference image of at least the part of the face of the user 204 may be captured.
- the generating of the corrected display data may include applying at least one image transform on the at least one display data based on the at least one spatial parameter change.
- the wearable display device 200 may include at least one actuator coupled to the display device 206 and the support member 202. Further, the at least one actuator may be configured for modifying the spatial relationship based on a correction data.
- the spatial relationship between the display device 206 and the user 204 may include at least one of a distance 218 and an orientation.
- the disturbance in the spatial relationship between the display device 206 and the user 204 may include a change in at least one of the distance 218, the angle, the direction, and the orientation.
- the distance 218 may include a perceived distance between the user 204 and the at least one display data. For instance, as shown in FIG. 3, the disturbance in the spatial relationship may originate due to a forward acceleration 304 of the user 204 and the wearable display device 200.
- the deformation of the deformable layer 216 may lead to a disturbance in the spatial relationship leading to a change in the distance 218 to a reduced distance 302 between the display device 206 and the user 204.
- the correction data may include transforming of the at least one display data through object level processing and restoring the at least one display data to the distance 218 from the user 204.
- the object level processing may include projecting one or more objects in the display data at the distance 218 instead of the distance 302 to oppose the disturbance in the spatial relationship.
- the disturbance in the spatial relationship may include a change in the angle between the display device 206 and the user 204.
- the angle between the display device 206 and the user 204 in the spatial relationship may be related to an original viewing angle related to the display data.
- the original viewing angle related to the display data may be a viewing angle at which the user 204 may view the display data through the display device 206.
- the disturbance in the spatial relationship may lead to a change in the original viewing angle related to the display data.
- the at least one display data may be transformed through pixel level processing to restore the original viewing angle related to the display data.
- the pixel level processing may include translation of the display data to compensate for the change in the angle in the spatial relationship.
- the display data may be translated along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data to compensate for the change in the angle in the spatial relationship, and to restore the original viewing angle related to the display data.
- FIG. 3 is a block diagram of a system 300 for facilitating provisioning of a virtual experience in accordance with some embodiments.
- the system 300 may include a communication device 302, a processing device 304 and a storage device 306.
- the communication device 302 may be configured for receiving at least one first sensor data corresponding to at least one first sensor 310 associated with a first vehicle 308. Further, the at least one first sensor 310 may be communicatively coupled to a first transmitter 312 configured for transmitting the at least one first sensor data over a first communication channel. In some embodiments, the first vehicle 308 may be a first aircraft. Further, the first user may be a first pilot. [0048] Further, the communication device 302 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 320 associated with a second vehicle 318. Further, the at least one second sensor 320 may be communicatively coupled to a second transmitter 322 configured for transmitting the at least one second sensor data over a second communication channel. In some embodiments, the second vehicle 818 may be a second aircraft. Further, the second user may be a second pilot.
- the at least one first sensor data may be received from a first On-Board-Diagnostics (OBD) system of the first vehicle 308, the at least one second sensor data may be received from a second On-Board-Diagnostics (OBD) system of the second vehicle 318.
- OBD On-Board-Diagnostics
- the communication device 302 may be configured for receiving at least one first presentation sensor data from at least one first presentation sensor 328 associated with the first vehicle 308. Further, the at least one first presentation sensor 328 may be communicatively coupled to the first transmitter configured for transmitting the at least one first presentation sensor data over the first communication channel. Further, in an embodiment, the at least one first presentation sensor 328 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a first spatial relationship between at least one first presentation device 314 associated with the first vehicle 308, and the first user. Further, the spatial relationship between the at least one first presentation device 314 and the first user may include at least one of a distance and an orientation.
- the first spatial relationship may include an exact distance, and an orientation, such as a precise angle between the at least one first presentation device 314 and the eyes of the first user.
- the disturbance in the first spatial relationship may include a change in the at least of the distance and the orientation between the at least one first presentation device 314 and the first user.
- the communication device 302 may be configured for receiving at least one second presentation sensor data from at least one second presentation sensor 330 associated with the second vehicle 318.
- the at least one second presentation sensor 330 may include a disturbance sensor configured for sensing a disturbance in a second spatial relationship between at least one second presentation device 324 associated with the second vehicle 318, and the second user.
- the at least one second presentation sensor 330 may be communicatively coupled to the first transmitter configured for transmitting the at least one second presentation sensor data over the second communication channel.
- the communication device 302 may be configured for transmitting at least one first optimized presentation data to at least one first presentation device 314 associated with the first vehicle 808.
- at least one first presentation device 314 may include a wearable display device facilitating provisioning of a virtual experience, such as the wearable display device 200.
- the at least one first optimized presentation data may include a first corrected display data generated based on a first correction data.
- the at least one first presentation device 314 may include a first receiver 316 configured for receiving the at least one first optimized presentation data over the first communication channel. Further, the at least one first presentation device 314 may be configured for presenting the at least one first optimized presentation data.
- the communication device 302 may be configured for transmitting at least one second optimized presentation data to at least one first presentation device 314 associated with the first vehicle 308.
- the first receiver 316 may be configured for receiving the at least one second optimized presentation data over the first communication channel.
- the at least one first presentation device 314 may be configured for presenting the at least one second optimized presentation data.
- the at least one second optimized presentation data may include a second corrected display data generated based on a second correction data.
- the communication device 302 may be configured for transmitting at least one second optimized presentation data to at least one second presentation device 324 associated with the second vehicle 318.
- the at least one second presentation device 324 may include a second receiver 326 configured for receiving the at least one second optimized presentation data over the second communication channel.
- the at least one first presentation device 324 may be configured for presenting the at least one second optimized presentation data.
- the processing device 304 may be configured for analyzing the at least one first presentation sensor data associated with the first vehicle 308.
- the processing device 304 may be configured for analyzing the at least one second presentation sensor data associated with the second vehicle 318.
- the processing device 304 may be configured for generating the first correction data based on the analyzing the at least one first presentation sensor data associated with the first vehicle 308.
- the first correction data may include an instruction to shift a perspective view of the at least one first optimized presentation data to compensate for the disturbance in the first spatial relationship between the first presentation device 314 and the first user. Accordingly, the first correction data may be generated contrary to the disturbance in the first spatial relationship.
- the disturbance may include an angular disturbance, wherein the first presentation device 314 may undergo an angular displacement as a result of the angular disturbance.
- the first correction data may include an instruction of translation to generate the first corrected display data included in the first optimized presentation data to compensate for the angular disturbance.
- the processing device 304 may be configured for generating the second correction data based on the analyzing the at least one second presentation sensor data associated with the second vehicle 318.
- the second correction data may include an instruction to shift a perspective view of the at least one second optimized presentation data to compensate for the disturbance in the second spatial relationship between the second presentation device 324 and the second user. Accordingly, the second correction data may be generated contrary to the disturbance in the second spatial relationship.
- the disturbance may include an angular disturbance, wherein the second presentation device 324 may undergo an angular displacement as a result of the angular disturbance.
- the second correction data may include an instruction of translation to generate the second corrected display data included in the second optimized presentation data to compensate for the angular disturbance.
- the processing device 304 may be configured for generating the at least one first optimized presentation data based on the at least one second sensor data.
- the processing device 304 may be configured for generating the at least one first optimized presentation data based on the at least one first presentation sensor data.
- the processing device 304 may be configured for generating the at least one second optimized presentation data based on the at least one first sensor data.
- the processing device 304 may be configured for generating the at least one second optimized presentation data based on the at least one second presentation sensor data.
- the storage device 306 may be configured for storing each of the at least one first optimized presentation data and the at least one second optimized presentation data.
- the at least one first sensor 310 may include one or more of a first orientation sensor, a first motion sensor, a first accelerometer, a first location sensor, a first speed sensor, a first vibration sensor, a first temperature sensor, a first light sensor and a first sound sensor.
- the at least one second sensor 320 may include one or more of a second orientation sensor, a second motion sensor, a second accelerometer, a second location sensor, a second speed sensor, a second vibration sensor, a second temperature sensor, a second light sensor and a second sound sensor.
- the at least one first sensor 310 may be configured for sensing at least one first physical variable associated with the first vehicle 308.
- the at least one second sensor 320 may be configured for sensing at least one second physical variable associated with the second vehicle 318.
- the at least one first physical variable may include one or more of a first orientation, a first motion, a first acceleration, a first location, a first speed, a first vibration, a first temperature, a first light intensity and a first sound.
- the at least one second physical variable may include one or more of a second orientation, a second motion, a second acceleration, a second location, a second speed, a second vibration, a second temperature, a second light intensity and a second sound.
- the at least one first sensor 310 may include a first environmental sensor configured for sensing a first environmental variable associated with the first vehicle 308.
- the at least one second sensor 320 may include a second environmental sensor configured for sensing a second environmental variable associated with the second vehicle 318.
- the at least one first sensor 310 may include a first user sensor configured for sensing a first user variable associated with a first user of the first vehicle 308.
- the at least one second sensor 320 may include a second user sensor configured for sensing a second user variable associated with a second user of the second vehicle 318.
- the first user variable may include a first user location and a first user orientation.
- the second user variable may include a second user location and a second user orientation.
- the first presentation device may include a first head mount display.
- the second presentation device may include a second head mount display.
- the first head mount display may include a first user location sensor of the at least one first sensor 310 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 310 configured for sensing the first user orientation.
- the second head mount display may include a second user location sensor of the at least one second sensor 320 configured for sensing the second user location, a second user orientation sensor of the at least one second sensor 320 configured for sensing the second user orientation.
- the first vehicle 308 may include a first user location sensor of the at least one first sensor 310 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 310 configured for sensing the first user orientation.
- the second vehicle 318 may include a second user location sensor of the at least one second sensor 320 configured for sensing the second user location, a second user orientation sensor of the at least one second sensor 320 configured for sensing the second user orientation.
- the first user orientation sensor may include a first gaze sensor configured for sensing a first eye gaze of the first user.
- the second user orientation sensor may include a second gaze sensor configured for sensing a second eye gaze of the second user.
- the first user location sensor may include a first proximity sensor configured for sensing the first user location in relation to the at least one first presentation device 314.
- the second user location sensor may include a second proximity sensor configured for sensing the second user location in relation to the at least one second presentation device 324.
- the at least one first presentation sensor 328 may include at least one sensor configured for sensing at least one first physical variable associated with the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one camera configured to monitor a movement of the first presentation device 314 associated with the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one accelerometer sensor configured to monitor an uneven movement of the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one gyroscope sensor configured to monitor an uneven orientation of the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
- the at least one second presentation sensor 330 may include at least one sensor configured for sensing at least one first physical variable associated with the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one camera configured to monitor a movement of the second presentation device 324 associated with the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one accelerometer sensor configured to monitor an uneven movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one gyroscope sensor configured to monitor an uneven orientation of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the first head mount display may include a first see-through display device.
- the second head mount display may include a second see-through display device.
- the first head mount display may include a first optical marker configured to facilitate determination of one or more of the first user location and the first user orientation.
- the at least one first sensor 310 may include a first camera configured for capturing a first image of the first optical marker.
- the at least one first sensor 310 may be communicatively coupled to a first processor associated with the vehicle.
- the first processor may be configured for determining one or more of the first user location and the first user orientation based on analysis of the first image.
- the second head mount display may include a second optical marker configured to facilitate determination of one or more of the second user location and the second user orientation.
- the at least one second sensor 320 may include a second camera configured for capturing a second image of the second optical marker. Further, the at least one second sensor 320 may be communicatively coupled to a second processor associated with the vehicle. Further, the second processor may be configured for determining one or more of the second user location and the second user orientation based on analysis of the second image.
- the first presentation device may include a first see-through display device disposed in a first windshield of the first vehicle 308.
- the second presentation device may include a second see-through display device disposed in a second windshield of the second vehicle 318.
- the first vehicle 308 may include a first watercraft, a first land vehicle, a first aircraft and a first amphibious vehicle.
- the second vehicle 318 may include a second watercraft, a second land vehicle, a second aircraft and a second amphibious vehicle.
- the at least one may include one or more of a first visual data, a first audio data and a first haptic data.
- the at least one second optimized presentation data may include one or more of a second visual data, a second audio data and a second haptic data.
- the at least one first presentation device 314 may include at least one environmental variable actuator configured for controlling at least one first environmental variable associated with the first vehicle 308 based on the first optimized presentation data.
- the at least one second presentation device 324 may include at least one environmental variable actuator configured for controlling at least one second environmental variable associated with the second vehicle 318 based on the second optimized presentation data.
- the at least one first environmental variable may include one or more of a first temperature level, a first humidity level, a first pressure level, a first oxygen level, a first ambient light, a first ambient sound, a first vibration level, a first turbulence, a first motion, a first speed, a first orientation and a first acceleration
- the at least one second environmental variable may include one or more of a second temperature level, a second humidity level, a second pressure level, a second oxygen level, a second ambient light, a second ambient sound, a second vibration level, a second turbulence, a second motion, a second speed, a second orientation and a second acceleration.
- the first vehicle 308 may include each of the at least one first sensor 310 and the at least one first presentation device 314.
- the second vehicle 318 may include each of the at least one second sensor 320 and the at least one second presentation device 324.
- the storage device 306 may be further configured for storing a first three-dimensional model corresponding to the first vehicle 308 and a second three-dimensional model corresponding to the second vehicle 318. Further, the generating of the first optimized presentation data may be based further on the second three-dimensional model. Further, the generating of the second optimized presentation data may be based further on the first three- dimensional model.
- the generating of the first optimized presentation data may be based on the determining of the unwanted movement of the associated with the first presentation device 314 associated with the first vehicle 308, such as due to a G- Force, a frictional force, and an uneven movement of the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one camera configured to monitor a movement of the first presentation device 314 associated with the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one accelerometer sensor configured to monitor an uneven movement of the first presentation device 314 associated with the first vehicle 308, such as due to a G- Force, a frictional force, and an uneven movement of the first vehicle 308.
- the at least one first presentation sensor 328 may include at least one gyroscope sensor configured to monitor an uneven orientation of the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
- the generating of the second optimized presentation data may be based on the determining of the unwanted movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one camera configured to monitor a movement of the second presentation device 324 associated with the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one accelerometer sensor configured to monitor an uneven movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the at least one second presentation sensor 330 may include at least one gyroscope sensor configured to monitor an uneven orientation of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
- the communication device 302 may be further configured for receiving an administrator command from an administrator device. Further, the generating of one or more of the first optimized presentation data and the second optimized presentation data may be based further on the administrator command.
- the at least one first presentation model may include at least one first virtual object model corresponding to at least one first virtual object. Further, the at least one second presentation model may include at least one second virtual object model corresponding to at least one second virtual object. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor model. Further, the generating of the at least one second virtual object model may be independent of the at least one first sensor model.
- the generating of one or more of the at least one first virtual object model and the at least one second virtual object model may be based on the administrator command.
- the storage device 806 may be configured for storing the at least one first virtual object model and the at least one second virtual object model.
- the administrator command may include a virtual distance parameter. Further, the generating of each of the at least one first optimized presentation data and the at least one second optimized presentation data may be based on the virtual distance parameter.
- the at least one first sensor data may include at least one first proximity data corresponding to at least one first external real object in a vicinity of the first vehicle 308.
- the at least one second sensor data may include at least one second proximity data corresponding to at least one second external real object in a vicinity of the second vehicle 318. Further, the generating of the at least one first optimized presentation data may be based further on the at least one second proximity data.
- the generating of the at least one second optimized presentation data may be based further on the at least one first proximity data.
- the at least one first external real object may include a first cloud, a first landscape feature, a first man-made structure and a first natural object.
- the at least one second external real object may include a second cloud, a second landscape feature, a second man-made structure and a second natural object.
- the at least one first sensor data may include at least one first image data corresponding to at least one first external real object in a vicinity of the first vehicle 308.
- the at least one second sensor data may include at least one second image data corresponding to at least one second external real object in a vicinity of the second vehicle 318.
- the generating of the at least one first optimized presentation data may be based further on the at least one second image data.
- the generating of the at least one second optimized presentation data may be based further on the at least one first image data.
- the communication device 302 may be further configured for transmitting a server authentication data to the first receiver 316.
- the first receiver 316 may be communicatively coupled to first processor associated with the first presentation device.
- the first processor may be communicatively coupled to a first memory device configured to store a first authentication data.
- the first processor may be configured for performing a first server authentication based on the first authentication data and the server authentication data.
- the first processor may be configured for controlling presentation of the at least one first optimized presentation data on the at least one first presentation device 314 based on the first server authentication.
- the communication device 302 may be configured for transmitting a server authentication data to the second receiver 326.
- the second receiver 326 may be communicatively coupled to second processor associated with the second presentation device.
- the second processor may be communicatively coupled to a second memory device configured to store a second authentication data.
- the second processor may be configured for performing a second server authentication based on the second authentication data and the server authentication data.
- the second processor may be configured for controlling presentation of the at least one second optimized presentation data on the at least one second presentation device 324 based on the second server authentication.
- the communication device 302 may be configured for receiving a first client authentication data from the first transmitter 312.
- the storage device 306 may be configured for storing the first authentication data.
- the communication device 302 may be configured for and receiving a second client authentication data from the second transmitter 322.
- the storage device 306 may be configured for storing the second authentication data.
- the processing device 304 may be further configured for performing a first client authentication based on the first client authentication data and the first authentication data. Further, the generating of the at least one second optimized presentation data may be further based on the first client authentication. Further, the processing device 304 may be configured for performing a second client authentication based on the second client authentication data and the second authentication data. Further, the generating of the at least one first optimized presentation data may be further based on the second client authentication.
- FIG. 4 is a block diagram of a first head mount display 400 for facilitating provisioning of a virtual experience in accordance with some embodiments.
- the first head mount display 400 may include a first user location sensor 402 of the at least one first sensor configured for sensing the first user location and a first user orientation sensor 404 of the at least one first sensor configured for sensing the first user orientation.
- the first head mount display 400 may include a display device 406 to present visuals. Further, in an embodiment, the display device 406 may be configured for displaying the first optimized display data, as generated by the processing device 404. [0096] Further, the first head mount display 400 may include a processing device 408 configured to obtain sensor data from the first user location sensor 402 and the first user orientation sensor 404. Further, the processing device 408 may be configured to send visuals to the display device 406.
- FIG. 5 is a block diagram of an apparatus 500 for facilitating provisioning of a virtual experience in accordance with some embodiments.
- the apparatus 500 may include at least one first sensor 502 (such as the at least one first sensor 310) configured for sensing at least one first sensor data associated with a first vehicle (such as the first vehicle 308).
- the apparatus 500 may include at least one first presentation sensor 510 (such as the at least one first presentation sensor 328) configured for sensing at least one first presentation sensor data associated with a first vehicle (such as the first vehicle 308).
- the at least one first presentation sensor 510 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a first spatial relationship between at least one first presentation device 508 associated with the first vehicle, and a first user.
- the spatial relationship between the at least one first presentation device 508 and the first user may include at least one of a distance and an orientation.
- the first spatial relationship may include an exact distance, and an orientation, such as a precise angle between the at least one first presentation device 508 and the eyes of the first user.
- the disturbance in the first spatial relationship may include a change in the at least of the distance and the orientation between the at least one first presentation device 314 and the first user.
- the apparatus 500 may include a first transmitter 504 (such as the first transmitter 312) configured to be communicatively coupled to the at least first sensor 502, and the at least one first presentation sensor 510. Further, the first transmitter 504 may be configured for transmitting the at least one first sensor data and the at least one first presentation sensor data to a communication device (such as the communication device 302) of a system over a first communication channel. [00100] Further, the apparatus 500 may include a first receiver 506 (such as the first receiver 316) configured for receiving the at least one first optimized presentation data from the communication device over the first communication channel.
- a first transmitter 504 such as the first transmitter 312 configured to be communicatively coupled to the at least first sensor 502, and the at least one first presentation sensor 510. Further, the first transmitter 504 may be configured for transmitting the at least one first sensor data and the at least one first presentation sensor data to a communication device (such as the communication device 302) of a system over a first communication channel.
- the apparatus 500 may include a first receiver 506
- the apparatus 500 may include the at least one first presentation device 508 (such as the at least one first presentation device 314) configured to be communicatively coupled to the first receiver 506.
- the at least one first presentation device 508 may be configured for presenting the at last one first optimized presentation data.
- the communication device may be configured for receiving at least one second sensor data corresponding to at least one second sensor (such as the at least one second sensor 320) associated with a second vehicle (such as the second vehicle 318). Further, the at least one second sensor may be communicatively coupled to a second transmitter (such as the second transmitter 322) configured for transmitting the at least one second sensor data over a second communication channel. Further, the system may include a processing device (such as the processing device 304) communicatively coupled to the communication device. Further, the processing device may be configured for generating the at least one first optimized presentation data based on the at least one second sensor data.
- FIG. 6 is a flowchart of a method 600 of facilitating provisioning of a virtual experience in accordance with some embodiments.
- the method 600 may include receiving, using a communication device (such as the communication device 302), at least one first sensor data corresponding to at least one first sensor (such as the at least one first sensor 310) associated with a first vehicle (such as the first vehicle 308). Further, the at least one first sensor may be communicatively coupled to a first transmitter (such as the first transmitter 312) configured for transmitting the at least one first sensor data over a first communication channel.
- a communication device such as the communication device 302
- at least one first sensor data corresponding to at least one first sensor such as the at least one first sensor 310) associated with a first vehicle (such as the first vehicle 308).
- the at least one first sensor may be communicatively coupled to a first transmitter (such as the first transmitter 312) configured for transmitting the at least one first sensor data over a first communication channel.
- the method 600 may include receiving, using the communication device, at least one second sensor data corresponding to at least one second sensor (such as the at least one second sensor 320) associated with a second vehicle (such as the second vehicle 318). Further, the at least one second sensor may be communicatively coupled to a second transmitter (such as the second transmitter 322) configured for transmitting the at least one second sensor data over a second communication channel.
- at least one second sensor such as the at least one second sensor 320
- a second vehicle such as the second vehicle 318
- the at least one second sensor may be communicatively coupled to a second transmitter (such as the second transmitter 322) configured for transmitting the at least one second sensor data over a second communication channel.
- the method 600 may include receiving, using the communication device, a first presentation sensor data corresponding to at least one first presentation sensor 328 associated with the first vehicle.
- the at least one first presentation sensor may be communicatively coupled to the first transmitter configured for transmitting the at least one first presentation sensor data over the first communication channel.
- the first presentation sensor may include at least one sensor configured to monitor a movement of at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle.
- the at least one first presentation sensor may include at least one camera configured to monitor a movement of the at least one first presentation device associated with the first vehicle.
- the at least one first presentation sensor may include at least one accelerometer sensor configured to monitor an uneven movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle.
- the at least one first presentation sensor may include at least one gyroscope sensor configured to monitor an uneven orientation of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle.
- the method 600 may include receiving, using the communication device, a second presentation sensor data corresponding to at least one second presentation sensor 330 associated with the second vehicle.
- the at least one second presentation sensor may be communicatively coupled to the second transmitter configured for transmitting the at least one second presentation sensor data over the second communication channel.
- the second presentation sensor may include at least one sensor configured to monitor a movement of at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle.
- the at least one second presentation sensor may include at least one camera configured to monitor a movement of the at least one second presentation device associated with the second vehicle.
- the at least one second presentation sensor may include at least one accelerometer sensor configured to monitor an uneven movement of the at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle.
- the at least one second presentation sensor may include at least one gyroscope sensor configured to monitor an uneven orientation of the at least one second presentation device associated with the second vehicle, such as due to a G- Force, a frictional force, and an uneven movement of the second vehicle.
- the method 600 may include analyzing, using a processing device, the at least one first sensor data and the at least one first presentation sensor data to generate at least one first modified presentation data.
- the analyzing may include determining an unwanted movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. Further, the unwanted movement of the at least one first presentation device associated with the first vehicle may include an upward movement, a downward movement, a leftward movement, and a rightward movement.
- the generating of the at least one first optimized presentation data may be based on the unwanted movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. For instance, the generating of the at least one first optimized presentation data may be based on negating an effect of the unwanted movement of the at least one first presentation device associated with the first vehicle.
- the generating of the at least one first optimized presentation data may include moving one or more components of the at least one first modified presentation data in an oppositely downward direction, an upward direction, a rightward direction, and a leftward direction respectively.
- the method 600 may include analyzing, using a processing device, the at least one second sensor data and the at least one second presentation sensor data to generate at least one second presentation data.
- the analyzing may include determining an unwanted movement of the at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. Further, the unwanted movement of the at least one second presentation device associated with the second vehicle may include an upward movement, a downward movement, a leftward movement, and a rightward movement.
- the generating of the at least one second optimized presentation data may be based on the unwanted movement of the at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. For instance, the generating of the at least one second optimized presentation data may be based on negating an effect of the unwanted movement of the at least one second presentation device associated with the second vehicle.
- the generating of the at least one second optimized presentation data may include moving one or more components of the at least one second presentation data in an oppositely downward direction, an upward direction, a rightward direction, and a leftward direction respectively.
- the method 600 may include transmitting, using the communication device, at least one first optimized presentation data to at least one first presentation device associated with the first vehicle.
- the at least one first presentation device may include a first receiver (such as the first receiver 316) configured for receiving the at least one first modified presentation data over the first communication channel.
- the at least one presentation device may be configured for presenting the at least one first optimized presentation data.
- the method 600 may include transmitting, using the communication device, at least one second optimized presentation data to at least one second presentation device (such as the at least one second presentation device 324) associated with the second vehicle. Further, the at least one second presentation device may include a second receiver (such as the second receiver 326) configured for receiving the at least one second presentation data over the second communication channel. Further, the at least one presentation device may be configured for presenting the at least one second optimized presentation data. [00111] At 618, the method 600 may include storing, using a storage device (such as the storage device 306), each of the at least one first optimized presentation data and the at least one second optimized presentation data.
- a storage device such as the storage device 306
- FIG. 7 shows a system 700 for facilitating provisioning of a virtual experience, in accordance with some embodiments.
- the system 700 may include a communication device 702 configured for receiving at least one first sensor data corresponding to at least one first sensor 710 associated with a first vehicle 708. Further, the at least one first sensor 710 may be communicatively coupled to a first transmitter 712 configured for transmitting the at least one first sensor data over a first communication channel.
- the communication device 702 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 716 associated with a second vehicle 714.
- the at least one second sensor 716 may include a second location sensor configured to detect a second location associated with the second vehicle 714.
- the at least one second sensor 716 may be communicatively coupled to a second transmitter 718 configured for transmitting the at least one second sensor data over a second communication channel.
- the at least one second sensor 716 may include a second user sensor configured for sensing a second user variable associated with a second user of the second vehicle 714.
- the second user variable may include a second user location and a second user orientation.
- the at least one second sensor 716 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a spatial relationship between a second presentation device 720 associated with the second vehicle 714 and the second user of the second vehicle 714.
- the spatial relationship between the second presentation device 720 and the second user may include at least one of a distance and an orientation.
- the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the second presentation device 720 and the eyes of the second user.
- the disturbance in the spatial relationship may include a change in the at least of the distance and the orientation between the second presentation device 720 and the second user.
- the disturbance in the spatial relationship may lead to an alteration in how the second user may view at least one second presentation data. For instance, if the disturbance in the spatial relationship leads to a reduction in the distance between the second presentation device 720 and the second user, the second user may perceive one or more objects in the at least one second presentation data to be closer. For instance, if the spatial relationship between the second presentation device 720 and the second user specifies a distance of “x” centimeters, and the disturbance in the spatial relationship leads to a reduction in the distance between the second presentation device 720 and the second user to “y” centimeters, the second user may perceive the at least one second presentation data to be closer by “x-y” centimeters.
- the communication device 702 may be configured for transmitting the at least one second presentation data to the at least one second presentation device 720 associated with the second vehicle 714.
- the at least one second presentation data may include at least one second virtual object model corresponding to at least one second virtual object.
- the at least one second virtual object may include one or more of a navigational marker and an air-corridor.
- the at least one second presentation data may include a second corrected display data generated based on a second correction data.
- the at least one second presentation device 720 may include a second receiver 722 configured for receiving the at least one second presentation data over the second communication channel.
- the at least one second presentation device 720 may be configured for presenting the at least one second presentation data.
- the at least one second presentation device 720 may include a second head mount display.
- the second head mount display may include a second user location sensor of the at least one second sensor 716 configured for sensing the second user location and a second user orientation sensor of the at least one second sensor 716 configured for sensing the second user orientation.
- the second head mount display may include a second see-through display device.
- the at least one second virtual object model may include a corrected augmented reality view, such as the corrected augmented reality view 800.
- the augmented reality view 800 may include one or more second virtual objects such as a navigational marker 808, and a skyway 806 as shown in FIG. 8).
- the system 700 may include a processing device 704 configured for generating the at least one second presentation data based on the at least one first sensor data and the at least one second sensor data. Further, the generating of the at least one second virtual object model may be independent of the at least one first sensor data. Further, in some embodiments, the processing device 704 may be configured for determining a second airspace class associated with the second vehicle 714 based on the second location including a second altitude associated with the second vehicle 714. Further, the generating of the at least one second virtual object model may be based on the second airspace class.
- the processing device 704 may be configured for generating the second correction data based on the analyzing the at least one second sensor data associated with the second vehicle 714.
- the second correction data may include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user. Accordingly, the second correction data may be generated contrary to the disturbance in the spatial relationship.
- the disturbance may include an angular disturbance, wherein the second presentation device 720 may undergo an angular displacement as a result of the angular disturbance.
- the second correction data may include an instruction of translation to generate the second corrected display data included in the second presentation data to compensate for the angular disturbance.
- the at least one second presentation data includes the at least one second virtual object model may include a corrected augmented reality view, such as the corrected augmented reality view 800
- the second correction data may include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user (such as a pilot 802).
- the second correction data may include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user, such as by projection of the one or more second virtual objects, such as the navigational marker 808, and the skyway 806 at a distance to compensate the disturbance and to generate the corrected augmented reality view 800.
- the system 700 may include a storage device 706 configured for storing the at least one second presentation data. Further, in some embodiments, the storage device 706 may be configured for retrieving the at least one second virtual object model based on the second location associated with the second vehicle 714. Further, in some embodiments, the storage device 706 may be configured for storing a first three-dimensional model corresponding to the first vehicle 708. Further, the generating of the second presentation data may be based on the first three-dimensional model.
- the communication device 702 may be configured for receiving an administrator command from an administrator device. Further, the generating of the at least one second virtual object model may be based on the administrator command.
- the communication device 702 may be configured for transmitting at least one first presentation data to at least one first presentation device (not shown) associated with the first vehicle 708.
- the at least one first presentation device may include a first receiver configured for receiving the at least one first presentation data over the first communication channel.
- the at least one first presentation device may be configured for presenting the at least one first presentation data.
- the processing device 704 may be configured for generating the at least one first presentation data based on the at least one second sensor data.
- the storage device 706 may be configured for storing the at least one first presentation data.
- the storage device 706 may be configured for storing a second three- dimensional model corresponding to the second vehicle 714. Further, the generating of the first presentation data may be based on the second three-dimensional model.
- the at least one first presentation data may include at least one first virtual object model corresponding to at least one first virtual object. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor data. Further, the storage device 706 may be configured for storing the at least one first virtual object model.
- the communication device 702 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 716 associated with a second vehicle 714. Further, the at least one second sensor 716 may be communicatively coupled to a second transmitter 718 configured for transmitting the at least one second sensor data over a second communication channel. Further, the communication device 702 may be configured for receiving at least one first sensor data corresponding to at least one first sensor 710 associated with a first vehicle 708. Further, the at least one first sensor 710 may include a first location sensor configured to detect a first location associated with the first vehicle 708.
- the at least one first sensor 710 may be communicatively coupled to a first transmitter 712 configured for transmitting the at least one first sensor data over a first communication channel.
- the at least one first sensor 710 may include a first user sensor configured for sensing a first user variable associated with a first user of the first vehicle 708.
- the first user variable may include a first user location and a first user orientation.
- the communication device 702 configured for transmitting at least one first presentation data to at least one first presentation device (not shown) associated with the first vehicle 708.
- the at least one first presentation data may include at least one first virtual object model corresponding to at least one first virtual object.
- the at least one first virtual object may include one or more of a navigational marker (such as a navigational marker 708, and/or a signboard 904 as shown in FIG. 9) and an air-corridor (such as a skyway 806 as shown in FIG. 8).
- the at least one first presentation device may include a first receiver configured for receiving the at least one first presentation data over the first communication channel. Further, the at least one first presentation device may be configured for presenting the at least one first presentation data. Further, in some embodiments, the at least one first presentation device may include a first head mount display.
- the first head mount display may include a first user location sensor of the at least one first sensor 710 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 710 configured for sensing the first user orientation.
- the first head mount display may include a first see- through display device.
- the processing device 704 may be configured for generating the at least one first presentation data based on the at least one second sensor data and the at least one first sensor data. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor data. Further, in some embodiments, the processing device 704 may be configured for determining a first airspace class associated with the first vehicle 708 based on the first location including a first altitude associated with the first vehicle 708.
- the generating of the at least one first virtual object model may be based on the administrator command.
- the communication device 702 may be configured for transmitting at least one second presentation data to at least one second presentation device (such as the second presentation device 720) associated with the second vehicle 714.
- the at least one second presentation device may include a second receiver (such as the second receiver 722) configured for receiving the at least one second presentation data over the second communication channel.
- the at least one second presentation device may be configured for presenting the at least one second presentation data.
- the processing device 704 may be configured for generating the at least one second presentation data based on the at least one first sensor data.
- the storage device 706 may be configured for storing the at least one second presentation data. Further, in some embodiments, the storage device 706 may be configured for storing a first three-dimensional model corresponding to the first vehicle 708. Further, the generating of the second presentation data may be based on the first three-dimensional model. Further, in some embodiments, the at least one second presentation data may include at least one second virtual object model corresponding to at least one second virtual object.
- the generating of the at least one second virtual object model may be independent of the at least one first sensor data.
- the storage device 706 may be configured for storing the at least one second virtual object model.
- FIG. 8 shows the corrected augmented reality view 800.
- the augmented reality view 800 may include a road drawn in the sky (such as the skyway 806) indicating a path that a civilian aircraft 804 may take in order to land at an airport.
- the augmented reality view 800 may include the navigation marker 808 indicating to a pilot 802 that the civilian aircraft 804 should take a left turn. The navigation marker 808 may assist the pilot 802 in navigating towards a landing strip to land the civilian aircraft 804.
- the corrected augmented reality view 800 may provide pilots with a similar view as seen by public transport drivers (e.g. taxi or bus) on the ground.
- the pilots (such as the pilot 802) may see roads (such as the skyway 806) that the pilot 802 need to drive on. Further, the pilot 802, in an instance, may see signs just like a taxi driver who may just look out of a window and see road signs.
- the corrected augmented reality view 800 may include (but not limited to) one or more of skyways (such the skyway 806), navigation markers (such as the navigation marker 808), virtual tunnels, weather information, an air corridor, speed, signboards for precautions, airspace class, one or more parameters shown on a conventional horizontal situation indicator (HSI) etc.
- the skyways may indicate a path that an aircraft (such as the civilian aircraft 804) should take.
- the skyways may appear similar to roads on the ground.
- the navigation markers may be similar to regulatory road signs used on the roads on the ground. Further, the navigation markers may instruct pilots (such as the pilot 802) on what they must or should do (or not do) under a given set of circumstances.
- the navigation markers may be used to reinforce air-traffic laws, regulations or requirements which apply either at all times or at specified times or places upon a flight path.
- the navigation markers may include one or more of a left curve ahead sign, a right curve ahead sign, a keep left sign, and a keep to right sign.
- the virtual tunnels may appear similar to tunnels on roads on the ground.
- the pilot 802 may be required to fly the aircraft through the virtual tunnel.
- the weather information may include real-time weather data that affects flying conditions.
- the weather information may include information related to one or more of wind speed, gust, and direction; variable wind direction; visibility, and variable visibility; temperature; precipitation; and cloud cover.
- the air corridor may indicate an air route along which the aircraft is allowed to fly, especially when the aircraft is over a foreign country.
- the corrected augmented reality view 800 may include speed information.
- the speed information may include one or more of a current speed, a ground speed, and a recommended speed.
- the signboards for precautions may be related to warnings shown to the pilot 802.
- the one or more parameters shown on a conventional horizontal situation indicator (HSI) include NAV warning flag, lubber line, compass warning flag, course select pointer, TO/FROM indicator, glideslope deviation scale, heading select knob, compass card, course deviation scale, course select knob, course deviation bar (CD I), symbolic aircraft, dual glideslope pointers, and heading select bug.
- information such as altitude, attitude, airspeed, the rate of climb, heading, autopilot and auto-throttle engagement status, flight director modes and approach status etc. that may be displayed on a conventional primary flight display may also be displayed in the corrected augmented reality view 800.
- the corrected augmented reality view 800 may include a one or more of other vehicles (such as another airplane 810).
- the one or more other vehicles in an instance, may include one or more live vehicles (such as representing real pilots flying real aircraft), one or more virtual vehicles (such as representing real people on the ground, flying virtual aircraft), and one or more constructed vehicles (such as representing aircraft generated and controlled using computer graphics and processing systems).
- a special use airspace class may be determined.
- the special use airspace class may include alert areas, warning areas, restricted areas, prohibited airspace, military operation area, national security area, controlled firing areas etc. For an instance, if an aircraft (such as the civilian aircraft 804) enters a prohibited area by mistake, then a notification may be displayed in the corrected augmented reality view 800. Accordingly, the pilot 802 may reroute the aircraft towards a permitted airspace.
- the corrected augmented reality view 800 may include one or more live aircraft (representing real pilots flying real aircraft), one or more virtual aircraft (representing real people on the ground, flying virtual aircraft) and one or more constructed aircraft (representing aircraft generated and controlled using computer graphics and processing systems). Further, the corrected augmented reality view 800 shown to a pilot (such as the pilot 802) in a first aircraft (such as the civilian aircraft 804) may be modified based on sensor data received from another aircraft (such as another airplane 1410). The sensor data may include data received from one or more internal sensors to track and localize the pilot's head within the cockpit of the aircraft. Further, the sensor data may include data received from one or more external sensors to track the position and orientation of the aircraft. Further, the data received from the one or more internal sensors and the one or more external sensors may be combined to provide a highly usable augmented reality solution in a fast-moving environment.
- FIG. 9 shows an augmented reality view 900 shown to a real pilot while a civilian aircraft 902 is taxiing at an airport, in accordance with an exemplary embodiment.
- the augmented reality view 900 may include one or more navigational markers (such as the navigation marker 808) and signboards (such as a signboard 904) that assist a pilot to taxi the civilian aircraft 902 at the airport.
- the navigational markers may indicate the direction of movement.
- the signboards may indicate the speed limits.
- the augmented reality view 900 may help the pilot to taxi the civilian aircraft 902 towards a parking location after landing. Further, augmented reality view 900 may help the pilot to taxi the civilian aircraft 902 towards a runway for taking- off Therefore, a ground crew may no longer be required to instruct the pilot while taxiing the civilian aircraft 902 at the airport.
- the augmented reality view 900 may include one or more live aircraft (such as a live aircraft 906) at the airport (representing real pilots in real aircraft), one or more virtual aircraft at the airport (representing real people on the ground, controlling a virtual aircraft) and one or more constructed aircraft at the airport (representing aircraft generated and controlled using computer graphics and processing systems).
- the augmented reality view 900 shown to a pilot in a first aircraft may be modified based on sensor data received from another aircraft.
- the sensor data may include data received from one or more internal sensors to track and localize the pilot's head within the cockpit of the aircraft.
- the sensor data may include data received from one or more external sensors to track the position and orientation of the aircraft.
- the data received from the one or more internal sensors and the one or more external sensors may be combined to provide a highly usable augmented reality solution in a fast-moving environment.
- the process of acquiring sensor information from one or more vehicles, maintaining a repository of data describing various real and virtual platforms and environments, and generating presentation data may be distributed among various platforms and among a plurality of processors.
- a system consistent with an embodiment of the disclosure may include a computing device or cloud service, such as computing device 1000.
- computing device 1000 may include at least one processing unit 1002 and a system memory 1004.
- system memory 1004 may include, but is not limited to, volatile (e.g. random-access memory (RAM)), non-volatile (e.g. read-only memory (ROM)), flash memory, or any combination.
- System memory 1004 may include operating system 1005, one or more programming modules 1006, and may include a program data 1007. Operating system 1005, for example, may be suitable for controlling computing device 1000’s operation.
- programming modules 1006 may include image-processing module, machine learning module and/or image classifying module. Furthermore, embodiments of the disclosure may be practiced in conjunction with a graphics library, other operating systems, or any other application program and is not limited to any particular application or system. This basic configuration is illustrated in FIG. 10 by those components within a dashed line 1008.
- Computing device 1000 may have additional features or functionality.
- computing device 1000 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape.
- additional storage is illustrated in FIG. 10 by a removable storage 1009 and a non-removable storage 1010.
- Computer storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data.
- System memory 1004, removable storage 1009, and non-removable storage 1010 are all computer storage media examples (i.e., memory storage.)
- Computer storage media may include, but is not limited to, RAM, ROM, electrically erasable read-only memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store information and which can be accessed by computing device 1000. Any such computer storage media may be part of device 1000.
- Computing device 1000 may also have input device(s) 1012 such as a keyboard, a mouse, a pen, a sound input device, a touch input device, a location sensor, a camera, a biometric sensor, etc.
- Output device(s) 1014 such as a display, speakers, a printer, etc. may also be included.
- the aforementioned devices are examples and others may be used.
- Computing device 1000 may also contain a communication connection 1016 that may allow device 1000 to communicate with other computing devices 1018, such as over a network in a distributed computing environment, for example, an intranet or the Internet.
- Communication connection 1016 is one example of communication media.
- Communication media may typically be embodied by computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media.
- modulated data signal may describe a signal that has one or more characteristics set or changed in such a manner as to encode information in the signal.
- communication media may include wired media such as a wired network or direct- wired connection, and wireless media such as acoustic, radio frequency (RF), infrared, and other wireless media.
- wireless media such as acoustic, radio frequency (RF), infrared, and other wireless media.
- RF radio frequency
- computer readable media may include both storage media and communication media.
- program modules and data files may be stored in system memory 1004, including operating system 1005.
- programming modules 1006 e.g., application 1020 such as a media player
- processing unit 1002 may perform other processes.
- Other programming modules that may be used in accordance with embodiments of the present disclosure may include sound encoding/decoding applications, machine learning application, acoustic classifiers etc.
- Asset operators, ground troops and others involved in military combat may find themselves in complex situations and they may have to make a series of decisions in quick succession to accomplish a mission. These individuals may have a plan and a leader, but each one, or groups of people, still have to make individual decisions based on their training and information that they have about the situation. Communication and adherence to validated tactics is vital in such situations and insightful guidance provides a path to success.
- Al systems may process vast amounts of combat field data and provide insightful guidance to individuals, groups, leaders, etc. while they are being trained and while they are in combat situations.
- a fighter pilot may be on a mission to escort and protect a strike package on a mission.
- the flight may encounter enemy fighters approaching to disrupt the package’s mission.
- the escorting fighter pilot(s) has to make a decision on how to deal with the incoming fighters.
- the enemy may be a simple configuration of a manageable few assets, but the enemy may be a well -organized force with an advanced Integrated Air Defense System (IADS).
- IADS advanced Integrated Air Defense System
- the inventors discovered new training systems for operators of vehicles.
- the training systems include presenting augmented reality content to vehicle operators as they operate real assets in real environments to provide the operators with experience in maneuvering programmed situations.
- the training systems also include debrief tools where recordings of the training sessions can be reviewed, and maneuver improvements can be taught to the operators.
- a human trainer may control the debrief tools in various ways to teach the operator what they might have done differently in the given situation.
- Al asset performance agents may also be used in the debrief to demonstrate what an Al trained asset may have done in the given situation that replicates what an experienced operator would have executed.
- An Al asset performance agent (sometimes referred to as an “Al agent” herein) is a computer program that is designed to mimic the performance of an operator based on a physics model and/or training data.
- a physics model represents the performance attributes of a particular type of asset. For example, one may want to develop an Al agent to fly and maneuver in situations mimicking an aircraft like an F- 22 or an adversary’s aircraft.
- the F-22 Al agent would be programmed with a physics engine that represents the main flight performance capabilities of the F-22.
- the F-22 Al agent might have, for example, top speed at given altitudes, fuel range predictions based on performance, climb rate, G-force limits, pilot limits, thrust, turn rate, turn radius, etc.
- a virtual representation of an F-22 in flight, on the ground, etc. may then be computer controlled in accordance with instructions as limited by or guided by the physics engine. This enables a computer to run simulations using the F-22 Al agent in various training situations to discover trends, develop new strategies for managing such environments and training based on an Al model.
- a human trainer may also give feedback to improve the Al model (e.g., modifying algorithmic weightings).
- the Al agent may also be trained using actual operator / asset performance data.
- An Al agent may be used as a virtual asset in a live augmented reality training environment to represent an asset to which an operator may have to interact. For example, a real pilot of a real airplane may fly the plane while wearing augmented reality gear.
- An adversary aircraft may be presented as augmented reality content to the pilot at a geospatial position based on virtual markers and the pilot may be able to engage in air combat with the adversary aircraft .
- the adversary aircraft may be controlled by, or in part controlled by, an Al agent trained to represent the particular type of adversary aircraft.
- the Al agent may be adjusted to represent an extreme version of the enemy (e.g., making all the right maneuvers) or it may be adjusted to represent something less (e.g., making most of the right maneuvers).
- the adjustment may call for ‘common’ mistakes or tendencies to be included in the control methodology.
- the adjustment may depend on the type of training designed for the pilot.
- a debrief tool may be used to illustrate how an operator of an asset (e.g., pilot of a plane) maneuvered with respect to the augmented reality training environment.
- the debrief environment may include a 2D, 3D or another virtual environment.
- the debrief tool may include a 3D virtual reality environment where the aircraft and training objects maneuvered.
- the training objects may have been live (e.g., real assets in the environment), virtual (e.g., computer generated and controlled), constructive (e.g., computer generated and at least partially human controlled).
- Data from a training session may be used in generating the presentation in the virtual environment.
- the data may represent the attributes of the various assets over time.
- the real plane(s) may have been positionally tracked as well as conditionally tracked to provide an understanding of where the plane was at a particular time and what was going on with the plane’s avionics, flight planes, speed, G-force experience, kinetic energy states, potential energy states, etc.
- the virtual and constructive asset may log similar information such that all of the relevant information can be presented in the virtual environment.
- All of the training assets may be presented in the virtual environment debrief tool.
- a human instructor and the trainee may talk about what was happening during training and the instructor may offer advice.
- One form of advice may be delivered by allowing the instructor to grab objects in the virtual environment to show their conditional data, move the object to talk about an alternate scenario, cause an object to make a maneuver different than what was performed, etc.
- Another form of training in the debrief virtual environment would be to allow an Al agent representing the type of asset being used by the trainee to show what it would have done in the given situation during a playback of a real training session that may diverge from that the trainee performed.
- the Al agent maneuvers may be shown in comparison to, or instead of, the playback of what the trainee did in their flight.
- a user of the debrief tools may stop the playback of the trainee’s flight at some point instruct an Al agent to take over from that point to demonstrate what it would have done.
- the Al agent becomes situationally educated by using situational information gathered by the exercise. For example, the trainee’s airplane data (e.g., speed, energy states, pose, orientation) was tracked up to the point of stopping the playback and some or all of the data may be incorporated into the Al agent to set it in a proper relational condition as the trainee’s airplane was at the time.
- the Al agent may also be programmed with data representing the environment, time of day, angle of the sun, landscape, threats, etc.
- the Al agent may also be programmed to mimic a condition of the trainee (e.g., how long they had been flying in the current mission, biometric information).
- the Al agent may be programmed to understand the mission or desired outcome.
- the Al agent Once the Al agent has been situationally educated it may be deployed within the debrief virtual environment. It may then be permitted to move forward in time to demonstrate alternate maneuvers. It may also be sent back in time to reset its situational education. A human instructor may talk about how the trainee reacted to situations in comparison to the Al agent’s performance. The debrief tool may compare the performance of the trainee’s performance with the Al agent’s performance to identify gaps, grades, critiques, tendencies, etc.
- FIG. 11 illustrates a debriefing virtual environment in accordance with the principles of the present invention.
- a trace 1102 representing the trainee’s flight path during a training event may be presented in geospatial relation with an enemy threat 1104 that was presented to the trainee during the training event. As illustrated, the trainee turned sharply right as they came into closer proximity to the threat.
- a human trainer conducting the debrief may see an opportunity to show the trainee alternatives to the maneuver made.
- the trainer may select a point 1106 along the trainee's flight path to start an Al agent’s evaluation and suggestions.
- the Al agent may begin by programming its current conditions with the conditions of the trainee’s real plane such that it is in a similar condition (e.g., speed, energy states, enemy asset situations). Then, the Al agent may be instructed to make what it sees as the best maneuver 1108 for the given situation.
- Al agent's performance suggestions There may be a number of ways to present the Al agent's performance suggestions in the debrief tools.
- a user may decide how to look at the data and/or data representations.
- the visual presentation of the actual operator’s performance may be positioned in the same virtual environment as the Al agent’s suggested performance (e.g., either the operator or the Al agent may be presented in a ‘ghosting’ fashion).
- the operator’s performance may be hidden so focus can be placed on the Al agent’s performance.
- the Al agent's performance may be hidden to focus on the operator's performance.
- the debrief tools described herein may be used in various debriefing and planning situations: debriefing an operator based on training session performance (e.g., AR training, live training) debriefing an operator based on live action performance (e.g., actual mission), understanding group tendencies, evaluating operators and groups of operators (e.g., gap analysis), planning a mission and simulating outcomes (e.g., where the simulation is based in part on Al agent performance and in part on identified operator tendencies), etc.
- the debriefing tools are not limited to military or defense situations.
- the debriefing tools may be used to teach civilian or commercial operators.
- the debrief tools described herein may be used to debrief a number of training, evaluation, learning, compliance, etc. activities (e.g., flying in a formation; coordination of air assets, ground assets, water assets, space assets, jamming assets; coordination between friendly assets and enemy assets).
- the tools may be used to debrief tactile maneuvers (e.g., tactical turning, landing, following navigation guidance).
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Aviation & Aerospace Engineering (AREA)
- Theoretical Computer Science (AREA)
- Business, Economics & Management (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Entrepreneurship & Innovation (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
A method of debriefing based on a recording of a real flight of an aircraft by an operator in which said operator encountered a situation requiring a reaction by said operator, said recording comprising data of said real flight, said data comprising geospatial locations of said aircraft throughout said real flight and optionally those of one or more objects associated with said situation if applicable, flight parameters of said aircraft, and environmental conditions, and, optionally, a mission objective, said method comprising: (a) displaying a replay of at least a portion of said real flight; and (b) displaying an alternative reaction to said reaction of said operator using an AI agent, said AI agent being initialized with at least a portion of said data.
Description
OPERATOR TRAINING USING ARTIFICIAL INTELLIGENCE AGENT
CROSS REFERENCE TO RELATED APPLICATIONS
[0001] The present patent application claims the benefit of U.S. Provisional Patent Application 63/600,311, filed November 17, 2023, the entire disclosure of which is hereby incorporated herein by reference.
FIELD OF THE INVENTION
[0002] Generally, the present disclosure relates to the field of data processing. More specifically, the present disclosure relates to methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience.
BACKGROUND OF THE INVENTION
[0003] Display devices are used for various types of training, such as in simulators. Such display devices may display virtual reality and augmented reality content.
[0004] However, in some situations, movement of a display device with respect to a user using the display device may alter a perception of the content that may be displayed. For instance, due to a movement of the display device due to external forces, such as movement of display devices in flight helmets due to acceleration of aircraft, the user’s perception of the displayed content may change, which is not desired.
[0005] Therefore, there is a need for improved methods, systems, apparatuses and devices for facilitating provisioning of a virtual experience that may overcome one or more of the above-mentioned problems and/or limitations.
SUMMARY OF INVENTION
[0006] The following presents a simplified summary of the invention in order to provide a basic understanding of the invention. This summary is not an extensive overview of the invention. It is not intended to identify key/critical elements of the invention or to delineate the scope of the invention. Its sole purpose is to present some concepts of the invention in a simplified form as a prelude to the more detailed description that is presented later.
[0007] In one embodiment, the present invention relates to a method of debriefing based on a recording of a real flight of an aircraft by an operator in which the operator encountered a situation requiring a reaction by the operator. In one embodiment, the recording comprising data of the real flight, the data comprising geospatial locations of the aircraft throughout the real flight and optionally those of one or more objects associated with the situation if applicable, flight parameters of the aircraft, and environmental conditions, and, optionally, a mission objective. In one embodiment, the method comprising: (a) displaying a replay of at least a portion of the real flight; and (b) displaying an alternative reaction to the reaction of the operator using an Al agent, the Al agent being initialized with at least a portion of the data.
[0008] In one embodiment, the invention relates to a system of debriefing based on a recording of a real flight of an aircraft by an operator in which the operator encountered a situation requiring a reaction by the operator. In one embodiment, the recording comprising data of the real flight, the data comprising geospatial locations of the aircraft throughout the real flight and optionally those of one or more objects associated with the situation if applicable, flight parameters of the aircraft, and environmental conditions, and, optionally, a mission objective. In one embodiment, the system comprises: (a) a processor configured with an Al agent and the recording; (b) at least one display communicatively coupled to the processor, and configured for displaying a replay of at least a portion of the real flight, and for displaying an alternative reaction to the reaction of the operator using the Al agent, the Al agent being initialized with at least a portion of the data.
BRIEF DESCRIPTION OF THE DRAWINGS
[0009] The accompanying drawings, which are incorporated in and constitute a part of this disclosure, illustrate various embodiments of the present disclosure. The drawings contain representations of various trademarks and copyrights owned by the Applicants. In addition, the drawings may contain other marks owned by third parties and are being used for illustrative purposes only. All rights to various trademarks and copyrights represented herein, except those belonging to their respective owners, are vested in and the property of the applicants. The applicants retain and reserve all rights in their trademarks and copyrights included herein, and grant permission to reproduce the material only in connection with reproduction of the granted patent and for no other purpose.
[0010] Furthermore, the drawings may contain text or captions that may explain certain embodiments of the present disclosure. This text is included for illustrative, non-limiting, explanatory purposes of certain embodiments detailed in the present disclosure.
[0011] FIG.l is an illustration of an online platform consistent with various embodiments of the present disclosure.
[0012] FIG. 2 shows a wearable display device for facilitating provisioning of a virtual experience, in accordance with some embodiments.
[0013] FIG. 3 is a block diagram of a system for facilitating provisioning of a virtual experience in accordance with some embodiments.
[0014] FIG. 4 is a block diagram of a first head mount display for facilitating provisioning of a virtual experience in accordance with some embodiments.
[0015] FIG. 5 is a block diagram of an apparatus for facilitating provisioning of a virtual experience in accordance with some embodiments.
[0016] FIG. 6 is a flowchart of a method of facilitating provisioning of a virtual experience in accordance with some embodiments.
[0017] FIG. 7 shows a system for facilitating provisioning of a virtual experience, in accordance with some embodiments.
[0018] FIG. 8 shows a corrected augmented reality view, in accordance with some embodiments.
[0019] FIG. 9 shows an augmented reality view shown to a real pilot while a civilian aircraft is taxiing at an airport, in accordance with an exemplary embodiment.
[0020] FIG. 10 is a block diagram of a computing device for implementing the methods disclosed herein, in accordance with some embodiments.
[0021] FIG. 11 illustrates a debriefing virtual environment in accordance with the principles of the present invention
DETAIL DESCRIPTIONS OF THE INVENTION
[0022] As a preliminary matter, it will readily be understood by one having ordinary skill in the relevant art that the present disclosure has broad utility and application. As should be understood, any embodiment may incorporate only one or a plurality of the above-disclosed aspects of the disclosure and may further incorporate only one or a plurality of the above-disclosed features. Furthermore, any embodiment discussed and identified as being “preferred” is considered to be part of a best mode contemplated for carrying out the embodiments of the present disclosure. Other embodiments also may be discussed for additional illustrative purposes in providing a full and enabling disclosure. Moreover, many embodiments, such as adaptations, variations, modifications, and equivalent arrangements, will be implicitly disclosed by the embodiments described herein and fall within the scope of the present disclosure.
[0023] Accordingly, while embodiments are described herein in detail in relation to one or more embodiments, it is to be understood that this disclosure is illustrative and exemplary of the present disclosure, and are made merely for the purposes of providing a full and enabling disclosure. The detailed disclosure herein of one or more embodiments is not intended, nor is to be construed, to limit the scope of
patent protection afforded in any claim of a patent issuing here from, which scope is to be defined by the claims and the equivalents thereof. It is not intended that the scope of patent protection be defined by reading into any claim a limitation found herein that does not explicitly appear in the claim itself.
[0024] Thus, for example, any sequence(s) and/or temporal order of steps of various processes or methods that are described herein are illustrative and not restrictive. Accordingly, it should be understood that, although steps of various processes or methods may be shown and described as being in a sequence or temporal order, the steps of any such processes or methods are not limited to being carried out in any particular sequence or order, absent an indication otherwise. Indeed, the steps in such processes or methods generally may be carried out in various different sequences and orders while still falling within the scope of the present invention. Accordingly, it is intended that the scope of patent protection is to be defined by the issued claim(s) rather than the description set forth herein.
[0025] Additionally, it is important to note that each term used herein refers to that which an ordinary artisan would understand such term to mean based on the contextual use of such term herein. To the extent that the meaning of a term used herein — as understood by the ordinary artisan based on the contextual use of such term — differs in any way from any particular dictionary definition of such term, it is intended that the meaning of the term as understood by the ordinary artisan should prevail.
[0026] Furthermore, it is important to note that, as used herein, “a” and “an” each generally denotes “at least one,” but does not exclude a plurality unless the contextual use dictates otherwise. When used herein to join a list of items, “or” denotes “at least one of the items,” but does not exclude a plurality of items of the list. Finally, when used herein to join a list of items, “and” denotes “all of the items of the list.”
[0027] The following detailed description refers to the accompanying drawings. Wherever possible, the same reference numbers are used in the drawings and the following description to refer to the same or similar elements. While many embodiments of the disclosure may be described, modifications, adaptations, and
other implementations are possible. For example, substitutions, additions, or modifications may be made to the elements illustrated in the drawings, and the methods described herein may be modified by substituting, reordering, or adding stages to the disclosed methods. Accordingly, the following detailed description does not limit the disclosure. Instead, the proper scope of the disclosure is defined by the appended claims. The present disclosure contains headers. It should be understood that these headers are used as references and are not to be construed as limiting upon the subjected matter disclosed under the header.
[0028] The present disclosure includes many aspects and features. Moreover, while many aspects and features relate to, and are described in the context of facilitating provisioning of a virtual experience, embodiments of the present disclosure are not limited to use only in this context.
[0029] FIG. 1 is an illustration of an online platform 100 consistent with various embodiments of the present disclosure. By way of non-limiting example, the online platform 100 to facilitate provisioning of a virtual experience may be hosted on a centralized server 102, such as, for example, a cloud computing service. The centralized server 102 may communicate with other network entities, such as, for example, an augmented and virtual reality display device 106, a sensor system 110 of an aircraft, database 114 (such as 3D model database) over a communication network 104, such as, but not limited to, the Internet. Further, users of the online platform 100 may include relevant parties such as, but not limited to, trainees, trainers, pilots, administrators, and so on.
[0030] A user 112, such as the one or more relevant parties, may access online platform 100 through a web based software application or browser. The web based software application may be embodied as, for example, but not be limited to, a website, a web application, a desktop application, and a mobile application compatible with a computing device 1000.
[0031] FIG. 2 shows a wearable display device 200 for facilitating provisioning of a virtual experience. In some embodiments, the wearable display device 200 may be utilized in conjunction with and/or to effectuate and/or facilitate operation of any element described elsewhere herein or illustrated in any figure
herein. Further, the wearable display device 200 may include a support member 202 configured to be mounted on a user 204. Further, the support member 202 may include a structure allowing the support member 202 to be easily mountable on the user 204. For instance, the wearable display device 200 may include a head mounted device (HMD). Further, the wearable display device 200 may include a display device 206 attached to the support member 202. For instance, if the wearable display device 200 is an HMD, the HMD may include a display device in front of one eye of the user 204, (a monocular HMD), in front of both eyes of the user 204, (a binocular HMD), an optical display device (which may reflect projected images), and so on. Further, the display device 206 may be configured for displaying at least one display data. Further, the display data may include virtual reality data related to a simulation, such as a training simulation. For instance, the training simulation may correspond to vehicular racing, such as Formula 1®, and may be used by race car drivers to train for race events. Further, in an instance, the training simulation may correspond to flight training, and may be used by air force pilots for flight training in fighter aircraft. Further, in some embodiments, the display data may include augmented reality data. Accordingly, the display data may include one or more augmented reality components overlaid on top of live image. For instance, the augmented reality data may be related to flight training including a first aircraft training simultaneously with a plurality of aircrafts in different locations. Accordingly, the augmented reality data may include augmented reality components displaying the plurality of plurality of aircrafts in different locations to a display device associated with a pilot of the first aircraft. Further, the wearable display device 200 may include at least one disturbance sensor 208 configured for sensing a disturbance in a spatial relationship between the display device 206 and the user 204. Further, the spatial relationship between the display device 206 and the user 204 may include at least one of a distance and an orientation. For instance, the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the display device 206 and the eyes of the user 204.
[0032] Further, the disturbance in the spatial relationship may include a change in at least one of the distance and the orientation between the display device 206 and the user 204. Further, the disturbance in the spatial relationship may lead to an alteration in how the user 204 may view the at least one display data. For instance,
if the disturbance in the spatial relationship leads to a reduction in the distance between the display device 206 and the user 204, the user 204 may perceive one or more objects in the at least one display data to be closer. For instance, if the spatial relationship between the display device 206 and the user 204 specifies a distance of “x” centimeters, and the disturbance in the spatial relationship leads to a reduction in the distance between the display device 206 and the user 204 to “y” centimeters, the user 204 may perceive the at least one display data to be closer by “x-y” centimeters.
[0033] Further, the wearable display device 200 may include a processing device 210 communicatively coupled with the display device 206. Further, the processing device 210 may be configured for receiving the at least one display data. Further, the processing device 210 may be configured for analyzing the disturbance in the spatial relationship. Further, the processing device 210 may be configured for generating a correction data based on the analyzing. Further, the processing device 210 may be configured for generating a corrected display data based on the at least one display data and the correction data. Further, the correction data may include an instruction to shift a perspective view of the at least one display data to compensate for the disturbance in the spatial relationship between the display device 206 and the user 204. Accordingly, the correction data may be generated contrary to the disturbance in the spatial relationship. For instance, the disturbance may include an angular disturbance, wherein the display device 206 may undergo an angular displacement as a result of the angular disturbance. Accordingly, the correction data may include an instruction of translation of the display data to compensate for the angular disturbance. Further, the display data may be translated along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data.
[0034] Further, in an instance, the disturbance may include a longitudinal disturbance, wherein the display device 206 may undergo a longitudinal displacement as a result of the longitudinal displacement. Accordingly, the correction data may include an instruction of translation of the display data to compensate for the longitudinal disturbance. Further, the display data may be projected along a distance perpendicular to a line of sight of the user 204 to negate the angular displacement of the display data. For instance, the display data may be projected along a distance
perpendicular to the line of sight of the user 204 opposite to a direction of the longitudinal disturbance to compensate for the longitudinal disturbance.
[0035] Further, the support member 202 may include a head gear configured to be mounted on a head of the user 204. Further, the head gear may include a helmet configured to be worn over a crown of the head. Further, the head gear may include a shell configured to accommodate at least a part of a head of the user 204. Further, a shape of the shell may define a concavity to facilitate accommodation of at least the part of the head. Further, the shell may include an interior layer 212, an exterior layer 214 and a deformable layer 216 disposed in between the interior layer 212 and the exterior layer 214. Further, the deformable layer 216 may be configured to provide cushioning. Further, the display device 206 may be attached to at least one of the interior layer 212 and the exterior layer 214.
[0036] Further, the disturbance in the spatial relationship may be based on a deformation of the deformable layer 216 due to an acceleration of the head gear. Further, the spatial relationship may include at least one vector representing at least one position of at least one part of the display device 206 in relation to at least one eye of the user 204. Further, a vector of the at least one vector may be characterized by an orientation and a distance. For instance, the spatial relationship between the display device 206 and the user 204 may include at least one of a distance and an orientation. For instance, the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the display device 206 and the eyes of the user 204. Further, the spatial relationship may describe an optimal arrangement of the display device 206 with respect to the user 204. Further, so that the optimal arrangement of the display device 206 with respect to the user 204 may allow the user to clearly view the display data without perceived distortion.
[0037] Further, in some embodiments, the at least one disturbance sensor 208 may include an accelerometer configured for sensing the acceleration. Further, in some embodiments, the at least one disturbance sensor 208 may include at least one proximity sensor configured for sensing at least one proximity between the at least one part of the display device 206 and the user 204. Further, in some embodiments, the at least one disturbance sensor 208 may include a deformation sensor configured for sensing a deformation of the deformable layer 216.
[0038] Further, in some embodiments, the display device 206 may include a see-through display device 206 configured to allow the user 204 to view a physical surrounding of the wearable device.
[0039] Further, in some embodiments, the at least one display data may include at least one object model associated with at least one object. Further, in some embodiments, the generating of the corrected display data may include applying at least one transformation to the at least one object model based on the correction data.
[0040] Further, the applying of the at least one transformation to the at least one object model based on the correction data may include translation of the display data to compensate for the angular disturbance. For instance, the correction data may include one or more instructions to translate the display data along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data. Accordingly, the applying of the at least one transformation to the at least one object model based on the correction data may include translation of the display data along the horizontal axis, the vertical axis, and the diagonal axis of the display data, to negate the angular displacement of the display data. Further, in an instance, if the correction data includes an instruction of translation of the display data to compensate for the longitudinal disturbance, the applying of the at least one transformation to the at least one object model based on the correction data may include translation may include projection of the display data along a distance perpendicular to a line of sight of the user 204 to negate the angular displacement of the display data. For instance, the applying of the at least one transform may include projection of the display data along a distance perpendicular to the line of sight of the user 204 opposite to a direction of the longitudinal disturbance to compensate for the longitudinal disturbance.
[0041] Further, in some embodiments, the at least one disturbance sensor 208 may include a camera configured to capture an image of each of a face of the user 204 and at least a part of the head gear. Further, the spatial relationship may include disposition of at least the part of the head gear in relation to the face of the user 204.
[0042] Further, in some embodiments, the at least one disturbance sensor 208 may include a camera disposed on the display device 206. Further, the camera may be
configured to capture an image of at least a part of a face of the user 204. Further, the wearable display device 200 may include a calibration input device configured to receive a calibration input. Further, the camera may be configured to capture a reference image of at least the part of the face of the user 204 based on receiving the calibration input. Further, the calibration input may be received in an absence of the disturbance. For instance, the calibration input device may include a button configured to be pushed by the user 204 in absence of the disturbance whereupon the reference image of at least the part of the face of the user 204 may be captured. Further, the analyzing of the disturbance may include comparing the reference image with a current image of at least the part of the face of the user 204. Further, the current image may be captured by the camera in a presence of the disturbance. Further, determining the correction data may include determining at least one spatial parameter change based on the comparing. Further, the at least one spatial parameter change may correspond to at least one of a displacement of at least the part of the face relative to the camera and a rotation about at least one axis of at least the part of the face relative to the camera.
[0043] Further, in some embodiments, the generating of the corrected display data may include applying at least one image transform on the at least one display data based on the at least one spatial parameter change.
[0044] Further, in some embodiments, the wearable display device 200 may include at least one actuator coupled to the display device 206 and the support member 202. Further, the at least one actuator may be configured for modifying the spatial relationship based on a correction data.
[0045] Further, the spatial relationship between the display device 206 and the user 204 may include at least one of a distance 218 and an orientation. Further, the disturbance in the spatial relationship between the display device 206 and the user 204 may include a change in at least one of the distance 218, the angle, the direction, and the orientation. Further, the distance 218 may include a perceived distance between the user 204 and the at least one display data. For instance, as shown in FIG. 3, the disturbance in the spatial relationship may originate due to a forward acceleration 304 of the user 204 and the wearable display device 200. Accordingly, the deformation of the deformable layer 216 may lead to a disturbance in the spatial relationship leading
to a change in the distance 218 to a reduced distance 302 between the display device 206 and the user 204. Accordingly, the correction data may include transforming of the at least one display data through object level processing and restoring the at least one display data to the distance 218 from the user 204. Further, the object level processing may include projecting one or more objects in the display data at the distance 218 instead of the distance 302 to oppose the disturbance in the spatial relationship. Further, the disturbance in the spatial relationship may include a change in the angle between the display device 206 and the user 204. Further, the angle between the display device 206 and the user 204 in the spatial relationship may be related to an original viewing angle related to the display data. Further, the original viewing angle related to the display data may be a viewing angle at which the user 204 may view the display data through the display device 206. Further, the disturbance in the spatial relationship may lead to a change in the original viewing angle related to the display data. Accordingly, the at least one display data may be transformed through pixel level processing to restore the original viewing angle related to the display data. Further, the pixel level processing may include translation of the display data to compensate for the change in the angle in the spatial relationship. Further, the display data may be translated along a horizontal axis of the display data, a vertical axis of the display data, a diagonal axis of the display data, and so on, to negate the angular displacement of the display data to compensate for the change in the angle in the spatial relationship, and to restore the original viewing angle related to the display data.
[0046] FIG. 3 is a block diagram of a system 300 for facilitating provisioning of a virtual experience in accordance with some embodiments. The system 300 may include a communication device 302, a processing device 304 and a storage device 306.
[0047] The communication device 302 may be configured for receiving at least one first sensor data corresponding to at least one first sensor 310 associated with a first vehicle 308. Further, the at least one first sensor 310 may be communicatively coupled to a first transmitter 312 configured for transmitting the at least one first sensor data over a first communication channel. In some embodiments, the first vehicle 308 may be a first aircraft. Further, the first user may be a first pilot.
[0048] Further, the communication device 302 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 320 associated with a second vehicle 318. Further, the at least one second sensor 320 may be communicatively coupled to a second transmitter 322 configured for transmitting the at least one second sensor data over a second communication channel. In some embodiments, the second vehicle 818 may be a second aircraft. Further, the second user may be a second pilot.
[0049] In some embodiments, the at least one first sensor data may be received from a first On-Board-Diagnostics (OBD) system of the first vehicle 308, the at least one second sensor data may be received from a second On-Board-Diagnostics (OBD) system of the second vehicle 318.
[0050] Further, the communication device 302 may be configured for receiving at least one first presentation sensor data from at least one first presentation sensor 328 associated with the first vehicle 308. Further, the at least one first presentation sensor 328 may be communicatively coupled to the first transmitter configured for transmitting the at least one first presentation sensor data over the first communication channel. Further, in an embodiment, the at least one first presentation sensor 328 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a first spatial relationship between at least one first presentation device 314 associated with the first vehicle 308, and the first user. Further, the spatial relationship between the at least one first presentation device 314 and the first user may include at least one of a distance and an orientation. For instance, the first spatial relationship may include an exact distance, and an orientation, such as a precise angle between the at least one first presentation device 314 and the eyes of the first user. Further, the disturbance in the first spatial relationship may include a change in the at least of the distance and the orientation between the at least one first presentation device 314 and the first user.
[0051] Further, the communication device 302 may be configured for receiving at least one second presentation sensor data from at least one second presentation sensor 330 associated with the second vehicle 318.
[0052] Further, in an embodiment, the at least one second presentation sensor 330 may include a disturbance sensor configured for sensing a disturbance in a second spatial relationship between at least one second presentation device 324 associated with the second vehicle 318, and the second user.
[0053] Further, the at least one second presentation sensor 330 may be communicatively coupled to the first transmitter configured for transmitting the at least one second presentation sensor data over the second communication channel.
[0054] Further, the communication device 302 may be configured for transmitting at least one first optimized presentation data to at least one first presentation device 314 associated with the first vehicle 808. Further, in an embodiment, at least one first presentation device 314 may include a wearable display device facilitating provisioning of a virtual experience, such as the wearable display device 200. Further, in an embodiment, the at least one first optimized presentation data may include a first corrected display data generated based on a first correction data.
[0055] Further, the at least one first presentation device 314 may include a first receiver 316 configured for receiving the at least one first optimized presentation data over the first communication channel. Further, the at least one first presentation device 314 may be configured for presenting the at least one first optimized presentation data.
[0056] Further, the communication device 302 may be configured for transmitting at least one second optimized presentation data to at least one first presentation device 314 associated with the first vehicle 308. Further, the first receiver 316 may be configured for receiving the at least one second optimized presentation data over the first communication channel. Further, the at least one first presentation device 314 may be configured for presenting the at least one second optimized presentation data.
[0057] Further, in an embodiment, the at least one second optimized presentation data may include a second corrected display data generated based on a second correction data.
[0058] Further, the communication device 302 may be configured for transmitting at least one second optimized presentation data to at least one second presentation device 324 associated with the second vehicle 318. Further, the at least one second presentation device 324 may include a second receiver 326 configured for receiving the at least one second optimized presentation data over the second communication channel. Further, the at least one first presentation device 324 may be configured for presenting the at least one second optimized presentation data.
[0059] Further, the processing device 304 may be configured for analyzing the at least one first presentation sensor data associated with the first vehicle 308.
[0060] Further, the processing device 304 may be configured for analyzing the at least one second presentation sensor data associated with the second vehicle 318.
[0061] Further, the processing device 304 may be configured for generating the first correction data based on the analyzing the at least one first presentation sensor data associated with the first vehicle 308. Further, the first correction data may include an instruction to shift a perspective view of the at least one first optimized presentation data to compensate for the disturbance in the first spatial relationship between the first presentation device 314 and the first user. Accordingly, the first correction data may be generated contrary to the disturbance in the first spatial relationship. For instance, the disturbance may include an angular disturbance, wherein the first presentation device 314 may undergo an angular displacement as a result of the angular disturbance. Accordingly, the first correction data may include an instruction of translation to generate the first corrected display data included in the first optimized presentation data to compensate for the angular disturbance.
[0062] Further, the processing device 304 may be configured for generating the second correction data based on the analyzing the at least one second presentation sensor data associated with the second vehicle 318. Further, the second correction data may include an instruction to shift a perspective view of the at least one second optimized presentation data to compensate for the disturbance in the second spatial relationship between the second presentation device 324 and the second user. Accordingly, the second correction data may be generated contrary to the disturbance in the second spatial relationship. For instance, the disturbance may include an
angular disturbance, wherein the second presentation device 324 may undergo an angular displacement as a result of the angular disturbance. Accordingly, the second correction data may include an instruction of translation to generate the second corrected display data included in the second optimized presentation data to compensate for the angular disturbance.
[0063] Further, the processing device 304 may be configured for generating the at least one first optimized presentation data based on the at least one second sensor data.
[0064] Further, the processing device 304 may be configured for generating the at least one first optimized presentation data based on the at least one first presentation sensor data.
[0065] Further, the processing device 304 may be configured for generating the at least one second optimized presentation data based on the at least one first sensor data.
[0066] Further, the processing device 304 may be configured for generating the at least one second optimized presentation data based on the at least one second presentation sensor data.
[0067] Further, the storage device 306 may be configured for storing each of the at least one first optimized presentation data and the at least one second optimized presentation data.
[0068] In some embodiments, the at least one first sensor 310 may include one or more of a first orientation sensor, a first motion sensor, a first accelerometer, a first location sensor, a first speed sensor, a first vibration sensor, a first temperature sensor, a first light sensor and a first sound sensor. Further, the at least one second sensor 320 may include one or more of a second orientation sensor, a second motion sensor, a second accelerometer, a second location sensor, a second speed sensor, a second vibration sensor, a second temperature sensor, a second light sensor and a second sound sensor.
[0069] In some embodiments, the at least one first sensor 310 may be configured for sensing at least one first physical variable associated with the first vehicle 308. Further, the at least one second sensor 320 may be configured for sensing at least one second physical variable associated with the second vehicle 318. In further embodiments, the at least one first physical variable may include one or more of a first orientation, a first motion, a first acceleration, a first location, a first speed, a first vibration, a first temperature, a first light intensity and a first sound. Further, the at least one second physical variable may include one or more of a second orientation, a second motion, a second acceleration, a second location, a second speed, a second vibration, a second temperature, a second light intensity and a second sound.
[0070] In some embodiments, the at least one first sensor 310 may include a first environmental sensor configured for sensing a first environmental variable associated with the first vehicle 308. Further, the at least one second sensor 320 may include a second environmental sensor configured for sensing a second environmental variable associated with the second vehicle 318.
[0071] In some embodiments, the at least one first sensor 310 may include a first user sensor configured for sensing a first user variable associated with a first user of the first vehicle 308. Further, the at least one second sensor 320 may include a second user sensor configured for sensing a second user variable associated with a second user of the second vehicle 318.
[0072] In further embodiments, the first user variable may include a first user location and a first user orientation. Further, the second user variable may include a second user location and a second user orientation. Further, the first presentation device may include a first head mount display. Further, the second presentation device may include a second head mount display.
[0073] In further embodiments, the first head mount display may include a first user location sensor of the at least one first sensor 310 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 310 configured for sensing the first user orientation. Further, the second head mount display may include a second user location sensor of the at least one second sensor 320 configured for sensing the second user location, a second user orientation sensor
of the at least one second sensor 320 configured for sensing the second user orientation.
[0074] In further embodiments, the first vehicle 308 may include a first user location sensor of the at least one first sensor 310 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 310 configured for sensing the first user orientation. Further, the second vehicle 318 may include a second user location sensor of the at least one second sensor 320 configured for sensing the second user location, a second user orientation sensor of the at least one second sensor 320 configured for sensing the second user orientation.
[0075] In further embodiments, the first user orientation sensor may include a first gaze sensor configured for sensing a first eye gaze of the first user. Further, the second user orientation sensor may include a second gaze sensor configured for sensing a second eye gaze of the second user.
[0076] In further embodiments, the first user location sensor may include a first proximity sensor configured for sensing the first user location in relation to the at least one first presentation device 314. Further, the second user location sensor may include a second proximity sensor configured for sensing the second user location in relation to the at least one second presentation device 324.
[0077] Further, in some embodiments, the at least one first presentation sensor 328 may include at least one sensor configured for sensing at least one first physical variable associated with the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308. For instance, the at least one first presentation sensor 328 may include at least one camera configured to monitor a movement of the first presentation device 314 associated with the first vehicle 308. Further, the at least one first presentation sensor 328 may include at least one accelerometer sensor configured to monitor an uneven movement of the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308. Further, the at least one first presentation sensor 328 may include at least one gyroscope sensor configured to monitor an uneven orientation of the first presentation device 314 associated with the first vehicle 308,
such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
[0078] Further, the at least one second presentation sensor 330 may include at least one sensor configured for sensing at least one first physical variable associated with the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318. For instance, the at least one second presentation sensor 330 may include at least one camera configured to monitor a movement of the second presentation device 324 associated with the second vehicle 318. Further, the at least one second presentation sensor 330 may include at least one accelerometer sensor configured to monitor an uneven movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318. Further, the at least one second presentation sensor 330 may include at least one gyroscope sensor configured to monitor an uneven orientation of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
[0079] In some embodiments, the first head mount display may include a first see-through display device. Further, the second head mount display may include a second see-through display device.
[0080] In some embodiments, the first head mount display may include a first optical marker configured to facilitate determination of one or more of the first user location and the first user orientation. Further, the at least one first sensor 310 may include a first camera configured for capturing a first image of the first optical marker. Further, the at least one first sensor 310 may be communicatively coupled to a first processor associated with the vehicle. Further, the first processor may be configured for determining one or more of the first user location and the first user orientation based on analysis of the first image. Further, the second head mount display may include a second optical marker configured to facilitate determination of one or more of the second user location and the second user orientation. Further, the at least one second sensor 320 may include a second camera configured for capturing a second image of the second optical marker. Further, the at least one second sensor
320 may be communicatively coupled to a second processor associated with the vehicle. Further, the second processor may be configured for determining one or more of the second user location and the second user orientation based on analysis of the second image.
[0081] In some embodiments, the first presentation device may include a first see-through display device disposed in a first windshield of the first vehicle 308. Further, the second presentation device may include a second see-through display device disposed in a second windshield of the second vehicle 318.
[0082] In some embodiments, the first vehicle 308 may include a first watercraft, a first land vehicle, a first aircraft and a first amphibious vehicle. Further, the second vehicle 318 may include a second watercraft, a second land vehicle, a second aircraft and a second amphibious vehicle.
[0083] In some embodiments, the at least one may include one or more of a first visual data, a first audio data and a first haptic data. Further, the at least one second optimized presentation data may include one or more of a second visual data, a second audio data and a second haptic data.
[0084] In some embodiments, the at least one first presentation device 314 may include at least one environmental variable actuator configured for controlling at least one first environmental variable associated with the first vehicle 308 based on the first optimized presentation data. Further, the at least one second presentation device 324 may include at least one environmental variable actuator configured for controlling at least one second environmental variable associated with the second vehicle 318 based on the second optimized presentation data. In further embodiments, the at least one first environmental variable may include one or more of a first temperature level, a first humidity level, a first pressure level, a first oxygen level, a first ambient light, a first ambient sound, a first vibration level, a first turbulence, a first motion, a first speed, a first orientation and a first acceleration, the at least one second environmental variable may include one or more of a second temperature level, a second humidity level, a second pressure level, a second oxygen level, a second ambient light, a second ambient sound, a second vibration level, a second
turbulence, a second motion, a second speed, a second orientation and a second acceleration.
[0085] In some embodiments, the first vehicle 308 may include each of the at least one first sensor 310 and the at least one first presentation device 314. Further, the second vehicle 318 may include each of the at least one second sensor 320 and the at least one second presentation device 324.
[0086] In some embodiments, the storage device 306 may be further configured for storing a first three-dimensional model corresponding to the first vehicle 308 and a second three-dimensional model corresponding to the second vehicle 318. Further, the generating of the first optimized presentation data may be based further on the second three-dimensional model. Further, the generating of the second optimized presentation data may be based further on the first three- dimensional model.
[0087] Further, the generating of the first optimized presentation data may be based on the determining of the unwanted movement of the associated with the first presentation device 314 associated with the first vehicle 308, such as due to a G- Force, a frictional force, and an uneven movement of the first vehicle 308. For instance, the at least one first presentation sensor 328 may include at least one camera configured to monitor a movement of the first presentation device 314 associated with the first vehicle 308. Further, the at least one first presentation sensor 328 may include at least one accelerometer sensor configured to monitor an uneven movement of the first presentation device 314 associated with the first vehicle 308, such as due to a G- Force, a frictional force, and an uneven movement of the first vehicle 308. Further, the at least one first presentation sensor 328 may include at least one gyroscope sensor configured to monitor an uneven orientation of the first presentation device 314 associated with the first vehicle 308, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle 308.
[0088] Further, the generating of the second optimized presentation data may be based on the determining of the unwanted movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318. For instance, the
at least one second presentation sensor 330 may include at least one camera configured to monitor a movement of the second presentation device 324 associated with the second vehicle 318. Further, the at least one second presentation sensor 330 may include at least one accelerometer sensor configured to monitor an uneven movement of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318. Further, the at least one second presentation sensor 330 may include at least one gyroscope sensor configured to monitor an uneven orientation of the second presentation device 324 associated with the second vehicle 318, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle 318.
[0089] In some embodiments, the communication device 302 may be further configured for receiving an administrator command from an administrator device. Further, the generating of one or more of the first optimized presentation data and the second optimized presentation data may be based further on the administrator command. In further embodiments, the at least one first presentation model may include at least one first virtual object model corresponding to at least one first virtual object. Further, the at least one second presentation model may include at least one second virtual object model corresponding to at least one second virtual object. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor model. Further, the generating of the at least one second virtual object model may be independent of the at least one first sensor model. Further, the generating of one or more of the at least one first virtual object model and the at least one second virtual object model may be based on the administrator command. Further, the storage device 806 may be configured for storing the at least one first virtual object model and the at least one second virtual object model.
[0090] In further embodiments, the administrator command may include a virtual distance parameter. Further, the generating of each of the at least one first optimized presentation data and the at least one second optimized presentation data may be based on the virtual distance parameter.
[0091] In further embodiments, the at least one first sensor data may include at least one first proximity data corresponding to at least one first external real object in a vicinity of the first vehicle 308. Further, the at least one second sensor data may include at least one second proximity data corresponding to at least one second external real object in a vicinity of the second vehicle 318. Further, the generating of the at least one first optimized presentation data may be based further on the at least one second proximity data. Further, the generating of the at least one second optimized presentation data may be based further on the at least one first proximity data. In further embodiments, the at least one first external real object may include a first cloud, a first landscape feature, a first man-made structure and a first natural object. Further, the at least one second external real object may include a second cloud, a second landscape feature, a second man-made structure and a second natural object.
[0092] In some embodiments, the at least one first sensor data may include at least one first image data corresponding to at least one first external real object in a vicinity of the first vehicle 308. Further, the at least one second sensor data may include at least one second image data corresponding to at least one second external real object in a vicinity of the second vehicle 318. Further, the generating of the at least one first optimized presentation data may be based further on the at least one second image data. Further, the generating of the at least one second optimized presentation data may be based further on the at least one first image data.
[0093] In some embodiments, the communication device 302 may be further configured for transmitting a server authentication data to the first receiver 316. Further, the first receiver 316 may be communicatively coupled to first processor associated with the first presentation device. Further, the first processor may be communicatively coupled to a first memory device configured to store a first authentication data. Further, the first processor may be configured for performing a first server authentication based on the first authentication data and the server authentication data. Further, the first processor may be configured for controlling presentation of the at least one first optimized presentation data on the at least one first presentation device 314 based on the first server authentication. Further, the communication device 302 may be configured for transmitting a server authentication
data to the second receiver 326. Further, the second receiver 326 may be communicatively coupled to second processor associated with the second presentation device. Further, the second processor may be communicatively coupled to a second memory device configured to store a second authentication data. Further, the second processor may be configured for performing a second server authentication based on the second authentication data and the server authentication data. Further, the second processor may be configured for controlling presentation of the at least one second optimized presentation data on the at least one second presentation device 324 based on the second server authentication. Further, the communication device 302 may be configured for receiving a first client authentication data from the first transmitter 312. Further, the storage device 306 may be configured for storing the first authentication data. Further, the communication device 302 may be configured for and receiving a second client authentication data from the second transmitter 322. Further, the storage device 306 may be configured for storing the second authentication data. Further, the processing device 304 may be further configured for performing a first client authentication based on the first client authentication data and the first authentication data. Further, the generating of the at least one second optimized presentation data may be further based on the first client authentication. Further, the processing device 304 may be configured for performing a second client authentication based on the second client authentication data and the second authentication data. Further, the generating of the at least one first optimized presentation data may be further based on the second client authentication.
[0094] FIG. 4 is a block diagram of a first head mount display 400 for facilitating provisioning of a virtual experience in accordance with some embodiments. The first head mount display 400 may include a first user location sensor 402 of the at least one first sensor configured for sensing the first user location and a first user orientation sensor 404 of the at least one first sensor configured for sensing the first user orientation.
[0095] Further, the first head mount display 400 may include a display device 406 to present visuals. Further, in an embodiment, the display device 406 may be configured for displaying the first optimized display data, as generated by the processing device 404.
[0096] Further, the first head mount display 400 may include a processing device 408 configured to obtain sensor data from the first user location sensor 402 and the first user orientation sensor 404. Further, the processing device 408 may be configured to send visuals to the display device 406.
[0097] FIG. 5 is a block diagram of an apparatus 500 for facilitating provisioning of a virtual experience in accordance with some embodiments. The apparatus 500 may include at least one first sensor 502 (such as the at least one first sensor 310) configured for sensing at least one first sensor data associated with a first vehicle (such as the first vehicle 308).
[0098] Further, the apparatus 500 may include at least one first presentation sensor 510 (such as the at least one first presentation sensor 328) configured for sensing at least one first presentation sensor data associated with a first vehicle (such as the first vehicle 308). Further, in an embodiment, the at least one first presentation sensor 510 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a first spatial relationship between at least one first presentation device 508 associated with the first vehicle, and a first user. Further, the spatial relationship between the at least one first presentation device 508 and the first user may include at least one of a distance and an orientation. For instance, the first spatial relationship may include an exact distance, and an orientation, such as a precise angle between the at least one first presentation device 508 and the eyes of the first user. Further, the disturbance in the first spatial relationship may include a change in the at least of the distance and the orientation between the at least one first presentation device 314 and the first user.
[0099] Further, the apparatus 500 may include a first transmitter 504 (such as the first transmitter 312) configured to be communicatively coupled to the at least first sensor 502, and the at least one first presentation sensor 510. Further, the first transmitter 504 may be configured for transmitting the at least one first sensor data and the at least one first presentation sensor data to a communication device (such as the communication device 302) of a system over a first communication channel.
[00100] Further, the apparatus 500 may include a first receiver 506 (such as the first receiver 316) configured for receiving the at least one first optimized presentation data from the communication device over the first communication channel.
[00101] Further, the apparatus 500 may include the at least one first presentation device 508 (such as the at least one first presentation device 314) configured to be communicatively coupled to the first receiver 506. The at least one first presentation device 508 may be configured for presenting the at last one first optimized presentation data.
[00102] Further, the communication device may be configured for receiving at least one second sensor data corresponding to at least one second sensor (such as the at least one second sensor 320) associated with a second vehicle (such as the second vehicle 318). Further, the at least one second sensor may be communicatively coupled to a second transmitter (such as the second transmitter 322) configured for transmitting the at least one second sensor data over a second communication channel. Further, the system may include a processing device (such as the processing device 304) communicatively coupled to the communication device. Further, the processing device may be configured for generating the at least one first optimized presentation data based on the at least one second sensor data.
[00103] FIG. 6 is a flowchart of a method 600 of facilitating provisioning of a virtual experience in accordance with some embodiments. At 602, the method 600 may include receiving, using a communication device (such as the communication device 302), at least one first sensor data corresponding to at least one first sensor (such as the at least one first sensor 310) associated with a first vehicle (such as the first vehicle 308). Further, the at least one first sensor may be communicatively coupled to a first transmitter (such as the first transmitter 312) configured for transmitting the at least one first sensor data over a first communication channel.
[00104] At 604, the method 600 may include receiving, using the communication device, at least one second sensor data corresponding to at least one second sensor (such as the at least one second sensor 320) associated with a second vehicle (such as the second vehicle 318). Further, the at least one second sensor may be communicatively coupled to a second transmitter (such as the second transmitter
322) configured for transmitting the at least one second sensor data over a second communication channel.
[00105] At 606, the method 600 may include receiving, using the communication device, a first presentation sensor data corresponding to at least one first presentation sensor 328 associated with the first vehicle. Further, the at least one first presentation sensor may be communicatively coupled to the first transmitter configured for transmitting the at least one first presentation sensor data over the first communication channel. Further, the first presentation sensor may include at least one sensor configured to monitor a movement of at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. For instance, the at least one first presentation sensor may include at least one camera configured to monitor a movement of the at least one first presentation device associated with the first vehicle. Further, the at least one first presentation sensor may include at least one accelerometer sensor configured to monitor an uneven movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. Further, the at least one first presentation sensor may include at least one gyroscope sensor configured to monitor an uneven orientation of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle.
[00106] At 608, the method 600 may include receiving, using the communication device, a second presentation sensor data corresponding to at least one second presentation sensor 330 associated with the second vehicle. Further, the at least one second presentation sensor may be communicatively coupled to the second transmitter configured for transmitting the at least one second presentation sensor data over the second communication channel. Further, the second presentation sensor may include at least one sensor configured to monitor a movement of at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. For instance, the at least one second presentation sensor may include at least one camera configured to monitor a movement of the at least one second presentation device associated with the second vehicle. Further, the at least one second presentation sensor may include at
least one accelerometer sensor configured to monitor an uneven movement of the at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. Further, the at least one second presentation sensor may include at least one gyroscope sensor configured to monitor an uneven orientation of the at least one second presentation device associated with the second vehicle, such as due to a G- Force, a frictional force, and an uneven movement of the second vehicle.
[00107] At 610, the method 600 may include analyzing, using a processing device, the at least one first sensor data and the at least one first presentation sensor data to generate at least one first modified presentation data. The analyzing may include determining an unwanted movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. Further, the unwanted movement of the at least one first presentation device associated with the first vehicle may include an upward movement, a downward movement, a leftward movement, and a rightward movement. Further, the generating of the at least one first optimized presentation data may be based on the unwanted movement of the at least one first presentation device associated with the first vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the first vehicle. For instance, the generating of the at least one first optimized presentation data may be based on negating an effect of the unwanted movement of the at least one first presentation device associated with the first vehicle. For instance, if the unwanted movement of the at least one first presentation device associated with the first vehicle includes an upward movement, a downward movement, a leftward movement, and a rightward movement, the generating of the at least one first optimized presentation data may include moving one or more components of the at least one first modified presentation data in an oppositely downward direction, an upward direction, a rightward direction, and a leftward direction respectively.
[00108] At 612, the method 600 may include analyzing, using a processing device, the at least one second sensor data and the at least one second presentation sensor data to generate at least one second presentation data. The analyzing may include determining an unwanted movement of the at least one second presentation
device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. Further, the unwanted movement of the at least one second presentation device associated with the second vehicle may include an upward movement, a downward movement, a leftward movement, and a rightward movement. Further, the generating of the at least one second optimized presentation data may be based on the unwanted movement of the at least one second presentation device associated with the second vehicle, such as due to a G-Force, a frictional force, and an uneven movement of the second vehicle. For instance, the generating of the at least one second optimized presentation data may be based on negating an effect of the unwanted movement of the at least one second presentation device associated with the second vehicle. For instance, if the unwanted movement of the at least one second presentation device associated with the second vehicle includes an upward movement, a downward movement, a leftward movement, and a rightward movement, the generating of the at least one second optimized presentation data may include moving one or more components of the at least one second presentation data in an oppositely downward direction, an upward direction, a rightward direction, and a leftward direction respectively.
[00109] At 614, the method 600 may include transmitting, using the communication device, at least one first optimized presentation data to at least one first presentation device associated with the first vehicle. Further, the at least one first presentation device may include a first receiver (such as the first receiver 316) configured for receiving the at least one first modified presentation data over the first communication channel. Further, the at least one presentation device may be configured for presenting the at least one first optimized presentation data.
[00110] At 616, the method 600 may include transmitting, using the communication device, at least one second optimized presentation data to at least one second presentation device (such as the at least one second presentation device 324) associated with the second vehicle. Further, the at least one second presentation device may include a second receiver (such as the second receiver 326) configured for receiving the at least one second presentation data over the second communication channel. Further, the at least one presentation device may be configured for presenting the at least one second optimized presentation data.
[00111] At 618, the method 600 may include storing, using a storage device (such as the storage device 306), each of the at least one first optimized presentation data and the at least one second optimized presentation data.
[00112] FIG. 7 shows a system 700 for facilitating provisioning of a virtual experience, in accordance with some embodiments. The system 700 may include a communication device 702 configured for receiving at least one first sensor data corresponding to at least one first sensor 710 associated with a first vehicle 708. Further, the at least one first sensor 710 may be communicatively coupled to a first transmitter 712 configured for transmitting the at least one first sensor data over a first communication channel.
[00113] Further, the communication device 702 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 716 associated with a second vehicle 714. Further, the at least one second sensor 716 may include a second location sensor configured to detect a second location associated with the second vehicle 714. Further, the at least one second sensor 716 may be communicatively coupled to a second transmitter 718 configured for transmitting the at least one second sensor data over a second communication channel. Further, in some embodiments, the at least one second sensor 716 may include a second user sensor configured for sensing a second user variable associated with a second user of the second vehicle 714. Further, the second user variable may include a second user location and a second user orientation.
[00114] Further, in some embodiments, the at least one second sensor 716 may include a disturbance sensor, such as the disturbance sensor 208 configured for sensing a disturbance in a spatial relationship between a second presentation device 720 associated with the second vehicle 714 and the second user of the second vehicle 714. Further, the spatial relationship between the second presentation device 720 and the second user may include at least one of a distance and an orientation. For instance, the spatial relationship may include an exact distance, and an orientation, such as a precise angle between the second presentation device 720 and the eyes of the second user.
[00115] Further, the disturbance in the spatial relationship may include a change in the at least of the distance and the orientation between the second presentation device 720 and the second user. Further, the disturbance in the spatial relationship may lead to an alteration in how the second user may view at least one second presentation data. For instance, if the disturbance in the spatial relationship leads to a reduction in the distance between the second presentation device 720 and the second user, the second user may perceive one or more objects in the at least one second presentation data to be closer. For instance, if the spatial relationship between the second presentation device 720 and the second user specifies a distance of “x” centimeters, and the disturbance in the spatial relationship leads to a reduction in the distance between the second presentation device 720 and the second user to “y” centimeters, the second user may perceive the at least one second presentation data to be closer by “x-y” centimeters.
[00116] Further, the communication device 702 may be configured for transmitting the at least one second presentation data to the at least one second presentation device 720 associated with the second vehicle 714. Further, the at least one second presentation data may include at least one second virtual object model corresponding to at least one second virtual object. Further, in some embodiments, the at least one second virtual object may include one or more of a navigational marker and an air-corridor.
[00117] Further, in an embodiment, the at least one second presentation data may include a second corrected display data generated based on a second correction data. Further, the at least one second presentation device 720 may include a second receiver 722 configured for receiving the at least one second presentation data over the second communication channel. Further, the at least one second presentation device 720 may be configured for presenting the at least one second presentation data. Further, in some embodiments, the at least one second presentation device 720 may include a second head mount display. Further, the second head mount display may include a second user location sensor of the at least one second sensor 716 configured for sensing the second user location and a second user orientation sensor of the at least one second sensor 716 configured for sensing the second user orientation.
Further, the second head mount display may include a second see-through display device.
[00118] Further, in some embodiments, the at least one second virtual object model may include a corrected augmented reality view, such as the corrected augmented reality view 800. Further, the augmented reality view 800 may include one or more second virtual objects such as a navigational marker 808, and a skyway 806 as shown in FIG. 8).
[00119] Further, the system 700 may include a processing device 704 configured for generating the at least one second presentation data based on the at least one first sensor data and the at least one second sensor data. Further, the generating of the at least one second virtual object model may be independent of the at least one first sensor data. Further, in some embodiments, the processing device 704 may be configured for determining a second airspace class associated with the second vehicle 714 based on the second location including a second altitude associated with the second vehicle 714. Further, the generating of the at least one second virtual object model may be based on the second airspace class.
[00120] Further, the processing device 704 may be configured for generating the second correction data based on the analyzing the at least one second sensor data associated with the second vehicle 714. Further, the second correction data may include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user. Accordingly, the second correction data may be generated contrary to the disturbance in the spatial relationship. For instance, the disturbance may include an angular disturbance, wherein the second presentation device 720 may undergo an angular displacement as a result of the angular disturbance. Accordingly, the second correction data may include an instruction of translation to generate the second corrected display data included in the second presentation data to compensate for the angular disturbance.
[00121] For instance, if the at least one second presentation data includes the at least one second virtual object model may include a corrected augmented reality view, such as the corrected augmented reality view 800, the second correction data may
include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user (such as a pilot 802). For instance, if the disturbance in the spatial relationship includes a reduction in the distance between the second presentation device 720, the second correction data may include an instruction to shift a perspective view of the at least one second presentation data to compensate for the disturbance in the spatial relationship between the second presentation device 720 and the second user, such as by projection of the one or more second virtual objects, such as the navigational marker 808, and the skyway 806 at a distance to compensate the disturbance and to generate the corrected augmented reality view 800.
[00122] Further, the system 700 may include a storage device 706 configured for storing the at least one second presentation data. Further, in some embodiments, the storage device 706 may be configured for retrieving the at least one second virtual object model based on the second location associated with the second vehicle 714. Further, in some embodiments, the storage device 706 may be configured for storing a first three-dimensional model corresponding to the first vehicle 708. Further, the generating of the second presentation data may be based on the first three-dimensional model.
[00123] Further, in some embodiments, the communication device 702 may be configured for receiving an administrator command from an administrator device. Further, the generating of the at least one second virtual object model may be based on the administrator command.
[00124] Further, in some embodiments, the communication device 702 may be configured for transmitting at least one first presentation data to at least one first presentation device (not shown) associated with the first vehicle 708. Further, the at least one first presentation device may include a first receiver configured for receiving the at least one first presentation data over the first communication channel. Further, the at least one first presentation device may be configured for presenting the at least one first presentation data. Further, in some embodiments, the processing device 704 may be configured for generating the at least one first presentation data based on the at least one second sensor data. Further, in some embodiments, the storage device 706
may be configured for storing the at least one first presentation data. Further, in some embodiments, the storage device 706 may be configured for storing a second three- dimensional model corresponding to the second vehicle 714. Further, the generating of the first presentation data may be based on the second three-dimensional model.
[00125] Further, in some embodiments, the at least one first presentation data may include at least one first virtual object model corresponding to at least one first virtual object. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor data. Further, the storage device 706 may be configured for storing the at least one first virtual object model.
[00126] Further, in some exemplary embodiment, the communication device 702 may be configured for receiving at least one second sensor data corresponding to at least one second sensor 716 associated with a second vehicle 714. Further, the at least one second sensor 716 may be communicatively coupled to a second transmitter 718 configured for transmitting the at least one second sensor data over a second communication channel. Further, the communication device 702 may be configured for receiving at least one first sensor data corresponding to at least one first sensor 710 associated with a first vehicle 708. Further, the at least one first sensor 710 may include a first location sensor configured to detect a first location associated with the first vehicle 708. Further, the at least one first sensor 710 may be communicatively coupled to a first transmitter 712 configured for transmitting the at least one first sensor data over a first communication channel. Further, in some embodiments, the at least one first sensor 710 may include a first user sensor configured for sensing a first user variable associated with a first user of the first vehicle 708. Further, the first user variable may include a first user location and a first user orientation. Further, the communication device 702 configured for transmitting at least one first presentation data to at least one first presentation device (not shown) associated with the first vehicle 708. Further, the at least one first presentation data may include at least one first virtual object model corresponding to at least one first virtual object. Further, in some embodiments, the at least one first virtual object may include one or more of a navigational marker (such as a navigational marker 708, and/or a signboard 904 as shown in FIG. 9) and an air-corridor (such as a skyway 806 as shown in FIG. 8). Further, the at least one first presentation device may include a first receiver
configured for receiving the at least one first presentation data over the first communication channel. Further, the at least one first presentation device may be configured for presenting the at least one first presentation data. Further, in some embodiments, the at least one first presentation device may include a first head mount display. Further, the first head mount display may include a first user location sensor of the at least one first sensor 710 configured for sensing the first user location and a first user orientation sensor of the at least one first sensor 710 configured for sensing the first user orientation. Further, the first head mount display may include a first see- through display device. Further, the processing device 704 may be configured for generating the at least one first presentation data based on the at least one second sensor data and the at least one first sensor data. Further, the generating of the at least one first virtual object model may be independent of the at least one second sensor data. Further, in some embodiments, the processing device 704 may be configured for determining a first airspace class associated with the first vehicle 708 based on the first location including a first altitude associated with the first vehicle 708. Further, the generating of the at least one first virtual object model may be based on the first airspace class. Further, in some embodiments, the storage device 706 may be configured for storing the at least one first presentation data. Further, in some embodiments, the storage device 706 may be configured for retrieving the at least one first virtual object model based on the first location associated with the first vehicle 708. Further, in some embodiments, the storage device 706 may be configured for storing a second three-dimensional model corresponding to the second vehicle 714. Further, the generating of the first presentation data may be based on the second three-dimensional model. Further, in some embodiments, the communication device 702 may be configured for receiving an administrator command from an administrator device. Further, the generating of the at least one first virtual object model may be based on the administrator command. Further, in some embodiments, the communication device 702 may be configured for transmitting at least one second presentation data to at least one second presentation device (such as the second presentation device 720) associated with the second vehicle 714. Further, the at least one second presentation device may include a second receiver (such as the second receiver 722) configured for receiving the at least one second presentation data over the second communication channel. Further, the at least one second presentation device may be configured for presenting the at least one second presentation data.
Further, in some embodiments, the processing device 704 may be configured for generating the at least one second presentation data based on the at least one first sensor data. Further, in some embodiments, the storage device 706 may be configured for storing the at least one second presentation data. Further, in some embodiments, the storage device 706 may be configured for storing a first three-dimensional model corresponding to the first vehicle 708. Further, the generating of the second presentation data may be based on the first three-dimensional model. Further, in some embodiments, the at least one second presentation data may include at least one second virtual object model corresponding to at least one second virtual object.
Further, the generating of the at least one second virtual object model may be independent of the at least one first sensor data. Further, the storage device 706 may be configured for storing the at least one second virtual object model.
[00127] FIG. 8 shows the corrected augmented reality view 800. Further, the augmented reality view 800 may include a road drawn in the sky (such as the skyway 806) indicating a path that a civilian aircraft 804 may take in order to land at an airport. Further, the augmented reality view 800 may include the navigation marker 808 indicating to a pilot 802 that the civilian aircraft 804 should take a left turn. The navigation marker 808 may assist the pilot 802 in navigating towards a landing strip to land the civilian aircraft 804.
[00128] Therefore, the corrected augmented reality view 800 may provide pilots with a similar view as seen by public transport drivers (e.g. taxi or bus) on the ground. The pilots (such as the pilot 802) may see roads (such as the skyway 806) that the pilot 802 need to drive on. Further, the pilot 802, in an instance, may see signs just like a taxi driver who may just look out of a window and see road signs.
[00129] Further, the corrected augmented reality view 800 may include (but not limited to) one or more of skyways (such the skyway 806), navigation markers (such as the navigation marker 808), virtual tunnels, weather information, an air corridor, speed, signboards for precautions, airspace class, one or more parameters shown on a conventional horizontal situation indicator (HSI) etc. The skyways may indicate a path that an aircraft (such as the civilian aircraft 804) should take. The skyways may appear similar to roads on the ground. The navigation markers may be similar to regulatory road signs used on the roads on the ground. Further, the navigation markers
may instruct pilots (such as the pilot 802) on what they must or should do (or not do) under a given set of circumstances. Further, the navigation markers may be used to reinforce air-traffic laws, regulations or requirements which apply either at all times or at specified times or places upon a flight path. For example, the navigation markers may include one or more of a left curve ahead sign, a right curve ahead sign, a keep left sign, and a keep to right sign. Further, the virtual tunnels may appear similar to tunnels on roads on the ground. The pilot 802 may be required to fly the aircraft through the virtual tunnel. Further, the weather information may include real-time weather data that affects flying conditions. For example, the weather information may include information related to one or more of wind speed, gust, and direction; variable wind direction; visibility, and variable visibility; temperature; precipitation; and cloud cover. Further, the air corridor may indicate an air route along which the aircraft is allowed to fly, especially when the aircraft is over a foreign country. Further, the corrected augmented reality view 800 may include speed information. The speed information may include one or more of a current speed, a ground speed, and a recommended speed. The signboards for precautions may be related to warnings shown to the pilot 802. The one or more parameters shown on a conventional horizontal situation indicator (HSI) include NAV warning flag, lubber line, compass warning flag, course select pointer, TO/FROM indicator, glideslope deviation scale, heading select knob, compass card, course deviation scale, course select knob, course deviation bar (CD I), symbolic aircraft, dual glideslope pointers, and heading select bug.
[00130] Further, in some embodiments, information such as altitude, attitude, airspeed, the rate of climb, heading, autopilot and auto-throttle engagement status, flight director modes and approach status etc. that may be displayed on a conventional primary flight display may also be displayed in the corrected augmented reality view 800.
[00131] Further, in some embodiments, the corrected augmented reality view 800 may include a one or more of other vehicles (such as another airplane 810). Further, the one or more other vehicles, in an instance, may include one or more live vehicles (such as representing real pilots flying real aircraft), one or more virtual vehicles (such as representing real people on the ground, flying virtual aircraft), and
one or more constructed vehicles (such as representing aircraft generated and controlled using computer graphics and processing systems).
[00132] In some embodiments, a special use airspace class may be determined. The special use airspace class may include alert areas, warning areas, restricted areas, prohibited airspace, military operation area, national security area, controlled firing areas etc. For an instance, if an aircraft (such as the civilian aircraft 804) enters a prohibited area by mistake, then a notification may be displayed in the corrected augmented reality view 800. Accordingly, the pilot 802 may reroute the aircraft towards a permitted airspace.
[00133] Further, the corrected augmented reality view 800 may include one or more live aircraft (representing real pilots flying real aircraft), one or more virtual aircraft (representing real people on the ground, flying virtual aircraft) and one or more constructed aircraft (representing aircraft generated and controlled using computer graphics and processing systems). Further, the corrected augmented reality view 800 shown to a pilot (such as the pilot 802) in a first aircraft (such as the civilian aircraft 804) may be modified based on sensor data received from another aircraft (such as another airplane 1410). The sensor data may include data received from one or more internal sensors to track and localize the pilot's head within the cockpit of the aircraft. Further, the sensor data may include data received from one or more external sensors to track the position and orientation of the aircraft. Further, the data received from the one or more internal sensors and the one or more external sensors may be combined to provide a highly usable augmented reality solution in a fast-moving environment.
[00134] FIG. 9 shows an augmented reality view 900 shown to a real pilot while a civilian aircraft 902 is taxiing at an airport, in accordance with an exemplary embodiment. The augmented reality view 900 may include one or more navigational markers (such as the navigation marker 808) and signboards (such as a signboard 904) that assist a pilot to taxi the civilian aircraft 902 at the airport. The navigational markers may indicate the direction of movement. The signboards may indicate the speed limits.
[00135] The augmented reality view 900 may help the pilot to taxi the civilian aircraft 902 towards a parking location after landing. Further, augmented reality view 900 may help the pilot to taxi the civilian aircraft 902 towards a runway for taking- off Therefore, a ground crew may no longer be required to instruct the pilot while taxiing the civilian aircraft 902 at the airport.
[00136] Further, the augmented reality view 900 may include one or more live aircraft (such as a live aircraft 906) at the airport (representing real pilots in real aircraft), one or more virtual aircraft at the airport (representing real people on the ground, controlling a virtual aircraft) and one or more constructed aircraft at the airport (representing aircraft generated and controlled using computer graphics and processing systems). Further, the augmented reality view 900 shown to a pilot in a first aircraft may be modified based on sensor data received from another aircraft. The sensor data may include data received from one or more internal sensors to track and localize the pilot's head within the cockpit of the aircraft. Further, the sensor data may include data received from one or more external sensors to track the position and orientation of the aircraft. Further, the data received from the one or more internal sensors and the one or more external sensors may be combined to provide a highly usable augmented reality solution in a fast-moving environment.
[00137] In accordance with exemplary and non-limiting embodiments, the process of acquiring sensor information from one or more vehicles, maintaining a repository of data describing various real and virtual platforms and environments, and generating presentation data may be distributed among various platforms and among a plurality of processors.
[00138] With reference to FIG. 10, a system consistent with an embodiment of the disclosure may include a computing device or cloud service, such as computing device 1000. In a basic configuration, computing device 1000 may include at least one processing unit 1002 and a system memory 1004. Depending on the configuration and type of computing device, system memory 1004 may include, but is not limited to, volatile (e.g. random-access memory (RAM)), non-volatile (e.g. read-only memory (ROM)), flash memory, or any combination. System memory 1004 may include operating system 1005, one or more programming modules 1006, and may include a program data 1007. Operating system 1005, for example, may be suitable
for controlling computing device 1000’s operation. In one embodiment, programming modules 1006 may include image-processing module, machine learning module and/or image classifying module. Furthermore, embodiments of the disclosure may be practiced in conjunction with a graphics library, other operating systems, or any other application program and is not limited to any particular application or system. This basic configuration is illustrated in FIG. 10 by those components within a dashed line 1008.
[00139] Computing device 1000 may have additional features or functionality. For example, computing device 1000 may also include additional data storage devices (removable and/or non-removable) such as, for example, magnetic disks, optical disks, or tape. Such additional storage is illustrated in FIG. 10 by a removable storage 1009 and a non-removable storage 1010. Computer storage media may include volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information, such as computer-readable instructions, data structures, program modules, or other data. System memory 1004, removable storage 1009, and non-removable storage 1010 are all computer storage media examples (i.e., memory storage.) Computer storage media may include, but is not limited to, RAM, ROM, electrically erasable read-only memory (EEPROM), flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store information and which can be accessed by computing device 1000. Any such computer storage media may be part of device 1000. Computing device 1000 may also have input device(s) 1012 such as a keyboard, a mouse, a pen, a sound input device, a touch input device, a location sensor, a camera, a biometric sensor, etc. Output device(s) 1014 such as a display, speakers, a printer, etc. may also be included. The aforementioned devices are examples and others may be used.
[00140] Computing device 1000 may also contain a communication connection 1016 that may allow device 1000 to communicate with other computing devices 1018, such as over a network in a distributed computing environment, for example, an intranet or the Internet. Communication connection 1016 is one example of communication media. Communication media may typically be embodied by
computer readable instructions, data structures, program modules, or other data in a modulated data signal, such as a carrier wave or other transport mechanism, and includes any information delivery media. The term “modulated data signal” may describe a signal that has one or more characteristics set or changed in such a manner as to encode information in the signal. By way of example, and not limitation, communication media may include wired media such as a wired network or direct- wired connection, and wireless media such as acoustic, radio frequency (RF), infrared, and other wireless media. The term computer readable media as used herein may include both storage media and communication media.
[00141] As stated above, a number of program modules and data files may be stored in system memory 1004, including operating system 1005. While executing on processing unit 1002, programming modules 1006 (e.g., application 1020 such as a media player) may perform processes including, for example, one or more stages of methods, algorithms, systems, applications, servers, databases as described above. The aforementioned process is an example, and processing unit 1002 may perform other processes. Other programming modules that may be used in accordance with embodiments of the present disclosure may include sound encoding/decoding applications, machine learning application, acoustic classifiers etc.
[00142] Asset operators, ground troops and others involved in military combat may find themselves in complex situations and they may have to make a series of decisions in quick succession to accomplish a mission. These individuals may have a plan and a leader, but each one, or groups of people, still have to make individual decisions based on their training and information that they have about the situation. Communication and adherence to validated tactics is vital in such situations and insightful guidance provides a path to success. Al systems may process vast amounts of combat field data and provide insightful guidance to individuals, groups, leaders, etc. while they are being trained and while they are in combat situations.
[00143] There are many combat situations where Al systems may provide useful suggestions to military personnel in training and combat situations. For example, in accordance with an exemplary and non-limiting embodiment, a fighter pilot may be on a mission to escort and protect a strike package on a mission. The flight may encounter enemy fighters approaching to disrupt the package’s mission.
The escorting fighter pilot(s) has to make a decision on how to deal with the incoming fighters. The enemy may be a simple configuration of a manageable few assets, but the enemy may be a well -organized force with an advanced Integrated Air Defense System (IADS). The fighter pilot, and his flight, must manage this complex situation to accomplish the mission and avoid losses.
[00144] The inventors discovered new training systems for operators of vehicles. The training systems include presenting augmented reality content to vehicle operators as they operate real assets in real environments to provide the operators with experience in maneuvering programmed situations. The training systems also include debrief tools where recordings of the training sessions can be reviewed, and maneuver improvements can be taught to the operators. A human trainer may control the debrief tools in various ways to teach the operator what they might have done differently in the given situation. Al asset performance agents may also be used in the debrief to demonstrate what an Al trained asset may have done in the given situation that replicates what an experienced operator would have executed.
[00145] An Al asset performance agent (sometimes referred to as an “Al agent” herein) is a computer program that is designed to mimic the performance of an operator based on a physics model and/or training data. A physics model represents the performance attributes of a particular type of asset. For example, one may want to develop an Al agent to fly and maneuver in situations mimicking an aircraft like an F- 22 or an adversary’s aircraft. The F-22 Al agent would be programmed with a physics engine that represents the main flight performance capabilities of the F-22. The F-22 Al agent might have, for example, top speed at given altitudes, fuel range predictions based on performance, climb rate, G-force limits, pilot limits, thrust, turn rate, turn radius, etc. A virtual representation of an F-22 in flight, on the ground, etc., may then be computer controlled in accordance with instructions as limited by or guided by the physics engine. This enables a computer to run simulations using the F-22 Al agent in various training situations to discover trends, develop new strategies for managing such environments and training based on an Al model. A human trainer may also give feedback to improve the Al model (e.g., modifying algorithmic weightings). The Al agent may also be trained using actual operator / asset performance data.
[00146] An Al agent may be used as a virtual asset in a live augmented reality training environment to represent an asset to which an operator may have to interact. For example, a real pilot of a real airplane may fly the plane while wearing augmented reality gear. An adversary aircraft may be presented as augmented reality content to the pilot at a geospatial position based on virtual markers and the pilot may be able to engage in air combat with the adversary aircraft . The adversary aircraft may be controlled by, or in part controlled by, an Al agent trained to represent the particular type of adversary aircraft. The Al agent may be adjusted to represent an extreme version of the enemy (e.g., making all the right maneuvers) or it may be adjusted to represent something less (e.g., making most of the right maneuvers). The adjustment may call for ‘common’ mistakes or tendencies to be included in the control methodology. The adjustment may depend on the type of training designed for the pilot.
[00147] A debrief tool may be used to illustrate how an operator of an asset (e.g., pilot of a plane) maneuvered with respect to the augmented reality training environment. The debrief environment may include a 2D, 3D or another virtual environment. For example, the debrief tool may include a 3D virtual reality environment where the aircraft and training objects maneuvered. The training objects may have been live (e.g., real assets in the environment), virtual (e.g., computer generated and controlled), constructive (e.g., computer generated and at least partially human controlled).
[00148] Data from a training session may be used in generating the presentation in the virtual environment. The data may represent the attributes of the various assets over time. The real plane(s) may have been positionally tracked as well as conditionally tracked to provide an understanding of where the plane was at a particular time and what was going on with the plane’s avionics, flight planes, speed, G-force experience, kinetic energy states, potential energy states, etc. The virtual and constructive asset may log similar information such that all of the relevant information can be presented in the virtual environment.
[00149] All of the training assets, including the real asset with the real operator that is being trained, may be presented in the virtual environment debrief tool. A human instructor and the trainee may talk about what was happening during training
and the instructor may offer advice. One form of advice may be delivered by allowing the instructor to grab objects in the virtual environment to show their conditional data, move the object to talk about an alternate scenario, cause an object to make a maneuver different than what was performed, etc.
[00150] Another form of training in the debrief virtual environment would be to allow an Al agent representing the type of asset being used by the trainee to show what it would have done in the given situation during a playback of a real training session that may diverge from that the trainee performed. The Al agent maneuvers may be shown in comparison to, or instead of, the playback of what the trainee did in their flight.
[00151] A user of the debrief tools may stop the playback of the trainee’s flight at some point instruct an Al agent to take over from that point to demonstrate what it would have done. The Al agent becomes situationally educated by using situational information gathered by the exercise. For example, the trainee’s airplane data (e.g., speed, energy states, pose, orientation) was tracked up to the point of stopping the playback and some or all of the data may be incorporated into the Al agent to set it in a proper relational condition as the trainee’s airplane was at the time. The Al agent may also be programmed with data representing the environment, time of day, angle of the sun, landscape, threats, etc. The Al agent may also be programmed to mimic a condition of the trainee (e.g., how long they had been flying in the current mission, biometric information). The Al agent may be programmed to understand the mission or desired outcome.
[00152] Once the Al agent has been situationally educated it may be deployed within the debrief virtual environment. It may then be permitted to move forward in time to demonstrate alternate maneuvers. It may also be sent back in time to reset its situational education. A human instructor may talk about how the trainee reacted to situations in comparison to the Al agent’s performance. The debrief tool may compare the performance of the trainee’s performance with the Al agent’s performance to identify gaps, grades, critiques, tendencies, etc.
[00153] FIG. 11 illustrates a debriefing virtual environment in accordance with the principles of the present invention. A trace 1102 representing the trainee’s flight
path during a training event may be presented in geospatial relation with an enemy threat 1104 that was presented to the trainee during the training event. As illustrated, the trainee turned sharply right as they came into closer proximity to the threat.
[00154] A human trainer conducting the debrief may see an opportunity to show the trainee alternatives to the maneuver made. The trainer may select a point 1106 along the trainee's flight path to start an Al agent’s evaluation and suggestions. The Al agent may begin by programming its current conditions with the conditions of the trainee’s real plane such that it is in a similar condition (e.g., speed, energy states, enemy asset situations). Then, the Al agent may be instructed to make what it sees as the best maneuver 1108 for the given situation.
[00155] There may be a number of ways to present the Al agent's performance suggestions in the debrief tools. A user may decide how to look at the data and/or data representations. For example, the visual presentation of the actual operator’s performance may be positioned in the same virtual environment as the Al agent’s suggested performance (e.g., either the operator or the Al agent may be presented in a ‘ghosting’ fashion). The operator’s performance may be hidden so focus can be placed on the Al agent’s performance. The Al agent's performance may be hidden to focus on the operator's performance.
[00156] The debrief tools described herein may be used in various debriefing and planning situations: debriefing an operator based on training session performance (e.g., AR training, live training) debriefing an operator based on live action performance (e.g., actual mission), understanding group tendencies, evaluating operators and groups of operators (e.g., gap analysis), planning a mission and simulating outcomes (e.g., where the simulation is based in part on Al agent performance and in part on identified operator tendencies), etc. Further, the debriefing tools are not limited to military or defense situations. The debriefing tools may be used to teach civilian or commercial operators.
[00157] The debrief tools described herein may be used to debrief a number of training, evaluation, learning, compliance, etc. activities (e.g., flying in a formation; coordination of air assets, ground assets, water assets, space assets, jamming assets; coordination between friendly assets and enemy assets). The tools may be used to
debrief tactile maneuvers (e.g., tactical turning, landing, following navigation guidance).
[00158] Although the invention has been explained in relation to its preferred embodiment, it is to be understood that many other possible modifications and variations can be made without departing from the spirit and scope of the invention.
Claims
1. A method of debriefing based on a recording of a real flight of an aircraft by an operator in which said operator encountered a situation requiring a reaction by said operator, said recording comprising data of said real flight, said data comprising geospatial locations of said aircraft throughout said real flight and optionally those of one or more objects associated with said situation if applicable, flight parameters of said aircraft, and environmental conditions, and, optionally, a mission objective, said method comprising: displaying a replay of at least a portion of said real flight; and displaying an alternative reaction to said reaction of said operator using an Al agent, said Al agent being initialized with at least a portion of said data.
2. The method of claim 1, further comprising interrupting displaying said replay with said displaying of said alternative reaction for a comparison.
3. The method of claim 2, wherein said interrupting is performed by an instructor.
4. The method of claim 1, wherein both said reply and said alternative reaction are displayed simultaneously.
5. The method of claim 4, wherein one of said one of reply or said alternative reaction is ghosted in said display.
6. The method of claim 1, wherein just one of said reply or said alternative reaction is displayed at a time.
7. The method of claim 1, wherein said Al agent comprises a physics model of said aircraft.
8. The method of claim 1, wherein said situation involves an object.
9. The method of claim 8, wherein said object is a non-adverse object
10. The method of claim 9, wherein said non-adverse object is terrain.
11. The method of claim 8, wherein said object is a target.
12. The method of claim 8, wherein said object is an adverse object comprising one or more of (1) adverse aircraft or (2) adverse anti-aircraft weapons.
13. The method of claim 12, wherein said Al agent is initialized with a physics model of said adverse object.
14. The method of claim 12, wherein said Al agent represents said adverse object.
15. The method of claim 14, wherein said Al agent represent a non-flawed version of said adverse object.
16. The method of claim 14, wherein said Al agent represent a diminished version of said adverse object.
17. The method of claim 1, further comprising comparing said alternative reaction to said reaction to show a comparison.
18. The method of claim 17, wherein said comparison is used to identify one or more of gaps, grades, critiques, tendencies on live action performance (e.g., actual mission), understanding group tendencies, evaluating operators and groups of operators (e.g., gap analysis).
19. The method of claim 1, wherein said alternative reaction is at least one of (1) a different flight maneuver, (2) a different flight path, (3) a different velocity, (4) a different choice of weapons, (5) a different target, (6) a different instruct! on/command, (7) a revised mission objective . . . .
20. The method of claim 1, wherein method displays said replay and alternative reaction is either 2D or 3D.
21. The method of claim 20, wherein 3D display is presented in a space in which observers are present.
22. The method of claim 20, wherein 3d display is configured to allow said instructor to select a visually displayed object and have that object display its data.
23. The method of claim 1, wherein said Al agent is configured with a condition of said operator.
24. The method of claim 23, wherein said Al agent is configured with one of fatigue of said operator or biometric information of said operator.
25. The method of claim 1, wherein said displaying said alternative reaction goes beyond said real flight in the future.
26. The method of claim 1, further comprising having said Al agent learn from said real flight to determine (1) trends or (2) new trends.
27. A system of debriefing based on a recording of a real flight of an aircraft by an operator in which said operator encountered a situation requiring a reaction by said operator, said recording comprising data of said real flight, said data comprising geospatial locations of said aircraft throughout said real flight and optionally those of one or more objects associated with said situation if applicable, flight parameters of said aircraft, and environmental conditions, and, optionally, a mission objective, said system comprising: a processor configured with an Al agent and said recording; at least one display communicatively coupled to said processor, and configured for displaying a replay of at least a portion of said real flight, and for displaying an alternative reaction to said reaction of said operator using said Al agent, said Al agent being initialized with at least a portion of said data.
28. The system of claim 27 wherein said at least one display comprises a plurality of wearables for attendees of said debrief to wear.
29. The system of claim 28 wherein said wearables are virtual reality googles.
Applications Claiming Priority (2)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202363600311P | 2023-11-17 | 2023-11-17 | |
| US63/600,311 | 2023-11-17 |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| WO2025106992A1 true WO2025106992A1 (en) | 2025-05-22 |
Family
ID=95743416
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2024/056416 Pending WO2025106992A1 (en) | 2023-11-17 | 2024-11-18 | Operator training using artificial intelligence agent |
Country Status (1)
| Country | Link |
|---|---|
| WO (1) | WO2025106992A1 (en) |
Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090099824A1 (en) * | 2005-11-28 | 2009-04-16 | L-3 Communications Corporation | Distributed Physics Based Training System and Methods |
| US20150081197A1 (en) * | 2013-09-14 | 2015-03-19 | The Boeing Company | Visual flight rules approach to airports |
| US20180165979A1 (en) * | 2016-12-09 | 2018-06-14 | The Boeing Company | Electronic Device and Method for Debriefing Evidence-Based Training Sessions |
| US20200128902A1 (en) * | 2018-10-29 | 2020-04-30 | Holosports Corporation | Racing helmet with visual and audible information exchange |
| US20230334788A1 (en) * | 2020-12-20 | 2023-10-19 | Wael Zohni | Mixed-Reality Visor For In-Situ Vehicular Operations Training |
-
2024
- 2024-11-18 WO PCT/US2024/056416 patent/WO2025106992A1/en active Pending
Patent Citations (5)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20090099824A1 (en) * | 2005-11-28 | 2009-04-16 | L-3 Communications Corporation | Distributed Physics Based Training System and Methods |
| US20150081197A1 (en) * | 2013-09-14 | 2015-03-19 | The Boeing Company | Visual flight rules approach to airports |
| US20180165979A1 (en) * | 2016-12-09 | 2018-06-14 | The Boeing Company | Electronic Device and Method for Debriefing Evidence-Based Training Sessions |
| US20200128902A1 (en) * | 2018-10-29 | 2020-04-30 | Holosports Corporation | Racing helmet with visual and audible information exchange |
| US20230334788A1 (en) * | 2020-12-20 | 2023-10-19 | Wael Zohni | Mixed-Reality Visor For In-Situ Vehicular Operations Training |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12437667B2 (en) | Augmented reality for vehicle operations | |
| US12437666B2 (en) | Augmented reality for vehicle operations | |
| US20210295732A1 (en) | In-flight training simulation displaying a virtual environment | |
| US20230108678A1 (en) | Methods and systems to allow real pilots in real aircraft using augmented and virtual reality to meet in a virtual piece of airspace | |
| US20250199301A1 (en) | Methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience | |
| US20230206780A1 (en) | Methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience in a racing environment | |
| US20240053609A1 (en) | Methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience | |
| US20240327027A1 (en) | Augmented reality system for aircraft pilots using third party data | |
| WO2025106992A1 (en) | Operator training using artificial intelligence agent | |
| WO2022094279A1 (en) | Augmented reality for vehicle operations | |
| WO2025170616A2 (en) | Training of artificial intelligence battleground guidance system | |
| US20250322620A1 (en) | Methods, systems, apparatuses, and devices for facilitating provisioning of a virtual experience in a racing environment |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 24892447 Country of ref document: EP Kind code of ref document: A1 |