[go: up one dir, main page]

WO2025240965A1 - Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers - Google Patents

Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers

Info

Publication number
WO2025240965A1
WO2025240965A1 PCT/US2025/030023 US2025030023W WO2025240965A1 WO 2025240965 A1 WO2025240965 A1 WO 2025240965A1 US 2025030023 W US2025030023 W US 2025030023W WO 2025240965 A1 WO2025240965 A1 WO 2025240965A1
Authority
WO
WIPO (PCT)
Prior art keywords
user
wearable device
context
guidance
orchestrated
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
PCT/US2025/030023
Other languages
French (fr)
Inventor
Nicholas Wallen
Benjamin Neal Bethurum
Willy Huang
Sean Garrett KELLY
Hayden Schoen
Shengzhi WU
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Meta Platforms Technologies LLC
Original Assignee
Meta Platforms Technologies LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US19/209,723 external-priority patent/US20250355916A1/en
Application filed by Meta Platforms Technologies LLC filed Critical Meta Platforms Technologies LLC
Publication of WO2025240965A1 publication Critical patent/WO2025240965A1/en
Pending legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • G06F3/013Eye tracking input arrangements
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/017Gesture based interaction, e.g. based on a set of recognized hand gestures
    • GPHYSICS
    • G06COMPUTING OR CALCULATING; COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback

Definitions

  • This relates generally to approaches for interacting with an artificially intelligent agent and, more specifically, utilizing artificially intelligent agent included at wearable devices to augment user experiences.
  • a method comprising: in response to an indication that an artificial intelligence (Al) agent trigger condition is present, providing an Al agent sensor data obtained by a wearable device; determining, by the Al agent, a context-based activity based on the sensor data obtained by the wearable device; generating, by the Al agent, orchestrated guidance based on the context-based activity, wherein the orchestrated guidance includes a recommended action for performing the context-based activity; and presenting the orchestrated guidance at the wearable device.
  • Al artificial intelligence
  • a user may be assisted in performing a technical task by means of a continued and/or guided human-machine interaction process.
  • the context-based activity may be a first context-based activity.
  • the sensor data may be first sensor data.
  • the orchestrated guidance may be first orchestrated guidance.
  • the recommended action may be a first recommended action.
  • the method may further comprise: in accordance with a determination that the first recommended action for performing the first context-based activity was performed, providing the Al agent second sensor data obtained by the wearable device; determining, by the Al agent, a second context-based activity based on the second sensor data obtained by the wearable device; generating, by the Al agent, second orchestrated guidance based on the second context-based activity, wherein the second orchestrated guidance includes a second recommended action for performing the second context-based activity; and presenting the second orchestrated guidance at the wearable device.
  • the context-based activity may be a first context-based activity of a plurality of context-based activities determined by the by the Al agent based on the sensor data.
  • the orchestrated guidance may include a plurality of recommended actions for performing the plurality of context-based activities.
  • the recommended action may be a first recommended action of the plurality of recommended actions.
  • the first recommended action may be configured to perform the first context-based activity.
  • Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action of the plurality of recommended actions.
  • Generating the orchestrated guidance may include determining a subset of the plurality of recommended actions for performing the first context-based activity.
  • Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action of the plurality of recommended actions and the subset of the plurality of recommended actions for performing the first context-based activity.
  • Generating the orchestrated guidance may include determining a sequence of context-based activities of the plurality of context-based activities to be performed, including a second context-based activity to follow the first context-based activity.
  • Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action and the second recommended action of the plurality of recommended actions for performing the plurality of context-based activities.
  • the method may further comprise: in response to a user input selecting the recommended action for performing the context-based activity, causing the wearable device to initiate a do-not-disturb mode, wherein, while in the do-not-disturb mode, the wearable device suppresses, at least, received notifications; and in response to an indication that participation in the context-based activity ceased: causing the wearable device to cease the do-not-disturb mode, generating, by the Al agent, a notification summary based on the notifications received while the wearable device was in the do-not-disturb mode, and presenting the notification summary at the wearable device.
  • the method may further comprise performing, by the Al agent, a search based on the recommended action; determining a task to perform based on the search; and presenting the task at the wearable device.
  • Presenting the orchestrated guidance at the wearable device may include at least one of: causing presentation of a user interface element associated with the orchestrated guidance at a communicatively coupled display, and/or causing presentation of audible guidance associated with the orchestrated guidance at a communicatively coupled speaker.
  • the context-based activity may be performed at a physical activity.
  • a wearable device comprising: a display; one or more sensors; and one or more programs, wherein the one or more programs are stored in memory and configured to be executed by one or more processors to cause the one or more processors to carry out the method of the first aspect.
  • a computer-readable storage medium including executable instructions that, when executed by one or more processors, cause the one or more processors to carry out the method of the first aspect.
  • the medium may be non-transitory.
  • a computer program product comprising instructions that, when executed by one or more processors of a computer, cause the one or more processors to carry out the method of the first aspect.
  • the example wearable device can be a headwearable device including a display, one or more sensors, and one or more programs.
  • the one or more programs are stored in memory and configured to be executed by one or more processors, the one or more programs including instructions for, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining data associated with an on-going activity performed by the user of the head-wearable device.
  • the one or more programs include instructions for generating, by an artificial intelligence (Al) agent, a contextbased response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device.
  • the one or more programs include instructions for presenting, at the head-wearable device, context-based response. The context-based response is presented within a portion of a field of view of the user.
  • a method for facilitating performance of a physical activity performed by user includes, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining data associated with an on-going activity performed by the user of the head-wearable device.
  • the method also includes generating, by an artificial intelligence (Al) agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the headwearable device.
  • the method further includes presenting, at the head-wearable device, context-based response, wherein the context-based response is presented within a portion of a field of view of the user.
  • Al artificial intelligence
  • a non-transitory, computer-readable storage medium including executable instructions that, when executed by one or more processors of a wearable device (e.g., a head-wearable device), cause the one or more processors to facilitate performance of a physical activity performed by user is described herein.
  • the executable instructions when executed by one or more processors, cause the one or more processors to, in response to an indication that a user of a head-wearable device is participating in an activity, obtain data associated with an on-going activity performed by the user of the headwearable device.
  • the executable instructions when executed by one or more processors, cause the one or more processors to generate, by an artificial intelligence (Al) agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device.
  • Al artificial intelligence
  • the executable instructions when executed by one or more processors, cause the one or more processors to present, at the head-wearable device, context-based response, wherein the context-based response is presented within a portion of a field of view of the user.
  • Instructions that cause performance of the methods and operations described herein can be stored on a non-transitory computer readable storage medium.
  • the non- transitory computer-readable storage medium can be included on a single electronic device or spread across multiple electronic devices of a system (computing system).
  • a non- exhaustive of list of electronic devices that can either alone or in combination (e.g., a system) perform the method and operations described herein include an extended-reality (XR) headset/glasses (e.g., a mixed-reality (MR) headset or a pair of augmented-reality (AR) glasses as two examples), a wrist-wearable device, an intermediary processing device, a smart textile-based garment, etc.
  • XR extended-reality
  • MR mixed-reality
  • AR augmented-reality
  • the instructions can be stored on a pair of AR glasses or can be stored on a combination of a pair of AR glasses and an associated input device (e.g., a wrist-wearable device) such that instructions for causing detection of input operations can be performed at the input device and instructions for causing changes to a displayed user interface in response to those input operations can be performed at the pair of AR glasses.
  • the devices and systems described herein can be configured to be used in conjunction with methods and operations for providing an XR experience.
  • the methods and operations for providing an XR experience can be stored on a non-transitory computer- readable storage medium.
  • the devices and/or systems described herein can be configured to include instructions that cause the performance of methods and operations associated with the presentation and/or interaction with an extended-reality (XR) headset. These methods and operations can be stored on a non-transitory computer-readable storage medium of a device or a system. It is also noted that the devices and systems described herein can be part of a larger, overarching system that includes multiple devices.
  • XR extended-reality
  • a non-exhaustive of list of electronic devices that can, either alone or in combination (e.g., a system), include instructions that cause the performance of methods and operations associated with the presentation and/or interaction with an XR experience include an extended-reality headset (e.g., a mixed-reality (MR) headset or a pair of augmented-reality (AR) glasses as two examples), a wrist-wearable device, an intermediary processing device, a smart textile-based garment, etc.
  • MR mixed-reality
  • AR augmented-reality
  • the XR headset can be in communication with one or more other devices (e.g., a wrist-wearable device, a server, intermediary processing device) which together can include instructions for performing methods and operations associated with the presentation and/or interaction with an extended- reality system (i.e., the XR headset would be part of a system that includes one or more additional devices).
  • other devices e.g., a wrist-wearable device, a server, intermediary processing device
  • the XR headset would be part of a system that includes one or more additional devices.
  • Figures 1A-1 N illustrate invocation of an artificially intelligent agent at one or more wearable devices for providing guidance based on an activity of a user.
  • Figures 2A-2L illustrate context-based responses generated by an artificially intelligent agent based on activities performed by a user.
  • Figures 3A-3D illustrate example user interfaces and additional features available at an Al assistive system.
  • Figures 4A and 4B illustrate example sequences of user interactions with personalized assistive systems.
  • Figure 5 illustrates a flow chart of a method for generating orchestrated guidance based on an activity of a user.
  • Figure 6 illustrates a flow chart of a method for facilitating performance of a physical activity performed by user.
  • Figures 7A - 7C-2 illustrate example XR systems that include AR and MR systems.
  • Embodiments of this disclosure can include or be implemented in conjunction with various types of extended-realities (XRs) such as mixed-reality (MR) and augmented- reality (AR) systems.
  • XRs extended-realities
  • MRs and ARs as described herein, are any superimposed functionality and/or sensory-detectable presentation provided by MR and AR systems within a user’s physical surroundings.
  • Such MRs can include and/or represent virtual realities (VRs) and VRs in which at least some aspects of the surrounding environment are reconstructed within the virtual environment (e g., displaying virtual reconstructions of physical objects in a physical environment to avoid the user colliding with the physical objects in a surrounding physical environment).
  • the surrounding environment that is presented through a display is captured via one or more sensors configured to capture the surrounding environment (e.g., a camera sensor, time-of-flight (ToF) sensor). While a wearer of an MR headset can see the surrounding environment in full detail, they are seeing a reconstruction of the environment reproduced using data from the one or more sensors (i.e. , the physical objects are not directly viewed by the user). An MR headset can also forgo displaying reconstructions of objects in the physical environment, thereby providing a user with an entirely VR experience.
  • sensors e.g., a camera sensor, time-of-flight (ToF) sensor.
  • An AR system provides an experience in which information is provided, e.g., through the use of a waveguide, in conjunction with the direct viewing of at least some of the surrounding environment through a transparent or semitransparent waveguide(s) and/or lens(es) of the AR glasses.
  • extended reality XR
  • this application also uses, at times, a head-wearable device or headset device as a catchall term that covers XR headsets such as AR glasses and MR headsets.
  • an MR environment can include, but is not limited to, non-immersive, semi-immersive, and fully immersive VR environments.
  • AR environments can include marker-based AR environments, markerless AR environments, location-based AR environments, and projection-based AR environments. The above descriptions are not exhaustive and any other environment that allows for intentional environmental lighting to pass through to the user would fall within the scope of an AR, and any other environment that does not allow for intentional environmental lighting to pass through to the user would fall within the scope of an MR.
  • the AR and MR content can include video, audio, haptic events, sensory events, or some combination thereof, any of which can be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to a viewer). Additionally, AR and MR can also be associated with applications, products, accessories, services, or some combination thereof, which are used, for example, to create content in an AR or MR environment and/or are otherwise used in (e.g., to perform activities in) AR and MR environments.
  • a user can perform a swiping in-air hand gesture to cause a song to be skipped by a song-providing application programming interface (API) providing playback at, for example, a home speaker.
  • API application programming interface
  • a hand gesture can include an in-air gesture, a surfacecontact gesture, and or other gestures that can be detected and determined based on movements of a single hand (e.g., a one-handed gesture performed with a user’s hand that is detected by one or more sensors of a wearable device (e.g., electromyography (EMG) and/or inertial measurement units (IMUs) of a wrist-wearable device, and/or one or more sensors included in a smart textile wearable device) and/or detected via image data captured by an imaging device of a wearable device (e.g., a camera of a head-wearable device, an external tracking camera setup in the surrounding environment)).
  • EMG electromyography
  • IMUs inertial measurement units
  • “In-air” generally includes gestures in which the user’s hand does not contact a surface, object, or portion of an electronic device (e.g., a head-wearable device or other communicatively coupled device, such as the wristwearable device), in other words the gesture is performed in open air in 3D space and without contacting a surface, an object, or an electronic device.
  • Surface-contact gestures contacts at a surface, object, body part of the user, or electronic device
  • a contact or an intention to contact
  • a surface e.g., a single- or double-finger tap on a table, on a user’s hand or another finger, on the user’s leg, a couch, a steering wheel.
  • the different hand gestures disclosed herein can be detected using image data and/or sensor data (e.g., neuromuscular signals sensed by one or more biopotential sensors (e.g., EMG sensors) or other types of data from other sensors, such as proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors) detected by a wearable device worn by the user and/or other electronic devices in the user’s possession (e.g., smartphones, laptops, imaging devices, intermediary devices, and/or other devices described herein).
  • biopotential sensors e.g., EMG sensors
  • sensors such as proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors
  • a wearable device worn by the user and/or other electronic devices in the user’s possession e.g., smartphones, laptops, imaging devices, intermediary devices, and/or other devices described herein.
  • a user can provide inputs using in-air or surface-contact gestures that are detected using neuromuscular signal sensors of the wrist-wearable device.
  • alternative and entirely interchangeable input modalities can be used instead, such as camera(s) located on the headset/glasses or elsewhere to detect in-air or surface-contact gestures or inputs at an intermediary processing device (e.g., through physical input components (e.g., buttons and trackpads)).
  • intermediary processing device e.g., through physical input components (e.g., buttons and trackpads)
  • These different input modalities can be interchanged based on both desired user experiences, portability, and/or a feature set of the product (e.g., a low-cost product may not include hand-tracking cameras).
  • an in-air gesture input detected by a camera of a head-wearable device can cause an output to occur at a head-wearable device or control another electronic device different from the head-wearable device.
  • an input detected using data from a neuromuscular signal sensor can also cause an output to occur at a headwearable device or control another electronic device different from the head-wearable device. While only a couple examples are described above, one skilled in the art would understand that different input modalities are interchangeable along with different output modalities in response to the inputs.
  • a processor e.g., a central processing unit (CPU) or microcontroller unit (MCU)
  • CPU central processing unit
  • MCU microcontroller unit
  • an electronic device e.g., a wrist-wearable device, a head-wearable device, a handheld intermediary processing device (HIPD), a smart textilebased garment, or other computer system.
  • processors e.g., a central processing unit (CPU) or microcontroller unit (MCU)
  • CPU central processing unit
  • MCU microcontroller unit
  • a processor may be (i) a general processor designed to perform a wide range of tasks, such as running software applications, managing operating systems, and performing arithmetic and logical operations; (ii) a microcontroller designed for specific tasks such as controlling electronic devices, sensors, and motors; (iii) a graphics processing unit (GPU) designed to accelerate the creation and rendering of images, videos, and animations (e.g., VR animations, such as three-dimensional modeling); (iv) a field-programmable gate array (FPGA) that can be programmed and reconfigured after manufacturing and/or customized to perform specific tasks, such as signal processing, cryptography, and machine learning; or (v) a digital signal processor (DSP) designed to perform mathematical operations on signals such as audio, video, and radio waves.
  • a general processor designed to perform a wide range of tasks, such as running software applications, managing operating systems, and performing arithmetic and logical operations
  • a microcontroller designed for specific tasks such as controlling electronic devices, sensors, and motors
  • controllers are electronic components that manage and coordinate the operation of other components within an electronic device (e.g., controlling inputs, processing data, and/or generating outputs).
  • controllers can include (i) microcontrollers, including small, low-power controllers that are commonly used in embedded systems and Internet of Things (loT) devices; (ii) programmable logic controllers (PLCs) that may be configured to be used in industrial automation systems to control and monitor manufacturing processes; (iii) system-on-a-chip (SoC) controllers that integrate multiple components such as processors, memory, I/O interfaces, and other peripherals into a single chip; and/or (iv) DSPs.
  • a graphics module is a component or software module that is designed to handle graphical operations and/or processes and can include a hardware module and/or a software module.
  • memory refers to electronic components in a computer or electronic device that store data and instructions for the processor to access and manipulate.
  • the devices described herein can include volatile and non-volatile memory.
  • Examples of memory can include (i) random access memory (RAM), such as DRAM, SRAM, DDR RAM or other random access solid state memory devices, configured to store data and instructions temporarily; (ii) read-only memory (ROM) configured to store data and instructions permanently (e.g., one or more portions of system firmware and/or boot loaders); (iii) flash memory, magnetic disk storage devices, optical disk storage devices, other non-volatile solid state storage devices, which can be configured to store data in electronic devices (e.g., universal serial bus (USB) drives, memory cards, and/or solid-state drives (SSDs)); and (iv) cache memory configured to temporarily store frequently accessed data and instructions.
  • RAM random access memory
  • ROM read-only memory
  • flash memory magnetic disk storage devices
  • optical disk storage devices other non-volatile solid state storage devices, which
  • Memory can include structured data (e.g., SQL databases, MongoDB databases, GraphQL data, or JSON data).
  • Other examples of memory can include (i) profile data, including user account data, user settings, and/or other user data stored by the user; (ii) sensor data detected and/or otherwise obtained by one or more sensors; (iii) media content data including stored image data, audio data, documents, and the like; (iv) application data, which can include data collected and/or otherwise obtained and stored during use of an application; and/or (v) any other types of data described herein.
  • a power system of an electronic device is configured to convert incoming electrical power into a form that can be used to operate the device.
  • a power system can include various components, including (i) a power source, which can be an alternating current (AC) adapter or a direct current (DC) adapter power supply; (ii) a charger input that can be configured to use a wired and/or wireless connection (which may be part of a peripheral interface, such as a USB, micro-USB interface, near-field magnetic coupling, magnetic inductive and magnetic resonance charging, and/or radio frequency (RF) charging); (iii) a power-management integrated circuit, configured to distribute power to various components of the device and ensure that the device operates within safe limits (e.g., regulating voltage, controlling current flow, and/or managing heat dissipation); and/or (iv) a battery configured to store power to provide usable power to components of one or more electronic devices.
  • a power source which can be an alternating current (AC) adapter or a direct current (DC
  • peripheral interfaces are electronic components (e.g., of electronic devices) that allow electronic devices to communicate with other devices or peripherals and can provide a means for input and output of data and signals.
  • peripheral interfaces can include (i) USB and/or micro-USB interfaces configured for connecting devices to an electronic device; (ii) Bluetooth interfaces configured to allow devices to communicate with each other, including Bluetooth low energy (BLE); (iii) near-field communication (NFC) interfaces configured to be short-range wireless interfaces for operations such as access control; (iv) pogo pins, which may be small, spring-loaded pins configured to provide a charging interface; (v) wireless charging interfaces; (vi) global- positioning system (GPS) interfaces; (vii) Wi-Fi interfaces for providing a connection between a device and a wireless network; and (viii) sensor interfaces.
  • BLE Bluetooth low energy
  • NFC near-field communication
  • pogo pins which may be small, spring-loaded pins configured to provide a charging interface
  • wireless charging interfaces
  • sensors are electronic components (e.g., in and/or otherwise in electronic communication with electronic devices, such as wearable devices) configured to detect physical and environmental changes and generate electrical signals.
  • sensors can include (i) imaging sensors for collecting imaging data (e.g., including one or more cameras disposed on a respective electronic device, such as a simultaneous localization and mapping (SLAM) camera); (ii) biopotential-signal sensors; (iii) IMUs for detecting, for example, angular rate, force, magnetic field, and/or changes in acceleration; (iv) heart rate sensors for measuring a user’s heart rate; (v) peripheral oxygen saturation (SpO2) sensors for measuring blood oxygen saturation and/or other biometric data of a user; (vi) capacitive sensors for detecting changes in potential at a portion of a user’s body (e.g., a sensor-skin interface) and/or the proximity of other devices or objects; (vii) sensors for detecting some inputs (e.g., capaci
  • imaging sensors for collecting imaging data
  • biopotential-signal-sensing components are devices used to measure electrical activity within the body (e.g., biopotentialsignal sensors).
  • biopotential-signal sensors include (i) electroencephalography (EEG) sensors configured to measure electrical activity in the brain to diagnose neurological disorders; (ii) electrocardiography (ECG or EKG) sensors configured to measure electrical activity of the heart to diagnose heart problems; (iii) EMG sensors configured to measure the electrical activity of muscles and diagnose neuromuscular disorders; (iv) electrooculography (EOG) sensors configured to measure the electrical activity of eye muscles to detect eye movement and diagnose eye disorders.
  • an application stored in memory of an electronic device includes instructions stored in the memory.
  • applications include (i) games; (ii) word processors; (iii) messaging applications; (iv) media-streaming applications; (v) financial applications; (vi) calendars; (vii) clocks; (viii) web browsers; (ix) social media applications; (x) camera applications; (xi) web-based applications; (xii) health applications; (xiii) AR and MR applications; and/or (xiv) any other applications that can be stored in memory.
  • the applications can operate in conjunction with data and/or one or more components of a device or communicatively coupled devices to perform one or more operations and/or functions.
  • communication interface modules can include hardware and/or software capable of data communications using any of a variety of custom or standard wireless protocols (e.g, IEEE 802.15.4, Wi-Fi, ZigBee, 6L0WPAN, Thread, Z-Wave, Bluetooth Smart, ISA100.11a, WirelessHART, or MiWi), custom or standard wired protocols (e.g., Ethernet or HomePlug), and/or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document.
  • a communication interface is a mechanism that enables different systems or devices to exchange information and data with each other, including hardware, software, or a combination of both hardware and software.
  • a communication interface can refer to a physical connector and/or port on a device that enables communication with other devices (e.g., USB, Ethernet, HDMI, or Bluetooth).
  • a communication interface can refer to a software layer that enables different software programs to communicate with each other (e.g., APIs and protocols such as HTTP and TCP/IP).
  • a graphics module is a component or software module that is designed to handle graphical operations and/or processes and can include a hardware module and/or a software module.
  • non-transitory computer-readable storage media are physical devices or storage medium that can be used to store electronic data in a non- transitory form (e.g., such that the data is stored permanently until it is intentionally deleted and/or modified).
  • the systems and methods disclosed herein provide different ways in which wearable devices can utilize artificial intelligence (Al) and/or an Al Agent (also referred to as a Al digital assistant or Al assistant).
  • a head-wearable device can retrieve information and use the information with an Al agent to generate responses and/or recommendations that are displayed at the head-wearable device and/or another communicatively device.
  • the systems and method disclosed here can be used collaborate with other users (including wearers of other wearable devices), and interact with third party applications using built-in Al models.
  • the systems and methods disclosed herein can utilize a user interactable Al agent to perform various tasks at the user’s request, as well as utilize the Al agent to monitor situations and provide user-specific assistance.
  • the systems and methods disclosed herein utilize Al agent to work with wearable devices and other devices (e.g., laptop, tablet, watches, desktops, phones, and other internet connected devices) within an ecosystem to accomplish tasks across multiple devices (e.g., XR systems described below in reference to Figures 7A-7C-2).
  • an Al agent can be configured to control an aspect of one or more of the other devices based on a request from the user.
  • the Al agent can also be invoked on different devices based on a determination that the user is interacting with a device other than a wearable device.
  • the systems and methods disclosed herein can use an Al agent to augment a user experience.
  • the Al agent can receive sensor data and/or other information captured by a wearable device, and use the sensor data and/or other information to generate and provide recommended actions and/or context-based responses.
  • a head-wearable device worn by the user can capture information corresponding to a field of view of the user 105 and/or a location of the user to generate and provide recommended actions and/or context-based responses.
  • the systems and methods disclosed herein generate and provide tailored information to a user based on location and/or data received from one or more wearable devices (e.g., sensor data and/or image data of a wristwearable device, a head-wearable device, etc.).
  • the systems and methods disclosed herein utilize an Al agent to collate recorded information (e.g., camera photos and videos) across multiple wearable devices to produce unique media (e.g., a single video which stitches the multiple head-wearable devices video feed into a single viewing experience).
  • positional data of each communicatively coupled device e.g., wearable device, such as a head-wearable device
  • FIG. 1A-1 N illustrate invocation of an artificially intelligent agent at one or more wearable devices for providing guidance based on an activity of a user.
  • An Al guidance system 100 shown and described in reference to Figures 1A-1 N provides example orchestrated guidance provided to a user 105 visiting a museum.
  • the Al guidance system 100 includes at least a wrist-wearable device 1 10 and a head-wearable device 120 donned by the user 105.
  • the Al guidance system 100 can include other wearable devices worn by the user 105, such as smart textile-based garments (e.g., wearable bands, shirts, etc.), and/or other electronic devices, such as an HIPD 742, a computer 740 (e.g., a laptop), mobile devices 750 (e.g., smartphones, tablets), and/or other electronic devices described below in reference to Figures 7A-7C.
  • the Al guidance system 100, the wearable devices, and the electronic devices can be communicatively coupled via a network (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN).
  • a network e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN.
  • the Al guidance system 100 further includes an Al agent 115 (represented by star symbols) that can be invoked by the user 105 via one or more devices of the Al guidance system 100 (e.g., a wearable device, such as a wrist-wearable device 110 and/or a head-wearable device 120).
  • a wearable device such as a wrist-wearable device 110 and/or a head-wearable device 120.
  • the Al agent 1 15 can be invoked in accordance with a determination that an Al agent trigger condition is present (as discussed below).
  • the wrist-wearable device 1 10 can include a display 1 12, an imaging device 1 14 (e.g., a camera), a microphone, a speaker, input surfaces (e.g., touch input surfaces, mechanical inputs, etc.), and one or more sensors (e.g., biopotential sensors (e.g., EMG sensors), proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors, etc.).
  • a display 1 12 e.g., a camera
  • input surfaces e.g., touch input surfaces, mechanical inputs, etc.
  • sensors e.g., biopotential sensors (e.g., EMG sensors), proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors, etc.).
  • the head-wearable device 120 can include another imaging device 122, an additional microphone, an additional speaker, additional input surfaces (e.g., touch input surfaces, mechanical inputs, etc.), and one or more additional sensors (e.g., biopotential sensors (e.g., EMG sensors), gaze trackers, proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors, etc.).
  • the head-wearable device 120 includes a display.
  • the wrist-wearable device 110 provides first example orchestrated guidance. While the user 105 is at the museum, the wrist-wearable device 1 10 and the head-wearable device 120 capture at least sensor data and image data via one or more sensors and/or imaging devices (e.g., imaging devices 1 14 and 122). In some examples, the head-wearable device 120 captures audio data.
  • the Al guidance system 100 can determine, based on image data, sensor data, audio data, and/or any other data available to the Al guidance system 100, whether an Al agent trigger condition is satisfied and, in accordance with a determination that an Al agent trigger condition is satisfied, the Al guidance system 100 can provide the indication that an Al agent trigger condition is present.
  • the Al guidance system 100 provides the Al agent 1 15, at least, image data, sensor data, audio data, and/or any other data captured by the devices of the Al guidance system 100.
  • the Al guidance system 100 provides the Al agent 1 15, at least, image data, sensor data, audio data, and/or any other data captured by the devices of the Al guidance system 100 in response to user invocation of the Al agent 1 15.
  • the Al agent 1 15 can be invoked via touch inputs, voice commands, hand gestures detected by and/or received at the wristwearable device 1 10, the head-wearable device 120, and/or any other device of the Al guidance system 100.
  • the Al agent 115 can use, at least, the image data and/or the sensor data received from the Al guidance system 100 to determine a context-based activity.
  • the Al agent 1 15 can use the image data and/or the sensor data to determine that the user 115 is visiting or exploring the museum.
  • the Al agent 1 15 can also use audio data to determine a context-based activity.
  • the context-based activity can be a physical activity (e.g. running, walking, etc.) and/or participation in an event (e.g., sightseeing, performing a hobby, cooking, driving, participating in a meeting, etc.).
  • the Al agent 115 can further generate orchestrated guidance based on the context-based activity.
  • the orchestrated guidance includes a recommended action for performing the context-based activity.
  • the Al guidance system 100 can present the orchestrated guidance at a wearable device (e.g., the wrist-wearable device 1 10 and/or the head-wearable device 120) and/or any other communicatively coupled device.
  • the Al agent 1 15 provides orchestrated guidance for the user 105’s museum visit, the orchestrated guidance including one or more recommended actions for facilitating the museum visit.
  • the orchestrated guidance and the recommended actions are presented at a display 1 12 of the wrist-wearable device 1 10.
  • the wrist-wearable device 1 10 presents, via the display 112, the first orchestrated guidance 116 (e.g., “Welcome to the museum! Here are some things you can do!”) and the recommended actions (e.g. take tour user interface (Ul) element 1 18 and do-not-disturb Ul element 1 19) generated by the Al agent 1 15.
  • the Al guidance system 100 can tailor the guided tour for the user 105.
  • Figure 1 B shows a field of view 125 of the user 105 via the head-wearable device 120.
  • the orchestrated guidance generated by the Al agent 1 15 can also be presented via a display of the head-wearable device 120.
  • the field of view 125 of the user 105 includes a first orchestrated guidance Ul element 127 (e.g. “Welcome to the museum! Let’s take a look around”).
  • Figures 1A and 1 B show orchestrated guidance and recommended actions presented at displays of the wrist-wearable device 110 and/or the head-wearable device 120
  • the orchestrated guidance and recommended actions can be presented via a speaker of wrist-wearable device 110, the head-wearable device 120, and/or another communicatively coupled device.
  • Figure 1 C show the user 105 providing a first user input 129 selecting a recommended action of the first orchestrated guidance 1 16.
  • the user 105 performs a hand gesture (e.g. a pinch) to provide a first user input 129 selecting the do-not- disturb Ul element 1 19.
  • the first user input 129 selecting the do-not-disturb Ul element 119 causes the wrist-wearable device 110, the head-wearable device 120, and/or other devices of the Al guidance system 100 to initiate a do-not-disturb mode (or focus mode, away mode, etc.).
  • the Al guidance system 100 suppresses, at least, received notifications, calls, and/or messages.
  • the use 105 can provide a voice request and/or other input to the Al guidance system 100 to silence notifications and provide a summary of the notifications later.
  • Figure 1 D shows a confirmation message generated by the Al agent 1 15.
  • the Al agent in response to the first user input 129, generates a corresponding response or recommended action.
  • the field of view 125 of the user 105 includes a confirmation message Ul element 130 based on an accepted recommended action of the first orchestrated guidance 116.
  • Figure 1 E shows updates to the first orchestrated guidance 116 based on one or more user inputs.
  • the orchestrated guidance generated by the Al agent 115 can include a subset of a plurality of recommended actions for performing the context-based activity.
  • the orchestrated guidance when presented at a wearable device, can include at least the subset of the plurality of recommended actions for performing the context-based activity.
  • one or more recommended actions of an orchestrated guidance are updated based on a user input selecting the one or more recommended actions.
  • the first orchestrated guidance 116 includes at least two Ul elements - take tour Ul element 118 and do-not-disturb Ul element 1 19 - and the Al agent 1 15 updates the first orchestrated guidance 1 16 to replace the do-not-disturb Ul element 119 with a view map Ul element 131 after detecting the first user input 129 selecting the do-not-disturb Ul element 1 19.
  • the second user input 133 selecting the take tour Ul element 1 18 cause the Al agent 115 to present updated first orchestrated guidance 116 and/or updated recommended actions.
  • one or more recommended actions of an orchestrated guidance are updated based on the user 105 forgoing to select or ignoring one or more recommended actions.
  • the Al agent 115 can determine that a context-based activity is one of a plurality of context based activities and, when generating the orchestrated guidance, determine a sequence for performing the plurality of context based activities (or context-based activities to be performed together and/or on parallel).
  • the context-based activity can be a first context-based activity of a plurality of context-based activities determined by the by the Al agent 1 15 (based on the sensor data, audio data, and/or image data)
  • the orchestrated guidance can include a plurality of recommended actions for performing the plurality of context-based activities, and the recommended action is a first recommended action of the plurality of recommended actions, the first recommended action being configured to perform the first context-based activity.
  • the Al agent 1 15 can determine when one or more contextbased activities are completed, identify similar context-based activities, provide alternate context-based activities (if one or more specific context-based activities cannot be performed or alternate suggestion are present).
  • the user 105 can have a schedule including at least two events - the museum visit (e.g., a first context-based activity) and a dinner (e.g., a second context-based activity) - and the orchestrated guidance determined by the Al agent 115 can include a first set of recommended actions for augmenting the user 105’s museum visit and a second set of recommended actions for augmenting the user 105’s dinner, the second set of recommended actions being presented to the user 105 in accordance with a determination that the museum visit has concluded (e.g, the user 105 leaves the museum, the user 105 terminates an augmented experience for the museum visit provided by the Al agent, the scheduled museum visit time elapses, etc.).
  • Figure 1 F shows a context-based response generated by the Al agent 115.
  • the context-based response is generated in response to the second user input 133 selecting the take tour Ul element 1 18.
  • the Al agent 1 15 generates a context-based response to facilitate the museum tour.
  • the user 105 can view a piece of art and the Al agent 1 15 can recognize the art and provide contextual information (or the contextbased response) to the user 105 (e.g., by presenting the information at a wrist-wearable device).
  • the Al agent 1 15 can use the sensor data, audio data, and/or the image data to generate the context-based response.
  • the Al agent 1 15 uses the sensor data, audio data, and/or the image data to identify that the statute 134 is an object of interest to the user 105 and generates a context-based response based on the statute 134. Identification of an object of interest is discussed below in reference to Figure 1 G.
  • the context-based response can be presented at the wrist-wearable device 110, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100.
  • the Al agent 1 15 presents a first context-based response Ul element 135 via a display of the head-wearable device 120, as shown in field of view 125 of the user 105.
  • Figure 1 G shows identification of an object of interest.
  • the Al guidance system 100 can identify an object of interest based on user gaze (determined by one or more eye tackers, sensors, and/or imaging devices of the head-wearable device 120 (e.g., gaze of user focused on an object for a predetermined amount of time (e.g., 10 seconds, 30 seconds, etc.))), direction of a field of view of the user 105 (determined by one or more sensors and/or imaging devices of the head-wearable device 120), pointing gestures performed by the user 105 (determined by one or more sensors and/or imaging devices of the wrist-wearable device 110 and/or the head-wearable device 120), voice commands, and/or other inputs provided by the user 105 to select an object of interest.
  • user gaze determined by one or more eye tackers, sensors, and/or imaging devices of the head-wearable device 120 (e.g., gaze of user focused on an object for a predetermined amount of time (e.g., 10 seconds, 30 seconds, etc.)
  • the user 105 provides a voice command 137 describing an object of interest.
  • the user 105 can perform a pointing gesture 138 to identify the object of interest and/or to augment or supplement the voice command 137.
  • the Al guidance system 100 can use one or more inputs modalities to identify an object of interest. In this way, the Al guidance system 100 can provide the user 105 with a tailored guided tour of a venue or the museum based on user specific objects of interest (animate or inanimate) within the venue or the museum (e.g., artwork the user 105 spends time appreciating).
  • Figure 1 H shows one or more additional Ul elements associated with the orchestrated guidance.
  • the Al guidance system 100 can present a highlight and/or one or more animations to identify a target object or object of interest.
  • the head-wearable device 120 can include a dimmable lens controlled by the Al guidance system 100 and can provide additional information to the user 105 (e.g., directing the user 105’s focus to certain objects within their field of view).
  • the Al guidance system 100 cause selective dimming of a portion of a display of the head-wearable device 120 such that an animated dimming target Ul element 139 is presented to the user 105.
  • the animated dimming target Ul element 139 can be used to draw the user 105’s attention to a portion of the field of view 125 such that the use 105 can confirm a selected object of interest or be notified of a portion of the field of view 125 being analyzed by the Al guidance system 100.
  • Figure 1 H further shows a second context-based response presented at the wrist-wearable device 1 10, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100.
  • the Al agent 115 presents a second context-based response Ul element 141 via a display of the head-wearable device 120, as shown in field of view 125 of the user 105.
  • the second context-based response is based on the object of interest identified by the user 105 and highlighted by the Al guidance system 100.
  • the context-based responses can also be provided as audio responses (or audio guidance) via speakers of the wrist-wearable device 1 10, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100.
  • updated orchestrated guidance is presented to the user 105.
  • second orchestrated guidance 143 including a second set of recommended actions e.g., Ul elements 144, 145, and 146) are presented to the user 105 via one or more wearable devices.
  • the second orchestrated guidance 143 and the second set of recommended actions can be based on the user’s current and/or past experiences at the museum and/or during the museum tour.
  • the Al agent 1 15 can provide a recommended action to explore the unseen landmarks (e.g., as shown by explore landmarks Ul element 145).
  • the user 105 provides a third user input 147 selecting an end tour Ul element 146.
  • Figure 1 J shows a notification summary presented at a wearable device of the Al guidance system 100.
  • the Al guidance system 100 in accordance with a determination that the end tour Ul element 146 was selected, ceases the user 105’s participation in the context-based activity (e.g., the museum visit).
  • the Al guidance system 100 in accordance with a determination that the museum visit has ended, causes the wearable devices or other communicatively coupled devices to cease the do-not-disturb mode.
  • the Al guidance system 100 after detecting that the do-not-disturb mode ceased, generate, using the Al agent 115, a notification summary based on the notifications received while the wearable devices (or other devices of the Al guidance system 100) were in the do- not-disturb mode.
  • the summary can be a natural language summary provided by the Al agent 115 that summarized the received notifications.
  • the notification summary can be presented via visual feedback (e.g., notification summary Ul element 140 presented via a communicatively coupled display), audio feedback (e.g., text-to-speech presented via a communicatively coupled speaker), and/or haptic feedback.
  • Figure 1 K shows further updated orchestrated guidance.
  • Figure 1 K shows a third orchestrated guidance 153 and a third set of recommended actions (e.g., Ul elements 154 and 155) presented at a wearable device.
  • the Al agent 115 determines the third orchestrated guidance 153 and the third set of recommended actions based on the notifications received while the wearable devices (or other devices of the Al guidance system 100) were in the do-not-disturb mode.
  • the third orchestrated guidance 153 and the third set of recommended actions provide the user 105 with options for responding to received messages and missed calls.
  • the user 105 forgoes selecting the third set of recommended actions.
  • the user 105 provides a touch input 157 at the headwearable device 120 to initiate a microphone of the head-wearable device 120 (or other communicatively coupled device) and provide a voice command 151 to the Al guidance system 100.
  • the voice command provided to the Al guidance system 100 can be used by the Al agent to determine another context-based activity (e.g., organizing dinner plans).
  • the Al agent 115 can generate for the other context-based activity additional orchestrated guidance recommended action for performing the other context-based activity.
  • the Al agent 115 can generate orchestrated guidance for organizing dinner plans and recommended actions.
  • Figure 1 L shows the Al guidance system 100 utilizing a web-agent to assist the user in the performance of the other context-based activity and/or determine recommended actions.
  • the Al guidance system 100 in response to a user input selecting the recommended action for performing the context-based activity, can perform, using the Al agent, a (web or application) search based on the recommended action.
  • the Al guidance system 100 can further determine a task to perform based on the search, and presenting the task at the wearable device.
  • the Al guidance system 100 receives a request from a user to cause an Al agent to perform a task (e.g., “find a restaurant for dinner tomorrow downtown and make a reservation for 4”) and, based on content of the request, the Al guidance system 100 can determine that traversal of one or more web pages is required to perform the task that fulfills the request from the user. Further, the Al guidance system 100, responsive to the request, can traverse, using a web-based Al agent, one or more web pages and/or applications and, after the traversing, process the collected data to generate, via the Al agent, the response for the user 105 (e.g., response identifying a restaurant for 4 people and a time for making reservations).
  • a request from a user to cause an Al agent to perform a task (e.g., “find a restaurant for dinner tomorrow downtown and make a reservation for 4”) and, based on content of the request, the Al guidance system 100 can determine that traversal of one or more web pages is required to perform the task that fulfills the request from the user
  • the Al guidance system 100 use the web-agent to autonomously carry out requests made by the user 105 even when the request is not associated with an API. In some examples, the Al guidance system 100 will report back on progress made in fulfilling the request of the user 105. For example, the Al guidance system 100 can report to the user 105 restaurant availability, restaurant wait times, errors in booking, reservation confirmations, etc. For example, as shown in Figure 1 L, the Al agent 115 identifies a restaurant and a reservation time for organizing the user 105’s dinner plans, and the Al guidance system 100 presents the restaurant and the reservation time to the user 105 via the wearable device (e.g., response Ul element 159).
  • the wearable device e.g., response Ul element 159
  • the Al guidance system 100 can utilize the web-agent (application-agent and/or other computer implemented agent) to assist the user 105 in collecting additional information for fulfilling the request from the user 105.
  • the Al guidance system 100 can search information related to social media posts to identify restaurant recommendations and/or restaurants in proximity and provide the information related to the social media posts to the Al agent 1 15 for generating a response and/or providing recommended actions.
  • the information is determined through the use of an Al model that is configured to determine additional information from images/videos/audio to provide contextual information (e.g., a picture of a posted restaurant and use an Al to determine which restaurant the poster was at).
  • the Al guidance system 100 can provide the user 105 with information about a previously seen social media post. In some examples, the Al guidance system 100 can be used to find additional information on posts or other content the user 105 has previously viewed via one or more devices, thereby providing unique results specific to the user’s viewing history.
  • the Al guidance system 100 can perform additional Al actions to assist the user 105 and/or augment the user 105’s experience.
  • the Al guidance system 100 can proactively provide or silence notifications based on user situations determined by the Al agent 1 15 (e.g. the Al guidance system 100 can detect ongoing activities of the user 105 based on sensor data, audio data, and/or image data, and determine situations would benefit from additional focus (e.g., productivity tasks, participation in events, etc.) and silence non-essential notifications until the situations are complete).
  • the Al guidance system 100 can also proactively display information that is determined to be essential to the user 105 and/or predicted to be useful to the user 105 based on the environment of the wearable devices and/or other devices of the Al guidance system 100.
  • a wearable device such as the head-wearable device 120
  • the Al guidance system 100 operations can occur without the need of user input (e.g. .touch inputs, voice commands, etc.).
  • Figure 1 M illustrate orchestrated guidance based on the restaurant identified by the Al guidance system 100.
  • the Al guidance system 100 presents via the wearable devices a fourth orchestrated guidance 162 and a fourth set of recommended actions (e.g., Ul elements 163, 164, and 165).
  • the user 105 provides another voice command 161 to the Al guidance system 100 for performing an action corresponding to the orchestrated guidance for organizing dinner plans.
  • the user 105 performs a pinch and hold gesture to initiate a microphone of the head-wearable device 120 (or other communicatively coupled device) and provide the other voice command 161 to the Al guidance system 100.
  • Figure 1 N shows the Al guidance system 100 providing confirmation of a completed task and generating an event for the user 105.
  • the Al guidance system 100 causes presentation of a task completion Ul element 167 via a display of the headwearable device 120. Additionally, the Al guidance system 100 also presents a calendar Ul element 169 showing an event or calendar invite generated by the Al agent 1 15.
  • the Al guidance system 100 can be used to augment user experience of other activities.
  • the Al guidance system 100 can be used for a cooking context-based activity and the Al guidance system 100 can be used by the user 105 to find a recipe, make a dish based on the recipe, present guidance on preparation of the dish based on the recipe (e.g., step-by-step instructions, illustration, and/or video).
  • the Al guidance system 100 can use sensor data, audio data, and/or image data of wearable devices and/or other devices to determine a current step of the recipe and/or progress made by the user 105.
  • the user 105 can query the Al guidance system 100 on the next step of the recipe, and the Al guidance system 100 can provide tailored instructions to the user 105.
  • the Al guidance system 100 can provide information about steps of the recipe, how much time is left, determinations of food preparedness based on sensor data, audio data, image data, etc.
  • the Al guidance system 100 can augment a user experience of a game application.
  • a user can query the Al guidance system 100 to perform a task in game, and the Al guidance system 100 can leverage the one or more sensors of the wearable devices (e.g., the head-wearable device 120) and/or other devices in communication with the Al guidance system 100 to satisfy the request of the user 105.
  • the Al guidance system 100 can provide natural language responses to guide a user 105 within an augmented reality environment by using IMU data and image data (e.g., the device can state “There is a monster behind you, watch out!”).
  • the request to the Al guidance system 100 can initiate the game without the need for the user 105 to open the application themselves.
  • the Al guidance system 100 could output audio spatially to the user to help them identify where an interactable object is in a game.
  • the Al guidance system 100 can augment a user experience of a sports event or sports application.
  • the user 105 can ask the Al guidance system 100 a question about an ongoing Formula 1 race to understand the positions of the drivers - e.g., “compare the pace between two drivers.”
  • the Al guidance system 100 can be configured to use live data from the application or the sports stream to provide the appropriate response. For sports that are heavily data driven, there is a lot of data that is not provided to the user 105, but the Al guidance system 100 can access any available data (e.g., microphone communications of one driver, tire data, lap times, showing different cameras of different drivers including selecting specific cameras on each car, etc.).
  • Figures 2A-2L illustrate context-based responses generated by an artificially intelligent agent based on activities performed by a user. Similar to Figures 1A-1 N, the operations shown in Figure 2A-2L can be performed by any XR systems described below in reference to Figures 7A-7C. For example, the operations of Figures 2A-2L can be performed by wearable devices, such as a wrist-wearable device 1 10 and/or a head-wearable device 120. The operations of Figures 2A-2L are performed by an Al assistive system 200 including at least a wrist-wearable device 1 10 and a head-wearable device 120 donned by the user 105 and/or other electronic devices described below in reference to Figures 7A-7C. The Al assistive system 200 can include the Al agent 115.
  • the Al assistive system 200 is analogous to the Al guidance system 100 shown and described in reference to Figures 1A-1 N. In some examples, the Al assistive system 200 and the Al guidance system 100 are the same. Alternatively, In some examples, the Al assistive system 200 and the Al guidance system 100 are distinct system implemented at any XR systems described below in reference to Figures 7A-7C. Operations of the Al assistive system 200 and the Al guidance system 100 can be performed in parallel, sequentially, concurrently, and/or in a predetermined order.
  • the Al assistive system 200 can augment the user 105’s experience in performing a physical activity and/or user experience while using a fitness application.
  • the Al assistive system 200 can assist the use in the performance of different physical or fitness activities.
  • the Al assistive system 200 can operate as a virtual coach and emulate a coach’s voice, provide specific instructions, and/or provide feedback to the user 105.
  • one or more sensors of the head-wearable device 120 and/or communicative coupled devices to determine whether the user 105 is performing the physical activity correctly.
  • the Al assistive system 200 can provide guidance to the user 105 to improve performance of the exercise.
  • the user 105 is participating in an activity with at least one other user 205.
  • the activity is physical exercise.
  • the user 105 and the at least one other user 205 are at a gym and start performing an exercise (e.g. a run).
  • the Al assistive system 200 in response to an indication that the user 105 of a wearable device, such as the wrist-wearable device 1 10 and/or the head-wearable device 120, is participating in an activity, obtains data associated with an on-going activity performed by the user 105 of the wearable device.
  • the indication can be provided in response to a user input. For example, first user input 207 at the head-wearable device 120 initiating a workout.
  • the Al assistive system 200 can generate the indication based on sensor data, audio data, and/or image data captured by one or more devices of the Al assistive system 200.
  • the Al assistive system 200 can detect that the user 105 is engaging in a physical activity, such as running, cycling, weightlifting, skiing, etc., and generate the indication that the user 105 is participating in an activity.
  • the Al assistive system 200 can generate the indication based on audio cues or context. For example, the user comment “ready for the run?” can be used to initiate and identify the activity.
  • the Al assistive system 200 generates, using the Al agent 1 15, a contextbased response based, in part, on the data associated with the on-going activity performed by the user 105 of the wearable device and presents, at the wearable device, the context-based response.
  • the Al agent 115 can generate a workout Ul 21 1 including activity information and a first context-based response (represented by first context-based response Ul element 209), and cause presentation of the workout Ul 21 1 and the first context-based response Ul element 209 at the head-wearable device 120.
  • the context-based response is presented within a portion of a field of view 212 of the user 105.
  • the context-based response and/or the workout Ul 211 are presented such that they are always visible to the user 105.
  • the context-based response and/or the workout Ul 21 1 can be positioned at a portion of the display of the wearable device reserved for the context-based response and/or the workout Ul 21 1.
  • the context-based response and/or the workout Ul 211 can be configured such that they are always overlayed over other applications and/or Uls.
  • the context-based response is a coaching response to assist the user 105 on performance of the activity.
  • the first contextbased response Ul element 209 prompts the user 105 if they would like help with their workout.
  • the context-based response can include navigation instructions.
  • the workout Ul 211 includes activity information, such as activity information Ul element 212 and activity route 217 (or activity map).
  • the workout Ul 21 1 includes biometric data to allow the user 105 to easily track their workout.
  • the workout Ul 211 can include real-time statistics including, but not limited to, speed, pace, splits, total distance, total duration, map, segments, elevation, gradient, heart rate, cadence, persona records (or PRs), challenges, and segment comparisons.
  • the Al assistive system 200 operates in conjunction with the wearable devices to automatically select information about the physical activity to present within the user interface elements.
  • the workout Ul 211 includes one or more quick access applications 215 that allow the user 105 to initiate one or more applications.
  • the Al assistive system 200 can present and/or share data rich overlay Uls that can include image data (e.g., Figures 2L and 2C) and/or other data about activities that the user 105 is performing.
  • the Al assistive system 200 allows the user 105 to connect and engage with their communities in more interesting and engaging ways, by curating informative overlays to captured activities. For example, by providing the user 105 with capabilities for sharing personal states about physical activities that the user is performing, the Al assistive system 200 allows the user 105 to elevate and showcase their efforts and progress.
  • the Al assistive system 200 can provide visual feedback to the user 105 via frames of the head-wearable device 120.
  • the head-wearable device 120 includes one or more indicators for assisting the user 105 in performance of the activity.
  • Figure 2B shows an interior portion 219 (e.g., face-facing portion of the frames) of the head-wearable device 120, the interior portion including a first light emitter portion 221 and a second light emitter portion 223.
  • the first and the second light emitter portions 221 and 223 can be light-emitting diodes (LEDs).
  • the Al assistive system 200 can use the first light emitter portion 221 and the second light emitter portion 223 to provide directions to the user (e.g., turn the first light emitter portion 221 on and the second light emitter portion 223 off to direct the user to the left, turn on both the first and the second light emitter portions 221 and 223 to direct the user to go forward; etc.).
  • the first and the second light emitter portions 221 and 223 can turn different colors, illuminate in different patterns and/or frequencies, and/or illuminate with different brightness to provide the user 105 with biometric information (e.g., green to indicate that the heart rate of the user 105 is in a first target threshold, yellow to indicate that the heart rate of the user 105 is in a second target threshold, red to indicate that the heart rate of the user 105 is in a third target threshold, etc.)
  • biometric information e.g., green to indicate that the heart rate of the user 105 is in a first target threshold, yellow to indicate that the heart rate of the user 105 is in a second target threshold, red to indicate that the heart rate of the user 105 is in a third target threshold, etc.
  • the user 105 responds to the first context-based response via a voice command 225.
  • the user 105 requests that the Al assistive system 200 assist the user 105 in setting a PR.
  • the user 105 can provide different types of request to the Al assistive system 200
  • the user 105 can provide the voice command 225 requesting that the Al agent 1 15 notify the user 105 when their heart rate is above a predefined threshold (e.g., heart rate goes above 165 BPM).
  • the Al assistive system 200 can provides a series of visual and/or audio response to the user 105 based on the voice command 225 or other user request.
  • the visual and/or audio response can be encouragement, suggestions, instructions, updates to biometric data, etc.
  • the Al assistive system 200 can provide the audio response in distinct vocal personalities and/or other characteristics, which may be based on the type of physical activity the user is performing (e.g., a personified Al agent).
  • the Al assistive system 200 can use the voice of a famous motivational runner in accordance with detecting that the user 105 is running a 10K.
  • the Al assistive system 200 generates, via the Al agent 115, second context-based response and updates to the workout Ul 211.
  • the Al assistive system 200 can generate a response to the voice command 225 and present the response to the user 105 via a wearable device (e.g., the second context-based response Ul element 227 presented within field of view 212).
  • the workout Ul 211 can be updated to show changes to biometric data (e.g., changes to calories burned, heart rate, etc.), workout completion, split times, etc.
  • the user 105 provides the Al assistive system 200 a request to live stream their activity.
  • the Al assistive system 200 can allow the user 105 to enable a live stream using wearable devices (e.g., the head-wearable device 120 and/or the wrist-wearable device 1 10) and/or other communicatively coupled device capture and transmit image data, audio data, and/or sensor data.
  • the user 105 can provide another voice command 229 requesting that the Al assistive system 200 initiate a stream to capture their run.
  • the user 105 can initiate the live stream via a touch input at the wrist-wearable device 1 10 and/or the head-wearable device 120.
  • the user 105 can perform a gesture to select one or more Ul element for selecting a particular functionality.
  • the user can perform a pinch gesture to select the streaming Ul element 234.
  • the Al assistive system 200 provides a third context-based response confirming the initiation of the stream to the user 105 (e.g., third context-based response Ul element 231).
  • the Al assistive system 200 can further present a streaming Ul 233 at the head-wearable device 120 and/or another streaming Ul 237 at the wrist-wearable device (or other communicatively coupled display).
  • the Al assistive system 200 can present static holographic elements 235 that provide simple information and/or images to the user 105.
  • the static holographic elements 235 can includes battery information, simplified notifications corresponding to stream interactions, and/or other Al agent 115 information (such as a camera view finder) can be presented.
  • the Al assistive system 200 can initiate the live stream on one or more platforms associated with the user 105.
  • the Al assistive system 200 can automatically select the streaming platform for the user 105 (e.g., based on user behavior).
  • the user 105 can provide a user input (e.g., voice command, touch input, gesture, etc.) identifying a streaming platform and/or selecting from one or more suggested streaming platforms identified by the Al assistive system 200.
  • the Al assistive system 200 notifies one or more followers of the user 105 that the live stream has been initiated.
  • the Al assistive system 200 can perform a complimentary operations to a requested operation of the user 105, which may be based on data about the user’s interaction history with the respective social platforms.
  • the streaming Ul 233 and the other streaming Ul 237 include a chat of the live stream.
  • the streaming Ul 233 and the other streaming Ul 237 can present the broadcasted stream (e.g., captured and transmitted image data, audio data, sensor data, and/or other transmitted data).
  • the user 105 can toggle information presented via the streaming Ul 233 and/or the other streaming Ul 237. For example, the user 105 can select one or more Ul elements within the streaming Ul 233 and/or the other streaming Ul 237 to toggle the presented information. Additionally, the user 105 can select a share Ul element to share additional content or information.
  • the Al assistive system 200 can apply one or more overlays and/or Ul elements to the streamed data such that the one or more overlays and/or Ul elements are viewable by devices receiving the streamed data.
  • the streamed image data can include information on the user’s current activity (e.g., current progress, percentage complete, and/or other information shared by the user 105).
  • the Al assistive system 200 can provides automatic user interactions by automatically engaging the user 105 and/or with other communicatively coupled devices with streamed data.
  • Figure 2G and 2H shows the Al assistive system 200 connecting the user 105 with the at least one other user 205.
  • the Al assistive system 200 in accordance with a determination that the activity is a group activity performed with at least one contact ofthe user 105 (e.g., a friend or connection of the user 105 ), obtains from an electronic device associated with the at least one contact of the user 105 additional data associated with a respective on-going activity performed by the at least one contact of the user 105.
  • the context-based response can further be based on the additional data associated with the respective on-going activity performed by the at least one contact of the user.
  • the Al assistive system 200 presents via a display of the headwearable device 120 a context-based response (e.g., a fourth context-based response 239) prompting the user 105 if they would like to connect with a contact (e.g., the at least one contact 205), as well as an updated workout Ul 21 1 including a pin 241 or flag of a position of the at least one contact 205 relative to a current position of the user 105.
  • Figure 2H further shows the user 105 providing a user input (e.g., yet another voice command 243) requesting that data be shared with the at least one contact 205.
  • the Al assistive system 200 provides a plurality communication modalities in which the user 105 can quickly connect with friends and/or contacts.
  • the Al assistive system 200 can be used to contact a single contact participating in a group activity or all contacts participating in the group activity.
  • the Al assistive system 200 can include one or more communication channels.
  • the Al assistive system 200 can include a walkie-talkie feature to quickly and effortlessly connect with one or more contacts.
  • the Al assistive system 200 can identify one or more participants in a group activity based on proximity data of one or more devices adjacent to wearable devices of the Al assistive system 200.
  • the Al assistive system 200 can identify one or more participants in a group activity based on electronic devices attempting to communicatively couple with the wearable devices and/or other devices of the Al assistive system 200. In some examples, the Al assistive system 200 can identify one or more participants in a group activity based on the user 105’s contact list and/or by reviewing recent group conversations about an event or activity. In some examples, the Al assistive system 200 uses natural language systems to invoke a conversation with a group and quickly communicate with the group.
  • Figure 2I and 2J show a perspective of the at least one contact 205.
  • Figures 2I and 2J show another Al assistive system (analogous to the Al assistive system 200) implemented on one or more wearable devices or other devices of the at least one contact 205.
  • the other Al assistive system presents via a speaker of a headwearable device 253 of the at least one contact 205 a context-based response 245 prompting the at least one contact 205 if they would like to connect with the user 105.
  • the at least one contact 205 further provides a voice command confirming that they would like to connect with the user 105.
  • Figure 2J shows a field of view 246 of the at least one contact 205 as viewed by the head-wearable device 253.
  • the field of view 246 of the at least one contact 205 includes a first workout Ul 249 tracking the at least one contact 205’s workout and a second workout Ul 247 including shared workout information from the user 105.
  • the first workout Ul 249 further includes a respective pin 250 identifying the location of the user 105 relative to the at least one contact 205.
  • Figure 2J further shows the at least one contact 205 providing the other Al assistive system a request. For example, the request 251 from the at least one contact 205 asks the other Al assistive system to send an encouraging message to user 105.
  • the Al assistive system 200 of the user 105 can receive the encouraging message and automatically cause presentation of the visual and/or audio message.
  • the encouraging message can include a haptic feedback response.
  • the Al assistive system 200 presents the encouraging message after determining that the user 105 has achieved a particular milestone related to the performance of the activity.
  • users are able to unlock pre-recorded praise from the Al assistive system 200 (e.g., personified Al agents) and/or pre-recorded audio by professional athletes related to the physical activities that the user is performing.
  • Figure 2J further shows one or more indicators 255 on the head-wearable device 253 of the at least one contact 205.
  • the indicators 255 of the head-wearable device 253 of the at least one contact 205 can be one or more light-emitters (e.g., LEDs). Similar to the first and second light emitter portions 221 and 223, the indicators 255 can communicate information to the at least one contact 205.
  • the indicators 255 can illuminate in different colors, patterns and/or frequencies, and/or brightness to convey information to the at least one contact 205.
  • the indicators 255 can illuminate to notify the at least one contact 205 when they are within target activity thresholds, performing an activity at a predetermined pace or speed, etc.
  • the indicators 255 provides a persistent indication to the at least one contact 205 based on whether a particular condition satisfies a predefined threshold. For example, based on the at least one contact 205 providing a user input activating the indicators 255, the indicators 255 can remain active until disabled.
  • the head-wearable device 253 are a pair of low-cost headwearable device that do not include a display and opt for presenting information via audio outputs and/or haptic feedback to the at least one contact 205.
  • the head-wearable device 253 can include low fidelity display that is configured to provide glanceable information.
  • this information may be text and glyphs (e.g., emoji’s, gifs, or low-resolution images) only, as opposed to media rich images (e.g., video or color images).
  • the low-fidelity display can be configured to display a single color (e.g., green) or grayscale.
  • the head-wearable device 253 can include an outward facing projector configured for displaying information.
  • the head wearable device 253 can be configured to display a text message onto a wearer’s hand or other surface.
  • the head-wearable device can project user interfaces such that a wearer can interact with a desktop-like user interface without needing to bring a laptop with them.
  • head-wearable devices are shown as having different features it is envisioned that a single head-wearable device could be configured to use all or a subset of these information presenting modalities.
  • the Al assistive system 200 can include different modalities for presenting and/or sharing information. While numerous modalities are discussed, it is envisioned that an operating system would be configured to present the information based on the device, and the developer would only need to specify the content to be presented and not the specific modality. In this way software can be produced to work across head-wearable devices with different capabilities (e.g., information output modalities). All of these devices described are configured to work with Al models for presenting information to users.
  • Figure 2K and 2L show additional data collected and/or shared during the performance of an activity (or group activity).
  • Figures 2K and 2L show image data collected during the performance of the group activity, shared image data between the members of the group activity, and/or synchronization of the image data.
  • the other Al assistive system presents via a speaker or a display of the head-wearable device 253 of the at least one contact 205 another context-based response 257 prompting the at least one contact 205 if they would like receive and synchronize image data shared by the user 105.
  • the at least one contact 205 further provides a voice command confirming that they would like to connect receive and synchronize image data shared by the user 105.
  • the Al assistive system 200 include sharing operations for creating and sharing user interfaces that include imaging data captured by the intelligent autocapture assistive operations.
  • the Al assistive system 200 provides user interfaces that include image data that is captured while a user performing a physical activity (e.g., a fitness activity, such as performing a bike ride).
  • the user interfaces also include user interface elements generated based on other data, different than the image data, related to the user’s performance of the respective physical activity.
  • the Al assistive system 200 is configured to allow users to tag captured media with personal metadata (e.g., real-time statistics).
  • the user interfaces may include engaging montages of captured images and other content about the performance of the physical activity.
  • an image sync Ul 259 can be configured display captured image data, combined image data (e.g., combined first image data 261 and second image data 263), and/or image montages.
  • the image sync Ul 259 can be presented at other devices of the Al assistive system 200.
  • the Al assistive system 200 in accordance with a determination that a plurality of video streams are (i) captured within a predefined amount of time of each other and (ii) within a predefined distance of each other, prepares a collated video of two or more of the plurality of video streams in a time- synchronized fashion.
  • Figures 3A-3D illustrate example user interfaces and additional features available at the Al assistive system 200.
  • Figures 3A and 3B show a map application and directions provided via the Al assistive system 200.
  • Figures 3C and 3D show automatic image capture capabilities of the Al assistive system 200.
  • the Al assistive system 200 presents a map Ul 307.
  • the map Ul 307 can include one or more Ul elements providing directions to the user 105.
  • the map Ul 307 can include a next step Ul element 309 including the next directions to take, as well as a path highlight 305 (which can be overlaid over the next path in the directions).
  • the user 105 can toggle between application via one or more user inputs.
  • the user 105 can cause presentation of the map Ul 307, via a wearable device of the Al assistive system 200, in response to user selection of the map application Ul element 308.
  • the Al assistive system 200 presents context-based responses 305 providing directions to the user 105.
  • Figure 3B shows a map settings Ul 313.
  • the map settings Ul 313 can be presented in response to user input 31 1 (selecting the downward arrow).
  • the map settings Ul 313 provides one or more options for allowing the user 105 to select settings for voiced directions (e.g., on, off, and/or a particular voice), visual direction indicators (e.g., path highlights, next step Ul elements, etc.), view (e.g., setting 2D, 3D, and/or street views), location sharing (e.g., privacy setting for sharing location, automatic sharing of location, etc.), etc.
  • voiced directions e.g., on, off, and/or a particular voice
  • visual direction indicators e.g., path highlights, next step Ul elements, etc.
  • view e.g., setting 2D, 3D, and/or street views
  • location sharing e.g., privacy setting for sharing location, automatic sharing of location, etc.
  • the Al assistive system 200 presents an image capture Ul 317.
  • the image capture Ul 317 can include one or more Ul elements for showing captured image data and/or options 323 for modifying, sharing, and/or dismissing the captured image data.
  • the image capture Ul 317 can include first and second image data 319 and 321 captured during the activity of the user 105.
  • the user 105 can toggle between application via one or more user inputs.
  • the user 105 can cause presentation of the image capture Ul 317, via a wearable device of the Al assistive system 200, in response to user selection of the image application Ul element 318.
  • the Al assistive system 200 presents contextbased responses 315 providing information on the automatically captured image data.
  • Figure 3D shows a capture settings Ul 327.
  • the capture settings Ul 327 can be presented in response to user input 325 (selecting the downward arrow).
  • the capture settings Ul 327 provides one or more options for allowing the user 105 to select settings for capture triggers (e.g., triggers that cause the automatic capture of image data, such as changes in movement, instant spikes in acceleration, activity milestones (e.g., hitting a baseball with the baseball bat), changes in vibration, etc.), capture settings (e.g., image capture setting such as resolution, format, frames per second, etc.), tagging options (e.g., settings identifying people and/or objects to be tagged), sharing options (e.g., privacy setting for sharing image data, identifying images that can be shared, frequency at which image data is shared, etc.), etc.
  • the Al assistive system 200 is configured to perform sharing operations based on the user input in accordance with determining that the user has already enabled the automatic image-capture operations.
  • the Al assistive system 200 is configured
  • Figures 4A and 4B illustrate example sequences of user interactions with personalized assistive systems (e.g., the Al guidance system 100 and/or the Al assistive system 200; Figures 1 A-2L).
  • the legend in the top right of Figures 4A and 4B indicates types of interactions and input modes for each respective segment of the timeline flow.
  • the task icon 401 indicates a productivity-based interaction
  • media-play icon 405 indicates media and/or an “edutainment” interaction
  • the messaging icon 407 indicates a communicationbased interaction
  • the information icon 409 indicates an information-based interaction
  • the solid line 41 1 indicates a touch input
  • the double line 413 indicates a wake word input
  • the triple line 415 indicates an Al chat session.
  • the interaction sequences of Figures 4A and 4B can be performed by a user that is wearing a head-worn device 120 (e.g., AR device 728) while the user of the device is performing a sequence of daily activities.
  • the head-worn device 120 ( Figures 1A-3D) includes or is in electronic communication with an assistive system for assisting in interactions with the head-worn device 120 to cause operations to be performed.
  • the head-worn device 120 may provide information (e.g., information related to data collected about a physical activity that a user is performing, an alert about an incoming message) without explicit user input to do so.
  • the user can perform voice commands to cause operations to be performed at the head-worn device 120.
  • the user can provide a voice command to turn on do-not-disturb (DND) at their head-worn device 120, with an option set for VIP exceptions, which would allow for certain users’ messages or other requests may be allowed.
  • DND do-not-disturb
  • the assistive system in accordance with receiving the request to turn on do not disturb, determines a set of potential operation commands that the request may correspond to.
  • the assistive system can determine to check one or more messenger threads accessible via the head-worn device 120 to determine a bike ride location for a bike ride that the user is participating in.
  • the assistive system performs the operations in response to a question by the user that does not directly provide instructions to search the user’s messages for the bike ride location.
  • the assistive system is capable of performing a set of operations based on a general prompt provided by the user.
  • the head-worn device 120 can automatically begin providing real-time navigation (e.g., via the assistive system or a different navigational application) to the user based on determining that the user is performing a bike ride along a particular navigational route. That is, the assistive system may be capable of determining when a user is performing an activity that can be enhanced by content from a different application stored in memory or otherwise in electronic communication with the head-worn device 120 (e.g., an application stored on the user’s smart phone).
  • the head-worn device 120 can provide message readouts from a group message for fellow cyclists to keep the user informed about updates in the chat while the user is performing the physical activity.
  • the head-worn device 120 can provide capabilities for the user to send and receive voice messages to other members of the cycling group chat.
  • the head-worn device 120 can cause the userto receive a text message (e.g., an audio readout of the text message) based on a determination that the message sender is from a user that qualifies under the VIP exceptions for the do not disturb setting that was instantiated at block 402. That is, In some examples, the assistive system can determine whether a particular received message should be provided for audio readout to the user based on settings of a different application.
  • a text message e.g., an audio readout of the text message
  • the assistive system can determine whether a particular received message should be provided for audio readout to the user based on settings of a different application.
  • the head-worn device 120 can cause a different group thread (e.g., a noisy group thread) to be silenced, such that audio readouts are not provided by the particular messaging thread.
  • the assistive system can unmute and catch up on soccer group thread in messenger after ride.
  • the assistive system can allow the user to message soccer group thread in messenger in response to a received message from the group thread.
  • the assistive system can allow a user to record a voice note about new commitments to soccer group, which may be provided to the user by the assistive system based on a prompt inquiring about the user’s availability for a particular event and/or time.
  • the assistive system can allow the user to look up local family events happening this weekend (e.g., by providing a general prompt about the user’s availability).
  • the assistive system can provide the information to the user about the family events based on a different event that has occurred at the head-worn device 120 (e.g., receiving a different message from a different user about the user’s availability to participate in a cycling event).
  • the user can receive a summary of a specific family event, for example, in accordance to provide an input in response to receiving the information about local family events happening that weekend.
  • the user can provide an input (e.g., “Hey Al assistant, repeat that on my phone”) to cause a previous audio message from the assistive system to be provided at a different electronic device (e.g., “Play last Al response on phone speaker for child to hear”).
  • the user can also share feedback from the assistive system (e.g, an Al response) with another user (e.g., the user’s partner) on a different application, different than the application that is providing the assistive system (e.g., a messaging application).
  • the user can receive a real-time game notification from sports app.
  • the user can cause the assistive system to provide on- demand translation for audio or textual content in another language.
  • the on-demand translation can be provided automatically based on a user request to read out content that is not in the user’s native language.
  • the user can request slower speed translation.
  • the user can receive voice messages from the cycling group on messenger.
  • the user can mute a noisy messenger group chat, which the assistive system may be configured to automatically recognize based on a frequency that electronic messages are being received by the head-worn device 120 or another electronic device in electronic communication with the head-worn device 120.
  • the user can check messages.
  • the assistive system can provide a notification to the user about a geographic landmark that the user is in proximity too (e.g., as determined by a navigational application on the user’s phone (e.g., “Location Update: At Farmer’s Market”).
  • the assistive system can be configured to provide new recipe ideas for a new ingredient (e.g., an ingredient purchased at the farmer’s market).
  • the suggestions can be provided in accordance with receiving purchase confirmation at the head-wearable device about a purchase that the user made at the farmers’ market.
  • Figure 4B illustrates another timeline view of another interaction sequence with a head-worn device 120 (e.g., AR device 700) while a user of the device is performing a sequence of daily activities.
  • the user can engage in an Al chat session (as indicated by the red segment) to perform various activities to start their day (e.g., block 446 to check the local time while traveling, block 448 to set an alarm to leave for the airport later, block 450 to check the weather to decide what to wear, block 452 to check the calendar for a time and/or location of the next event, block 454 to look up local business address and hours, block 456 to message a colleague, and block 458 to listen to news on a podcast).
  • the assistive system can be configured to automatically stop the Al chat session.
  • the user can perform a sequence of touch inputs, which may be used to cause the assistive operations to perform various functions, including those related to the audio outputs of the assistive system (e.g., block 460 to receive a text message reply from a colleague, block 462 to replay to the text message, block 464 to resume a podcast, block 466 to book a rideshare to an upcoming event, block 468 to receive a notification about the arrival of the rideshare, block 470 to check status of the rideshare, block 472 to call the rideshare to clarify pickup location, block 474 to listen to a music playlist while chatting, block 476 to receive an alarm to leave for the airport, block 478 to check a status of a flight, block, 480 to receive a reminder to buy a gift before departure of the flight, block 482 to call a partner on a messaging application, and block 484 to listen to meditation for the user’s flight anxiety).
  • block 460 to receive a text message reply from a colleague
  • block 462 to replay to the text message
  • the touch inputs provided by the user corresponding to one or more of blocks are based on universal gestures corresponding to universal inputs at the AR device 728, while one or more other blocks may correspond to user inputs provided to contextual input prompts (e.g., in response to an assistive prompt provided by the head-worn device 120).
  • the systems described herein allow users to interact with an assistive system provided at the head-worn device 120 to allow for increased efficiency and effectiveness of the user’s interactions with the head-worn device 120.
  • the assistive system can allow for the user to use the head-worn device 120 as a tool to help level up their efficiencies, including by allowing for multi-tasking and productivity on the go.
  • the assistive systems and devices described herein also allow the user to interact with the assistive system relatively inconspicuously, allowing for them to perform actions without distracting others around them.
  • Figures 5 and 6 illustrates flow diagrams of methods of generating Al contextbased response and actions.
  • Operations (e.g., steps) of the methods 500 and 600 can be performed by one or more processors (e.g., central processing unit and/or MCU) of an system XR system (e.g., XR systems of Figures 7A-7C-2). At least some of the operations shown in Figures 5 and 6 correspond to instructions stored in a computer memory or computer-readable storage medium (e.g., storage, RAM, and/or memory).
  • processors e.g., central processing unit and/or MCU
  • system XR system e.g., XR systems of Figures 7A-7C-2
  • Figures 5 and 6 correspond to instructions stored in a computer memory or computer-readable storage medium (e.g., storage, RAM, and/or memory).
  • Operations of the methods 500 and 600 can be performed by a single device alone or in conjunction with one or more processors and/or hardware components of another communicatively coupled device (e.g., wristwearable device 1 10 and a head-wearable device 120) and/or instructions stored in memory or computer-readable medium of the other device communicatively coupled to the system.
  • another communicatively coupled device e.g., wristwearable device 1 10 and a head-wearable device 120
  • the various operations of the methods described herein are interchangeable and/or optional, and respective operations of the methods are performed by any of the aforementioned devices, systems, or combination of devices and/or systems.
  • the method operations will be described below as being performed by particular component or device, but should not be construed as limiting the performance of the operation to the particular device in all examples.
  • FIG. 5 shows a flow chart of a method 500 for generating orchestrated guidance based on an activity of a user.
  • the method 500 occurs at a wrist-wearable device 110, head-wearable device 120, and/or other wearable device including one or more sensors, imaging devices, displays, and/or other components described herein.
  • the method 500 includes in response to an indication received at a wearable device that an artificial intelligence (Al) agent trigger condition is present, providing (502) an Al agent sensor data obtained by the wearable device.
  • Al artificial intelligence
  • a wrist-wearable device 1 10 and/or a head-wearable device 120 of a user can use image data, location data, audio data, and/or other data to detect the presence of an Al agent trigger condition.
  • Al agent trigger conditions include user queries, objects of interest, locations of interest, people of interest, time of day, user invocation, etc.
  • the method 500 includes determining (504), by the Al agent, a context-based activity based on the sensor data obtained by the wearable device.
  • the context-based activity is an interpretation of a particular activity, action, and/or event with which the user is engaged.
  • the context-based activity is a museum visit or museum tour.
  • Non-limiting examples of context-based activities include shopping, driving, sightseeing, traveling, exploring, cooking, gardening, tours, social meetings, productivity based tasks (e.g., working, note takings, etc.), exercising, etc.
  • the method 500 includes generating (506), by the Al agent, orchestrated guidance based on the context-based activity and presenting (508) the orchestrated guidance at the wearable device.
  • the orchestrated guidance includes a recommended action for performing the context-based activity.
  • the orchestrated guidance can be a single recommended action, a sequence of recommended actions, and/or or concurrent (and/or parallel) recommended actions for performing the context-based activity.
  • the orchestrated guidance can be one or more recommended actions for facilitating the user’s museum tour, such as a recommended action for placing the user devices on “do not disturb,’’ a recommended action for initiating a guided tour, recommended actions for exploring museum exhibits, presentation of a summary collating missed notifications and/or messages while the user was engaged in the tour, and recommended actions for responding to the missed notifications and/or messages.
  • the orchestrated guidance can be number of recommended actions for assisting the user in performance of the context-based activity - e.g., actions to be performed, during, or after the context-based activity.
  • the context-based activity is a first context-based activity
  • the sensor data is first sensor data
  • the orchestrated guidance is first orchestrated guidance
  • the recommended action is a first recommended action
  • the method 500 further includes, in accordance with a determination that the first recommended action for performing the first context-based activity was performed (or was ignored), providing the Al agent second sensor data obtained by the wearable device, determining, by the Al agent, a second contextbased activity based on the second sensor data obtained by the wearable device, generating, by the Al agent, second orchestrated guidance based on the second context-based activity and presenting the second orchestrated guidance at the wearable device.
  • the second orchestrated guidance including a second recommended action for performing the second context-based activity.
  • the method can build on different recommended actions and/or orchestrated guidance.
  • the user can accept one or more recommended actions (e.g., Figures 1A-1J) and/or cause the Al agent to generate new recommended actions (e.g., Figures 1 K-1 N - initiating a new context-based activity of searching for a restaurant).
  • the context-based activity is a first context-based activity of a plurality of context-based activities determined by the by the Al agent based on the sensor data
  • the orchestrated guidance includes a plurality of recommended actions for performing the plurality of context-based activities
  • the recommended action is a first recommended action of the plurality of recommended actions, the first recommended action being configured to perform the first context-based activity
  • presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions.
  • generating the orchestrated guidance includes determining a subset of the plurality of recommended actions for performing the first contextbased activity, and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions and the subset of the plurality of recommended actions for performing the first context-based activity.
  • a plurality recommended of actions associated with a context-based activity can be presented to the user. For example, as shown and described in reference to at least Figure 1A, at least two recommended actions are presented to the user in accordance with a determination that the user is visiting a museum.
  • generating the orchestrated guidance includes determining a sequence of context-based activities of the plurality of context-based activities to be performed, including a second context-based activity to follow the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action and the second recommended action of the plurality of recommended actions for performing the plurality of context-based activities.
  • a string of recommended actions are presented to the user and the recommended actions are updated based oner inputs selecting one or more user inputs.
  • deviations from the recommended actions are shown and described in reference to at least Figures 1 K- 1 N.
  • the method 500 includes, in response to a user input selecting the recommended action for performing the context-based activity, causing the wearable device to initiate a do-not-disturb mode (or focus mode, away mode, etc.). While in the do-not-disturb mode, the wearable device suppresses, at least, received notifications, and in response to an indication that participation in the context-based activity ceased causing the wearable device to cease the do-not-disturb mode; generating, by the Al agent, a notification summary based on the notifications received while the wearable device was in the do-not-disturb mode; and presenting the notification summary at the wearable device. Examples of the do-not-disturb mode and the notification summary are shown and described in reference to at least Figures 1A-1 J.
  • the method 500 includes, in response to a user input selecting the recommended action for performing the context-based activity, performing, by the Al agent, a search based on the recommended action, determining a task to perform based on the search, and presenting the task at the wearable device.
  • An example search request provided by a user is shown and described in reference to at least Figures 1 K and 1 L.
  • presenting the orchestrated guidance at the wearable device includes, at least one of causing presentation of a user interface element associated with the orchestrated guidance at a communicatively coupled display, and causing presentation of audible guidance associated with the orchestrated guidance at a communicatively coupled speaker.
  • An examples of one or more user interface elements associated with the orchestrated guidance and audible guidance are shown and described in reference to at least Figure 1 H.
  • the context-based activity is to be performed at a physical activity.
  • the context-based activity can be an exercise and a recommended action is performance of a particular routine or exercise (detected by the wearable device or another communicatively coupled device).
  • a method includes receiving sensor data from one or more sensors of a head-wearable device and in response to receiving the data from the one or more sensors of the head-wearable device, processing the data, via an Al agent, to analyze the sensor data to identify a task performed or to be performed by a user, and causing the Al agent to provide guidance associated with performance of the task.
  • a head-wearable device 120 can cause performance of the operations shown and described in reference to Figures 1A-1 N.
  • the causing occurs in response to a selection at a wrist-wearable device of a user interface element that indicates that a guided tour is available.
  • a user interface element that indicates that a guided tour is available.
  • user interface elements corresponding to a guided tour can be presented at a head-wearable device 120 and/or a wrist-wearable device 1 10.
  • the sensor data from the one or more sensors is one or more of microphone data, camera data, movement data, and positioning data.
  • sensor data captured by the wrist-wearable device, the headwearable device, and/or any other communicatively couple device can be used by the Al agent.
  • the method further includes, after causing the Al agent to provide guidance associated with the task, receiving additional sensor data from the one or more sensors of the head-wearable device, in response to receiving the additional sensor data from the one or more sensors of the head-wearable device, processing the additional sensor data, via the Al agent, to identify an additional task performed or to be performed by the user, and causing the Al agent to provide guidance associated with the additional task.
  • the Al agent can determine subsequent tasks based on additional data received.
  • the additional task is related to the task.
  • a method includes receiving a request at an Al agent to (i) forgo immediate output of incoming notifications and (ii) provide a summary of the incoming notifications at a later time, receiving a plurality of notifications, providing the notifications to a large language model (LLM), producing, using the LLM, a summary of the plurality of notifications, and providing a natural language summary, via an output modality of a head-wearable device, at the later time. Examples of summarized notifications are shown and described in reference to Figure 1 J.
  • a method includes receiving a request from a user interacting with an Al agent, the request requiring traversing content on a website using the Al agent. The method also includes, in response to receiving the request, traversing, using an computer-implemented agent associated with the Al agent, one or more graphical user interfaces associated with the website to collect data needed to formulate a response to the request from the user, and after the traversing, processing the data collected by the computer-implemented agent associated with the Al agent to generate the response and providing the response to the user.
  • the Al agent can utilize a web agent to search webpages and/or perform a web search to complete a user request and provide a corresponding response.
  • the web-based Al agent is distinct from the Al agent that received the task request. In some examples, different training data used to train that Al agent and the web-based agent. In some examples, the traversing the one or more web pages includes obtaining data needed to formulate a response to the request from the user. In some examples, surface Ul element related to progress of the Al agent in performing the traversal is presented (e.g., an Al agent symbol moving or spinning to show progress). In some examples, the web-based agent can be used to inquire about a contact (e.g., ask about a particular person that may be a contact of the user - e.g., “What kind of trip would Mike go on?”).
  • the context-based activities are further determined based on stored user data (e.g., use data about the user’s previous experiences and/or interests to curate the information about the guided tour). For example, if the user previously participated in an experience that was relevant to an aspect of the guided tour (e.g., Figures 1A-1 N), the Al agent may cause information about the previous event to surface or otherwise be integrated into the guided tour.
  • stored user data e.g., use data about the user’s previous experiences and/or interests to curate the information about the guided tour.
  • the Al agent may cause information about the previous event to surface or otherwise be integrated into the guided tour.
  • a head-wearable device is a display-less AR headset.
  • the input/output interface of the head-wearable device only includes one or more speakers.
  • the operations of the head-wearable device can be performed by a set of earbuds or other head-worn speaker device.
  • a user interface associated with the orchestrated set of guidance instructions is provided by the Al agent via a Lo-Fi display, the Lo-Fi being a glanceable display that presents notifications, live activities, Al agent information, and messages.
  • a user interface associated with the orchestrated set of guidance instructions is provided by the Al agent via a projector display, the projector display configured to project information at a hand of the user (e.g., at a palm or other body part of the user).
  • a non-textual user interface element is presented at the head-wearable device (e.g., an audio message, an arrow or similar symbol), and the non-textual user interface element is configured to direct a user of the head-wearable device toward a physical landmark as part of the orchestrated set of guidance instructions.
  • the user can select objects within a field of view of the user (e.g., captured by one or more sensors of a wearable device, such as an imaging device) to receive additional information on the selected object.
  • a wearable device such as an imaging device
  • the Al agent may cause some notifications to be muted during the guided tour, and then provide with an Al-generated summary of the conversations later so that the user can quickly catch up without reviewing many different messages right away.
  • FIG. 6 shows a flow chart of a method 600 for facilitating performance of a physical activity performed by user.
  • the method 600 occurs at a wrist-wearable device 110, a head-wearable device 120, and/or other wearable device including one or more sensors, imaging devices, displays, and/or other components described herein.
  • the method 600 includes, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining (602) data associated with an on-going activity performed by the user of the head-wearable device.
  • the method 600 includes generating (604), by an Al agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device, and presenting (606), at the headwearable device, context-based response.
  • the context-based response is presented within a portion of a field of view of the user. For example, as shown and described in reference to Figures 2A-2H, a head-wearable device 120 can present different context-based responses to the user based on a physical activity being performed.
  • the method 600 includes, in accordance with a determination that the activity is a group activity performed with at least one contact of the user, obtaining, from an electronic device associated with the at least one contact of the user, additional data associated with a respective on-going activity performed by the at least one contact of the user.
  • the context-based response is further based on the additional data associated with the respective on-going activity performed by the at least one contact of the user.
  • an Al agent can detect other contacts performing an activity with a user and share information between the users.
  • the data associated with the on-going activity performed by the user of the head-wearable device and the additional data associated with the respective on-going activity performed by the at least one contact of the user includes respective image data and/or audio data
  • the context-based response is an image response including a combination of the respective image data.
  • image data captured between the wearable devices can be synchronized, combined into a single image, and/or combined into an image collage.
  • the respective image data includes a plurality of video streams from a plurality of respective head-wearable devices
  • generating, by the Al agent, the context-based response includes in accordance with a determination that the plurality of video streams are (i) captured within a predefined amount of time of each other and (ii) within a predefined distance of each other, preparing a collated video of two or more of the plurality of video streams in a time-synchronized fashion.
  • the method includes providing to each of the respective head-wearable devices the collated video. At least one aspect of the collated video provided to each of the respective head-wearable devices is tailored to that respective head-wearable device.
  • the activity is a physical exercise; and the context-based response is a coaching response to assist the user on performance of the physical exercise.
  • an Al agent can coach a user through an exercise.
  • the activity is outdoor physical activity (e.g., running, biking, hiking, etc.), and the context-based response is a navigation instructions.
  • the Al agent can provide navigation instructions to the user.
  • E7 In some examples of any one of E1 -E6, the activity is participation in a note-taking session (e.g., a meeting, class, lecture, etc.), and the context-based response is a request to generate notes. While the primary example shown in Figures 2A-2L is an exercise, the Al agent can be used with other activities performed by the user.
  • a note-taking session e.g., a meeting, class, lecture, etc.
  • the context-based response is a request to generate notes. While the primary example shown in Figures 2A-2L is an exercise, the Al agent can be used with other activities performed by the user.
  • a method is performed at a headwearable device including (i) one or more cameras, and (ii) a display component configured to display digital content.
  • the method includes determining that a user wearing the head- wearable device is performing a physical activity and, in accordance with determining that the user wearing the head-wearable device is performing the physical activity, automatically, without additional user input, initializing assistive operations based on data provided by the one or more cameras of the head-wearable device.
  • the method also includes, while the assistive operations are being performed based on image data from the one or more cameras of the head-wearable device, identifying, based on the assistive operations, that at least a portion of a respective field of view of a respective camera of the one or more cameras satisfies automatic-image-capture criteria for automatically capturing an image.
  • the method further includes, based on the identifying, causing the respective camera to capture an image automatically, without further user input. For example, as shown and described in reference to Figure 3A, a wearable device can automatically capture image data.
  • the method further includes detecting a user input directed to a universal action button on a peripheral portion of the head-wearable device.
  • the assistive operations are initialized based on the user input being detected while the user is performing the physical activity. For example, as shown and described in reference to Figure 2A, the user can perform a tap gesture at a wearable device, such as the head-wearable device, to initiate the Al agent and/or other operations.
  • a method includes receiving (i) performance data corresponding to a physical activity that a user of a head-wearable device is performing, and (ii) capturing image data by the head-wearable device during performance of the physical activity.
  • the method also includes causing presentation, at a display component of the head-wearable device, a user interface element that includes one or more representations of the performance data, and responsive to provided user preferences, automatically sharing a field of view of the user in conjunction with sharing the user interface element as a composite user interface element to one or more other electronic devices.
  • information captured by wearable devices can be shared between users.
  • the performance data is received from a software application different than another software application that is performing operations at the head-wearable device for capturing the image data.
  • the information can be received from a streaming application and/or other application.
  • a method includes determining that a user of a head-wearable device is beginning performance of a physical activity while data about the physical activity is configured to be obtained by the head-wearable device of the user and, in accordance with the determining that the user of the head-wearable device is beginning performance of the physical activity, identifying an assistive module that uses one or more specialized artificial-intelligence models.
  • the method also includes causing interactive content to be provided to the user via the assistive module based on the data obtained about the physical activity that the user is performing. For example, as shown and described in reference to Figures 2A-2D, information captured by wearable devices can be used to assist the user in performance of the activity.
  • the method further includes generating an audio message using an artificial intelligence model of the assistive module performing operations during performance of the physical activity by the user and determining based on data obtained about performance of the physical activity by the user, that one or more messageproviding criteria are satisfied.
  • the method also includes, in accordance with the determining that the one or more message-providing criteria are satisfied, generating, using an Al model, a message related to the performance of the physical activity, and providing the generated electronic message to the user via one or of (i) a microphone of the head-wearable device, and (ii) a display component within a frame of the head-wearable device.
  • a method includes, at a head-worn device including a user interface for providing user interface elements to a user based on physical activities that the user is performing, receiving an update about a location of a user, based on a physical activity that the user is performing, and in accordance with receiving the indication, presenting a navigational user interface to the user providing navigation to the user based on an identified activity that the user is performing while wearing a head-worn device.
  • a head-worn device including a user interface for providing user interface elements to a user based on physical activities that the user is performing, receiving an update about a location of a user, based on a physical activity that the user is performing, and in accordance with receiving the indication, presenting a navigational user interface to the user providing navigation to the user based on an identified activity that the user is performing while wearing a head-worn device.
  • navigation instructions can be provided to the user.
  • a system that includes one or more wrist wearable devices and a pair of augmented-reality glasses, and the system is configured to perform operations corresponding to any of A1-I1.
  • a non-transitory computer readable storage medium including instructions that, when executed by a computing device in communication with a pair of augmented-reality glasses, cause the computer device to perform operations corresponding to any of A1- 11.
  • a wearable device (a head-wearable device and/or a wrist-wearable device) configured to perform or cause performance of operations corresponding to any of A1- 11.
  • an intermediary processing device In accordance with some examples, an intermediary processing device
  • Figures 7A - 7C-2 illustrate example XR systems that include AR and MR systems.
  • Figure 7A shows a first XR system 700a and first example user interactions using a wrist-wearable device 726, a head-wearable device (e.g., AR device 728), and/or a HIPD 742.
  • Figure 7B shows a second XR system 700b and second example user interactions using a wrist-wearable device 726, AR device 728, and/or an HIPD 742.
  • Figures 7C-1 and 7C-2 show a third MR system 700c and third example user interactions using a wrist-wearable device 726, a head-wearable device (e.g., an MR device such as a VR device), and/or an HIPD 742.
  • a head-wearable device e.g., an MR device such as a VR device
  • an HIPD 742 e.g., an HIPD 742.
  • the aboveexample AR and MR systems can perform various functions and/or operations.
  • the wrist-wearable device 726, the head-wearable devices, and/or the HIPD 742 can communicatively couple via a network 725 (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN). Additionally, the wrist-wearable device 726, the head-wearable device, and/or the HIPD 742 can also communicatively couple with one or more servers 730, computers 740 (e.g., laptops, computers), mobile devices 750 (e.g., smartphones, tablets), and/or other electronic devices via the network 725 (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN).
  • a network 725 e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN.
  • a smart textile-based garment when used, can also communicatively couple with the wrist-wearable device 726, the head-wearable device(s), the HIPD 742, the one or more servers 730, the computers 740, the mobile devices 750, and/or other electronic devices via the network 725 to provide inputs.
  • a user 702 is shown wearing the wrist-wearable device 726 and the AR device 728 and having the HIPD 742 on their desk.
  • the wrist-wearable device 726, the AR device 728, and the HIPD 742 facilitate user interaction with an AR environment.
  • the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 cause presentation of one or more avatars 704, digital representations of contacts 706, and virtual objects 708.
  • the user 702 can interact with the one or more avatars 704, digital representations of the contacts 706, and virtual objects 708 via the wrist-wearable device 726, the AR device 728, and/or the HIPD 742.
  • the user 702 is also able to directly view physical objects in the environment, such as a physical table 729, through transparent lens(es) and waveguide(s) of the AR device 728.
  • an MR device could be used in place of the AR device 728 and a similar user experience can take place, but the user would not be directly viewing physical objects in the environment, such as table 729, and would instead be presented with a virtual reconstruction of the table 729 produced from one or more sensors of the MR device (e.g., an outward facing camera capable of recording the surrounding environment).
  • the user 702 can use any of the wrist-wearable device 726, the AR device 728 (e.g., through physical inputs at the AR device and/or built-in motion tracking of a user’s extremities), a smart-textile garment, externally mounted extremity tracking device, the HIPD 742 to provide user inputs, etc.
  • the user 702 can perform one or more hand gestures that are detected by the wrist-wearable device 726 (e.g., using one or more EMG sensors and/or IMUs built into the wrist-wearable device) and/or AR device 728 (e.g., using one or more image sensors or cameras) to provide a user input.
  • the user 702 can provide a user input via one or more touch surfaces of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742, and/or voice commands captured by a microphone of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742.
  • the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 include an artificially intelligent digital assistant to help the user in providing a user input (e.g., completing a sequence of operations, suggesting different operations or commands, providing reminders, confirming a command).
  • the digital assistant can be invoked through an input occurring at the AR device 728 (e.g., via an input at a temple arm of the AR device 728).
  • the user 702 can provide a user input via one or more facial gestures and/or facial expressions.
  • cameras of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 can track the user 702’s eyes for navigating a user interface.
  • the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 can operate alone or in conjunction to allow the user 702 to interact with the AR environment.
  • the HIPD 742 is configured to operate as a central hub or control center for the wrist-wearable device 726, the AR device 728, and/or another communicatively coupled device.
  • the user 702 can provide an input to interact with the AR environment at any of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742, and the HIPD 742 can identify one or more back-end and front-end tasks to cause the performance of the requested interaction and distribute instructions to cause the performance of the one or more back-end and front-end tasks at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742.
  • a back-end task is a background-processing task that is not perceptible by the user (e.g., rendering content, decompression, compression, application-specific operations)
  • a front-end task is a user-facing task that is perceptible to the user (e.g., presenting information to the user, providing feedback to the user).
  • the HIPD 742 can perform the back-end tasks and provide the wrist-wearable device 726 and/or the AR device 728 operational data corresponding to the performed back-end tasks such that the wrist-wearable device 726 and/or the AR device 728 can perform the front-end tasks.
  • the HIPD 742 which has more computational resources and greater thermal headroom than the wrist-wearable device 726 and/or the AR device 728, performs computationally intensive tasks and reduces the computer resource utilization and/or power usage of the wrist-wearable device 726 and/or the AR device 728.
  • the HIPD 742 identifies one or more back-end tasks and front-end tasks associated with a user request to initiate an AR video call with one or more other users (represented by the avatar 704 and the digital representation of the contact 706) and distributes instructions to cause the performance of the one or more back-end tasks and front-end tasks.
  • the HIPD 742 performs back- end tasks for processing and/or rendering image data (and other data) associated with the AR video call and provides operational data associated with the performed back-end tasks to the AR device 728 such that the AR device 728 performs front-end tasks for presenting the AR video call (e.g., presenting the avatar 704 and the digital representation of the contact 706).
  • the HIPD 742 can operate as a focal or anchor point for causing the presentation of information. This allows the user 702 to be generally aware of where information is presented. For example, as shown in the first AR system 700a, the avatar 704 and the digital representation of the contact 706 are presented above the HIPD 742. In particular, the HIPD 742 and the AR device 728 operate in conjunction to determine a location for presenting the avatar 704 and the digital representation of the contact 706. In some examples, information can be presented within a predetermined distance from the HIPD 742 (e.g., within five meters). For example, as shown in the first AR system 700a, virtual object 708 is presented on the desk some distance from the HIPD 742.
  • the HIPD 742 and the AR device 728 can operate in conjunction to determine a location for presenting the virtual object 708.
  • presentation of information is not bound by the HIPD 742. More specifically, the avatar 704, the digital representation of the contact 706, and the virtual object 708 do not have to be presented within a predetermined distance of the HIPD 742.
  • an AR device 728 is described working with an HIPD, an MR headset can be interacted with in the same way as the AR device 728.
  • User inputs provided at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 are coordinated such that the user can use any device to initiate, continue, and/or complete an operation.
  • the user 702 can provide a user input to the AR device 728 to cause the AR device 728 to present the virtual object 708 and, while the virtual object 708 is presented by the AR device 728, the user 702 can provide one or more hand gestures via the wrist-wearable device 726 to interact and/or manipulate the virtual object 708.
  • an AR device 728 is described working with a wrist-wearable device 726, an MR headset can be interacted with in the same way as the AR device 728.
  • Figure 7A illustrates an interaction in which an artificially intelligent virtual assistant can assist in requests made by a user 702.
  • the Al virtual assistant can be used to complete open-ended requests made through natural language inputs by a user 702.
  • the user 702 makes an audible request 744 to summarize the conversation and then share the summarized conversation with others in the meeting.
  • the Al virtual assistant is configured to use sensors of the XR system (e.g., cameras of an XR headset, microphones, and various other sensors of any of the devices in the system) to provide contextual prompts to the user for initiating tasks.
  • sensors of the XR system e.g., cameras of an XR headset, microphones, and various other sensors of any of the devices in the system
  • Figure 7A also illustrates an example neural network 752 used in Artificial Intelligence applications.
  • Uses of Artificial Intelligence (Al) are varied and encompass many different aspects of the devices and systems described herein. Al capabilities cover a diverse range of applications and deepen interactions between the user 702 and user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726).
  • the Al discussed herein can be derived using many different training techniques. While the primary Al model example discussed herein is a neural network, other Al models can be used.
  • Nonlimiting examples of Al models include artificial neural networks (ANNs), deep neural networks (DNNs), convolution neural networks (CNNs), recurrent neural networks (RNNs), large language models (LLMs), long short-term memory networks, transformer models, decision trees, random forests, support vector machines, k-nearest neighbors, genetic algorithms, Markov models, Bayesian networks, fuzzy logic systems, and deep reinforcement learnings, etc.
  • the Al models can be implemented at one or more of the user devices, and/or any other devices described herein. For devices and systems herein that employ multiple Al models, different models can be used depending on the task. For example, for a natural-language artificially intelligent virtual assistant, an LLM can be used and for the object detection of a physical environment, a DNN can be used instead.
  • an Al virtual assistant can include many different Al models and based on the user’s request, multiple Al models may be employed (concurrently, sequentially or a combination thereof).
  • an LLM-based Al model can provide instructions for helping a user follow a recipe and the instructions can be based in part on another Al model that is derived from an ANN, a DNN, an RNN, etc. that is capable of discerning what part of the recipe the user is on (e.g., object and scene detection).
  • a user 702 can interact with an Al model through natural language inputs captured by a voice sensor, text inputs, or any other input modality that accepts natural language and/or a corresponding voice sensor module.
  • input is provided by tracking the eye gaze of a user 702 via a gaze tracker module.
  • the Al model can also receive inputs beyond those supplied by a user 702.
  • the Al can generate its response further based on environmental inputs (e.g., temperature data, image data, video data, ambient light data, audio data, GPS location data, inertial measurement (i.e., user motion) data, pattern recognition data, magnetometer data, depth data, pressure data, force data, neuromuscular data, heart rate data, temperature data, sleep data) captured in response to a user request by various types of sensors and/or their corresponding sensor modules.
  • environmental inputs e.g., temperature data, image data, video data, ambient light data, audio data, GPS location data, inertial measurement (i.e., user motion) data, pattern recognition data, magnetometer data, depth data, pressure data, force data, neuromuscular data, heart rate data, temperature data, sleep data
  • environmental inputs e.g., temperature data, image data, video data, ambient light data, audio data, GPS location data, inertial measurement (i.e., user motion) data, pattern recognition data, magnetometer data, depth data, pressure data, force data, neuromuscular data, heart
  • the sensors’ data can be retrieved entirely from a single device (e.g., AR device 728) or from multiple devices that are in communication with each other (e.g., a system that includes at least two of an AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726, etc.).
  • the Al model can also access additional information (e.g., one or more servers 730, the computers 740, the mobile devices 750, and/or other electronic devices) via a network 725.
  • Al-enhanced functions includes but is not limited to image recognition, speech recognition (e.g., automatic speech recognition), text recognition (e.g., scene text recognition), pattern recognition, natural language processing and understanding, classification, regression, clustering, anomaly detection, sequence generation, content generation, and optimization.
  • Al-enhanced functions are fully or partially executed on cloud-computing platforms communicatively coupled to the user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726) via the one or more networks.
  • the cloud-computing platforms provide scalable computing resources, distributed computing, managed Al services, interference acceleration, pre-trained models, APIs and/or other resources to support comprehensive computations required by the Al- enhanced function.
  • Example outputs stemming from the use of an Al model can include natural language responses, mathematical calculations, charts displaying information, audio, images, videos, texts, summaries of meetings, predictive operations based on environmental factors, classifications, pattern recognitions, recommendations, assessments, or other operations.
  • the generated outputs are stored on local memories of the user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726), storage options of the external devices (servers, computers, mobile devices, etc.), and/or storage options of the cloud-computing platforms.
  • the Al-based outputs can be presented across different modalities (e.g., audiobased, visual-based, haptic-based, and any combination thereof) and across different devices of the XR system described herein.
  • Some visual-based outputs can include the displaying of information on XR augments of an XR headset, user interfaces displayed at a wrist-wearable device, laptop device, mobile device, etc.
  • haptic feedback can provide information to the user 702.
  • FIG. 7B shows the user 702 wearing the wrist-wearable device 726 and the AR device 728 and holding the HIPD 742.
  • the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 are used to receive and/or provide one or more messages to a contact of the user 702.
  • the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 detect and coordinate one or more user inputs to initiate a messaging application and prepare a response to a received message via the messaging application.
  • the user 702 initiates, via a user input, an application on the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 that causes the application to initiate on at least one device.
  • an application on the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 that causes the application to initiate on at least one device.
  • the user 702 performs a hand gesture associated with a command for initiating a messaging application (represented by messaging user interface 712); the wrist-wearable device 726 detects the hand gesture; and, based on a determination that the user 702 is wearing the AR device 728, causes the AR device 728 to present a messaging user interface 712 of the messaging application.
  • the AR device 728 can present the messaging user interface 712 to the user 702 via its display (e.g., as shown by user 702’s field of view 710).
  • the application is initiated and can be run on the device (e.g., the wrist-wearable device 726, the AR device 728, and/or the HIPD 742) that detects the user input to initiate the application, and the device provides another device operational data to cause the presentation of the messaging application.
  • the wrist-wearable device 726 can detect the user input to initiate a messaging application, initiate and run the messaging application, and provide operational data to the AR device 728 and/or the HIPD 742 to cause presentation of the messaging application.
  • the application can be initiated and run at a device other than the device that detected the user input.
  • the wrist-wearable device 726 can detect the hand gesture associated with initiating the messaging application and cause the HIPD 742 to run the messaging application and coordinate the presentation of the messaging application.
  • the user 702 can provide a user input provided at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to continue and/or complete an operation initiated at another device.
  • the user 702 can provide an input at the HIPD 742 to prepare a response (e.g., shown by the swipe gesture performed on the HIPD 742).
  • the user 702’s gestures performed on the HIPD 742 can be provided and/or displayed on another device.
  • the user 702’s swipe gestures performed on the HIPD 742 are displayed on a virtual keyboard of the messaging user interface 712 displayed by the AR device 728.
  • the wrist-wearable device 726, the AR device 728, the HIPD 742, and/or other communicatively coupled devices can present one or more notifications to the user 702.
  • the notification can be an indication of a new message, an incoming call, an application update, a status update, etc.
  • the user 702 can select the notification via the wristwearable device 726, the AR device 728, or the HIPD 742 and cause presentation of an application or operation associated with the notification on at least one device.
  • the user 702 can receive a notification that a message was received at the wrist-wearable device 726, the AR device 728, the HIPD 742, and/or other communicatively coupled device and provide a user input at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to review the notification, and the device detecting the user input can cause an application associated with the notification to be initiated and/or presented at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742.
  • the AR device 728 can present to the user 702 game application data and the HIPD 742 can use a controller to provide inputs to the game.
  • the user 702 can use the wrist-wearable device 726 to initiate a camera of the AR device 728, and the user can use the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to manipulate the image capture (e.g., zoom in or out, apply filters) and capture image data.
  • an AR device 728 is shown being capable of certain functions, it is understood that an AR device can be an AR device with varying functionalities based on costs and market demands.
  • an AR device may include a single output modality such as an audio output modality.
  • the AR device may include a low-fidelity display as one of the output modalities, where simple information (e.g., text and/or low-fidelity images/video) is capable of being presented to the user.
  • the AR device can be configured with face-facing light emitting diodes (LEDs) configured to provide a user with information, e.g., an LED around the right-side lens can illuminate to notify the wearerto turn right while directions are being provided or an LED on the left-side can illuminate to notify the wearer to turn left while directions are being provided.
  • the AR device can include an outward-facing projector such that information (e.g., text information, media) may be displayed on the palm of a user’s hand or other suitable surface (e.g., a table, whiteboard).
  • information may also be provided by locally dimming portions of a lens to emphasize portions of the environment in which the user’s attention should be directed.
  • Some AR devices can present AR augments either monocularly or binocularly (e.g., an AR augment can be presented at only a single display associated with a single lens as opposed presenting an AR augmented at both lenses to produce a binocular image).
  • an AR device capable of presenting AR augments binocularly can optionally display AR augments monocularly as well (e.g., for power-saving purposes or other presentation considerations).
  • the user 702 is shown wearing the wristwearable device 726 and an MR device 732 (e.g., a device capable of providing either an entirely VR experience or an MR experience that displays object(s) from a physical environment at a display of the device) and holding the HIPD 742.
  • an MR device 732 e.g., a device capable of providing either an entirely VR experience or an MR experience that displays object(s) from a physical environment at a display of the device
  • the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 are used to interact within an MR environment, such as a VR game or other MR/VR application.
  • the MR device 732 presents a representation of a VR game (e.g., first MR game environment 720) to the user 702
  • the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 detect and coordinate one or more user inputs to allow the user 702 to interact with the VR game.
  • the user 702 can provide a user input via the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 that causes an action in a corresponding MR environment.
  • the user 702 in the third MR system 700c (shown in Figure 7C-1) raises the HIPD 742 to prepare for a swing in the first MR game environment 720.
  • the MR device 732 responsive to the user 702 raising the HIPD 742, causes the MR representation of the user 722 to perform a similar action (e.g., raise a virtual object, such as a virtual sword 724).
  • each device uses respective sensor data and/or image data to detect the user input and provide an accurate representation of the user 702’s motion.
  • image sensors e.g., SLAM cameras or other cameras
  • SLAM cameras or other cameras can be used to detect a position of the HIPD 742 relative to the user 702’s body such that the virtual object can be positioned appropriately within the first MR game environment 720
  • sensor data from the wrist-wearable device 726 can be used to detect a velocity at which the user 702 raises the HIPD 742 such that the MR representation of the user 722 and the virtual sword 724 are synchronized with the user 702’s movements
  • image sensors of the MR device 732 can be used to represent the user 702’s body, boundary conditions, or real-world objects within the first MR game environment 720.
  • the user 702 performs a downward swing while holding the HIPD 742.
  • the user 702’s downward swing is detected by the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 and a corresponding action is performed in the first MR game environment 720.
  • the data captured by each device is used to improve the user’s experience within the MR environment.
  • sensor data of the wrist-wearable device 726 can be used to determine a speed and/or force at which the downward swing is performed and image sensors of the HIPD 742 and/or the MR device 732 can be used to determine a location of the swing and how it should be represented in the first MR game environment 720, which, in turn, can be used as inputs for the MR environment (e.g., game mechanics, which can use detected speed, force, locations, and/or aspects of the user 702’s actions to classify a user’s inputs (e.g., user performs a light strike, hard strike, critical strike, glancing strike, miss) or calculate an output (e.g., amount of damage)).
  • game mechanics which can use detected speed, force, locations, and/or aspects of the user 702’s actions to classify a user’s inputs (e.g., user performs a light strike, hard strike, critical strike, glancing strike, miss) or calculate an output (e.g., amount of damage)
  • Figure 7C-2 further illustrates that a portion of the physical environment is reconstructed and displayed at a display of the MR device 732 while the MR game environment 720 is being displayed.
  • a reconstruction of the physical environment 746 is displayed in place of a portion of the MR game environment 720 when object(s) in the physical environment are potentially in the path of the user (e.g., a collision with the user and an object in the physical environment are likely).
  • this example MR game environment 720 includes (i) an immersive VR portion 748 (e.g., an environment that does not have a corollary counterpart in a nearby physical environment) and (ii) a reconstruction of the physical environment 746 (e.g., table 750 and cup 752).
  • While the example shown here is an MR environment that shows a reconstruction of the physical environment to avoid collisions, other uses of reconstructions of the physical environment can be used, such as defining features of the virtual environment based on the surrounding physical environment (e.g., a virtual column can be placed based on an object in the surrounding physical environment (e.g., a tree)).
  • the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 are described as detecting user inputs, In some examples, user inputs are detected at a single device (with the single device being responsible for distributing signals to the other devices for performing the user input).
  • the HIPD 742 can operate an application for generating the first MR game environment 720 and provide the MR device 732 with corresponding data for causing the presentation of the first MR game environment 720, as well as detect the user 702’s movements (while holding the HIPD 742) to cause the performance of corresponding actions within the first MR game environment 720.
  • operational data e.g., sensor data, image data, application data, device data, and/or other data
  • a single device e.g., the HIPD 742
  • process the operational data and cause respective devices to perform an action associated with processed operational data.
  • the user 702 can wear a wrist-wearable device 726, wear an MR device 732, wear smart textile-based garments 738 (e.g., wearable haptic gloves), and/or hold an HIPD 742 device.
  • the wrist-wearable device 726, the MR device 732, and/or the smart textile-based garments 738 are used to interact within an MR environment (e.g., any AR or MR system described above in reference to Figures 7A-7B).
  • the MR device 732 presents a representation of an MR game (e.g., second MR game environment 720) to the user 702
  • the wrist-wearable device 726, the MR device 732, and/or the smart textile-based garments 738 detect and coordinate one or more user inputs to allow the user 702 to interact with the MR environment.
  • the user 702 can provide a user input via the wrist-wearable device 726, an HIPD 742, the MR device 732, and/or the smart textile-based garments 738 that causes an action in a corresponding MR environment.
  • each device uses respective sensor data and/or image data to detect the user input and provide an accurate representation of the user 702’s motion. While four different input devices are shown (e.g., a wrist-wearable device 726, an MR device 732, an HIPD 742, and a smart textile-based garment 738) each one of these input devices entirely on its own can provide inputs for fully interacting with the MR environment.
  • the wrist-wearable device can provide sufficient inputs on its own for interacting with the MR environment.
  • multiple input devices e.g., a wrist-wearable device and the smart textile-based garment 7308
  • sensor fusion can be utilized to ensure inputs are correct. While multiple input devices are described, it is understood that other input devices can be used in conjunction or on their own instead, such as but not limited to external motion-tracking cameras, other wearable devices fitted to different parts of a user, apparatuses that allow for a user to experience walking in an MR environment while remaining substantially stationary in the physical environment, etc.
  • the data captured by each device is used to improve the user’s experience within the MR environment.
  • the smart textile-based garments 738 can be used in conjunction with an MR device and/or an HIPD 742.
  • experiences are described as occurring on an AR device and other experiences are described as occurring on an MR device, one skilled in the art would appreciate that experiences can be ported over from an MR device to an AR device, and vice versa.
  • Example devices and systems including electronic devices and systems, will be discussed. Such example devices and systems are not intended to be limiting, and one of skill in the art will understand that alternative devices and systems to the example devices and systems described herein may be used to perform the operations and construct the systems and devices that are described herein.
  • an electronic device is a device that uses electrical energy to perform a specific function. It can be any physical object that contains electronic components such as transistors, resistors, capacitors, diodes, and integrated circuits. Examples of electronic devices include smartphones, laptops, digital cameras, televisions, gaming consoles, and music players, as well as the example electronic devices discussed herein.
  • an intermediary electronic device is a device that sits between two other electronic devices, and/or a subset of components of one or more electronic devices and facilitates communication, and/or data processing and/or data transfer between the respective electronic devices and/or electronic components.
  • any data collection performed by the devices described herein and/or any devices configured to perform or cause the performance of the different examples described above in reference to any of the Figures, hereinafter the “devices,” is done with user consent and in a manner that is consistent with all applicable privacy laws. Users are given options to allow the devices to collect data, as well as the option to limit or deny collection of data by the devices. A user is able to opt in or opt out of any data collection at any time. Further, users are given the option to request the removal of any collected data.
  • the term “if’ can be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting,” that a stated condition precedent is true, depending on the context.
  • the phrase “if it is determined [that a stated condition precedent is true]” or “if [a stated condition precedent is true]” or “when [a stated condition precedent is true]” can be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

Systems and methods of generating orchestrated guidance based on an activity of a user are disclosed. An example method for generating orchestrated guidance based on an activity of a user includes in response to an indication received at a wearable device that an artificial intelligence (AI) agent trigger condition is present, providing an AI agent sensor data obtained by the wearable device. The method includes determining, by the AI agent, a context-based activity based on the sensor data obtained by the wearable device and generating, by the AI agent, orchestrated guidance based on the context-based activity. The orchestrated guidance includes a recommended action for performing the context-based activity. The method also includes presenting the orchestrated guidance at the wearable device.

Description

WEARABLE DEVICES INCLUDING ARTIFICIALLY INTELLIGENT SYSTEMS FOR GENERATING AND PRESENTING GUIDANCE TO WEARERS CROSS-REFERENCE TO RELATED APPLICATIONS
[0001] This application claims benefit of and priority to U.S. provisional patent application Ser. No. 63/649,289 filed May 17, 2024 and to U.S. non-provisional patent application Ser. No. 19/209,723 filed May 15, 2025.
TECHNICAL FIELD
[0002] This relates generally to approaches for interacting with an artificially intelligent agent and, more specifically, utilizing artificially intelligent agent included at wearable devices to augment user experiences.
BACKGROUND
[0003] While artificial intelligence is used in different manners, commercial Al is usually only accessible in inconvenient manners, such as interacting with an artificial intelligence on a website or receiving Al generated content in relation to an internet search. These examples have drawbacks as it limits the user’s experience with Al generated content to very siloed experiences and also has a high burden on the user for accessing/interacting with the Al.
[0004] As such, there is a need to address one or more of the above-identified challenges. A brief summary of solutions to the issues noted above are described below.
SUMMARY
[0005] According to a first aspect, there is provided a method, comprising: in response to an indication that an artificial intelligence (Al) agent trigger condition is present, providing an Al agent sensor data obtained by a wearable device; determining, by the Al agent, a context-based activity based on the sensor data obtained by the wearable device; generating, by the Al agent, orchestrated guidance based on the context-based activity, wherein the orchestrated guidance includes a recommended action for performing the context-based activity; and presenting the orchestrated guidance at the wearable device.
[0006] In this way, a user may be assisted in performing a technical task by means of a continued and/or guided human-machine interaction process.
[0007] The context-based activity may be a first context-based activity. The sensor data may be first sensor data. The orchestrated guidance may be first orchestrated guidance. The recommended action may be a first recommended action. The method may further comprise: in accordance with a determination that the first recommended action for performing the first context-based activity was performed, providing the Al agent second sensor data obtained by the wearable device; determining, by the Al agent, a second context-based activity based on the second sensor data obtained by the wearable device; generating, by the Al agent, second orchestrated guidance based on the second context-based activity, wherein the second orchestrated guidance includes a second recommended action for performing the second context-based activity; and presenting the second orchestrated guidance at the wearable device.
[0008] The context-based activity may be a first context-based activity of a plurality of context-based activities determined by the by the Al agent based on the sensor data. The orchestrated guidance may include a plurality of recommended actions for performing the plurality of context-based activities. The recommended action may be a first recommended action of the plurality of recommended actions. The first recommended action may be configured to perform the first context-based activity. Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action of the plurality of recommended actions.
[0009] Generating the orchestrated guidance may include determining a subset of the plurality of recommended actions for performing the first context-based activity. Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action of the plurality of recommended actions and the subset of the plurality of recommended actions for performing the first context-based activity.
[0010] Generating the orchestrated guidance may include determining a sequence of context-based activities of the plurality of context-based activities to be performed, including a second context-based activity to follow the first context-based activity. Presenting the orchestrated guidance at the wearable device may include presenting at least the first recommended action and the second recommended action of the plurality of recommended actions for performing the plurality of context-based activities.
[0011] The method may further comprise: in response to a user input selecting the recommended action for performing the context-based activity, causing the wearable device to initiate a do-not-disturb mode, wherein, while in the do-not-disturb mode, the wearable device suppresses, at least, received notifications; and in response to an indication that participation in the context-based activity ceased: causing the wearable device to cease the do-not-disturb mode, generating, by the Al agent, a notification summary based on the notifications received while the wearable device was in the do-not-disturb mode, and presenting the notification summary at the wearable device.
[0012] In response to a user input selecting the recommended action for performing the context-based activity, the method may further comprise performing, by the Al agent, a search based on the recommended action; determining a task to perform based on the search; and presenting the task at the wearable device.
[0013] Presenting the orchestrated guidance at the wearable device may include at least one of: causing presentation of a user interface element associated with the orchestrated guidance at a communicatively coupled display, and/or causing presentation of audible guidance associated with the orchestrated guidance at a communicatively coupled speaker. [0014] The context-based activity may be performed at a physical activity.
[0015] According to a second aspect, there is provided a wearable device, comprising: a display; one or more sensors; and one or more programs, wherein the one or more programs are stored in memory and configured to be executed by one or more processors to cause the one or more processors to carry out the method of the first aspect.
[0016] According to a third aspect, there is provided a computer-readable storage medium including executable instructions that, when executed by one or more processors, cause the one or more processors to carry out the method of the first aspect. The medium may be non-transitory.
[0017] According to a fourth aspect, there is provided a computer program product comprising instructions that, when executed by one or more processors of a computer, cause the one or more processors to carry out the method of the first aspect.
[0018] Also described is, a wearable device for facilitating performance of a physical activity performed by user is described herein. The example wearable device can be a headwearable device including a display, one or more sensors, and one or more programs. The one or more programs are stored in memory and configured to be executed by one or more processors, the one or more programs including instructions for, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining data associated with an on-going activity performed by the user of the head-wearable device. The one or more programs include instructions for generating, by an artificial intelligence (Al) agent, a contextbased response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device. The one or more programs include instructions for presenting, at the head-wearable device, context-based response. The context-based response is presented within a portion of a field of view of the user.
[0019] Also described is, a method for facilitating performance of a physical activity performed by user is described herein. The method includes, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining data associated with an on-going activity performed by the user of the head-wearable device. The method also includes generating, by an artificial intelligence (Al) agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the headwearable device. The method further includes presenting, at the head-wearable device, context-based response, wherein the context-based response is presented within a portion of a field of view of the user.
[0020] Also described is, a non-transitory, computer-readable storage medium including executable instructions that, when executed by one or more processors of a wearable device (e.g., a head-wearable device), cause the one or more processors to facilitate performance of a physical activity performed by user is described herein. The executable instructions, when executed by one or more processors, cause the one or more processors to, in response to an indication that a user of a head-wearable device is participating in an activity, obtain data associated with an on-going activity performed by the user of the headwearable device. The executable instructions, when executed by one or more processors, cause the one or more processors to generate, by an artificial intelligence (Al) agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device. The executable instructions, when executed by one or more processors, cause the one or more processors to present, at the head-wearable device, context-based response, wherein the context-based response is presented within a portion of a field of view of the user.
[0021] Instructions that cause performance of the methods and operations described herein can be stored on a non-transitory computer readable storage medium. The non- transitory computer-readable storage medium can be included on a single electronic device or spread across multiple electronic devices of a system (computing system). A non- exhaustive of list of electronic devices that can either alone or in combination (e.g., a system) perform the method and operations described herein include an extended-reality (XR) headset/glasses (e.g., a mixed-reality (MR) headset or a pair of augmented-reality (AR) glasses as two examples), a wrist-wearable device, an intermediary processing device, a smart textile-based garment, etc. For instance, the instructions can be stored on a pair of AR glasses or can be stored on a combination of a pair of AR glasses and an associated input device (e.g., a wrist-wearable device) such that instructions for causing detection of input operations can be performed at the input device and instructions for causing changes to a displayed user interface in response to those input operations can be performed at the pair of AR glasses. The devices and systems described herein can be configured to be used in conjunction with methods and operations for providing an XR experience. The methods and operations for providing an XR experience can be stored on a non-transitory computer- readable storage medium.
[0022] The devices and/or systems described herein can be configured to include instructions that cause the performance of methods and operations associated with the presentation and/or interaction with an extended-reality (XR) headset. These methods and operations can be stored on a non-transitory computer-readable storage medium of a device or a system. It is also noted that the devices and systems described herein can be part of a larger, overarching system that includes multiple devices. A non-exhaustive of list of electronic devices that can, either alone or in combination (e.g., a system), include instructions that cause the performance of methods and operations associated with the presentation and/or interaction with an XR experience include an extended-reality headset (e.g., a mixed-reality (MR) headset or a pair of augmented-reality (AR) glasses as two examples), a wrist-wearable device, an intermediary processing device, a smart textile-based garment, etc. For example, when an XR headset is described, it is understood that the XR headset can be in communication with one or more other devices (e.g., a wrist-wearable device, a server, intermediary processing device) which together can include instructions for performing methods and operations associated with the presentation and/or interaction with an extended- reality system (i.e., the XR headset would be part of a system that includes one or more additional devices). Multiple combinations with different related devices are envisioned, but not recited for brevity.
[0023] The features and advantages described in the specification are not necessarily all inclusive and, in particular, certain additional features and advantages will be apparent to one of ordinary skill in the art in view of the drawings, specification, and claims. Moreover, it should be noted that the language used in the specification has been principally selected for readability and instructional purposes.
[0024] Having summarized the above example aspects, a brief description of the drawings will now be presented.
BRIEF DESCRIPTION OF THE DRAWINGS
[0025] For a better understanding of the present disclosure, reference should be made to the Detailed Description below, in conjunction with the following drawings in which like reference numerals refer to corresponding parts throughout the figures.
[0026] Figures 1A-1 N illustrate invocation of an artificially intelligent agent at one or more wearable devices for providing guidance based on an activity of a user.
[0027] Figures 2A-2L illustrate context-based responses generated by an artificially intelligent agent based on activities performed by a user.
[0028] Figures 3A-3D illustrate example user interfaces and additional features available at an Al assistive system.
[0029] Figures 4A and 4B illustrate example sequences of user interactions with personalized assistive systems.
[0030] Figure 5 illustrates a flow chart of a method for generating orchestrated guidance based on an activity of a user.
[0031] Figure 6 illustrates a flow chart of a method for facilitating performance of a physical activity performed by user.
[0032] Figures 7A - 7C-2 illustrate example XR systems that include AR and MR systems.
[0033] In accordance with common practice, the various features illustrated in the drawings may not be drawn to scale. Accordingly, the dimensions of the various features may be arbitrarily expanded or reduced for clarity. In addition, some of the drawings may not depict all of the components of a given system, method, or device. Finally, like reference numerals may be used to denote like features throughout the specification and figures.
DETAILED DESCRIPTION
[0034] Numerous details are described herein to provide a thorough understanding of the examples illustrated in the accompanying drawings. However, some examples may be practiced without many of the specific details, and the scope of the claims is only limited by those features and aspects specifically recited in the claims. Furthermore, well-known processes, components, and materials have not necessarily been described in exhaustive detail so as to avoid obscuring pertinent aspects of the disclosure.
[0035] Embodiments of this disclosure can include or be implemented in conjunction with various types of extended-realities (XRs) such as mixed-reality (MR) and augmented- reality (AR) systems. MRs and ARs, as described herein, are any superimposed functionality and/or sensory-detectable presentation provided by MR and AR systems within a user’s physical surroundings. Such MRs can include and/or represent virtual realities (VRs) and VRs in which at least some aspects of the surrounding environment are reconstructed within the virtual environment (e g., displaying virtual reconstructions of physical objects in a physical environment to avoid the user colliding with the physical objects in a surrounding physical environment). In the case of MRs, the surrounding environment that is presented through a display is captured via one or more sensors configured to capture the surrounding environment (e.g., a camera sensor, time-of-flight (ToF) sensor). While a wearer of an MR headset can see the surrounding environment in full detail, they are seeing a reconstruction of the environment reproduced using data from the one or more sensors (i.e. , the physical objects are not directly viewed by the user). An MR headset can also forgo displaying reconstructions of objects in the physical environment, thereby providing a user with an entirely VR experience. An AR system, on the other hand, provides an experience in which information is provided, e.g., through the use of a waveguide, in conjunction with the direct viewing of at least some of the surrounding environment through a transparent or semitransparent waveguide(s) and/or lens(es) of the AR glasses. Throughout this application, the term “extended reality (XR)” is used as a catchall term to cover both ARs and MRs. In addition, this application also uses, at times, a head-wearable device or headset device as a catchall term that covers XR headsets such as AR glasses and MR headsets.
[0036] As alluded to above, an MR environment, as described herein, can include, but is not limited to, non-immersive, semi-immersive, and fully immersive VR environments. As also alluded to above, AR environments can include marker-based AR environments, markerless AR environments, location-based AR environments, and projection-based AR environments. The above descriptions are not exhaustive and any other environment that allows for intentional environmental lighting to pass through to the user would fall within the scope of an AR, and any other environment that does not allow for intentional environmental lighting to pass through to the user would fall within the scope of an MR.
[0037] The AR and MR content can include video, audio, haptic events, sensory events, or some combination thereof, any of which can be presented in a single channel or in multiple channels (such as stereo video that produces a three-dimensional effect to a viewer). Additionally, AR and MR can also be associated with applications, products, accessories, services, or some combination thereof, which are used, for example, to create content in an AR or MR environment and/or are otherwise used in (e.g., to perform activities in) AR and MR environments.
[0038] Interacting with these AR and MR environments described herein can occur using multiple different modalities and the resulting outputs can also occur across multiple different modalities. In one example AR or MR system, a user can perform a swiping in-air hand gesture to cause a song to be skipped by a song-providing application programming interface (API) providing playback at, for example, a home speaker.
[0039] A hand gesture, as described herein, can include an in-air gesture, a surfacecontact gesture, and or other gestures that can be detected and determined based on movements of a single hand (e.g., a one-handed gesture performed with a user’s hand that is detected by one or more sensors of a wearable device (e.g., electromyography (EMG) and/or inertial measurement units (IMUs) of a wrist-wearable device, and/or one or more sensors included in a smart textile wearable device) and/or detected via image data captured by an imaging device of a wearable device (e.g., a camera of a head-wearable device, an external tracking camera setup in the surrounding environment)). “In-air” generally includes gestures in which the user’s hand does not contact a surface, object, or portion of an electronic device (e.g., a head-wearable device or other communicatively coupled device, such as the wristwearable device), in other words the gesture is performed in open air in 3D space and without contacting a surface, an object, or an electronic device. Surface-contact gestures (contacts at a surface, object, body part of the user, or electronic device) more generally are also contemplated in which a contact (or an intention to contact) is detected at a surface (e.g., a single- or double-finger tap on a table, on a user’s hand or another finger, on the user’s leg, a couch, a steering wheel). The different hand gestures disclosed herein can be detected using image data and/or sensor data (e.g., neuromuscular signals sensed by one or more biopotential sensors (e.g., EMG sensors) or other types of data from other sensors, such as proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors) detected by a wearable device worn by the user and/or other electronic devices in the user’s possession (e.g., smartphones, laptops, imaging devices, intermediary devices, and/or other devices described herein). [0040] The input modalities as alluded to above can be varied and are dependent on a user’s experience. For example, in an interaction in which a wrist-wearable device is used, a user can provide inputs using in-air or surface-contact gestures that are detected using neuromuscular signal sensors of the wrist-wearable device. In the event that a wrist-wearable device is not used, alternative and entirely interchangeable input modalities can be used instead, such as camera(s) located on the headset/glasses or elsewhere to detect in-air or surface-contact gestures or inputs at an intermediary processing device (e.g., through physical input components (e.g., buttons and trackpads)). These different input modalities can be interchanged based on both desired user experiences, portability, and/or a feature set of the product (e.g., a low-cost product may not include hand-tracking cameras).
[0041] While the inputs are varied, the resulting outputs stemming from the inputs are also varied. For example, an in-air gesture input detected by a camera of a head-wearable device can cause an output to occur at a head-wearable device or control another electronic device different from the head-wearable device. In another example, an input detected using data from a neuromuscular signal sensor can also cause an output to occur at a headwearable device or control another electronic device different from the head-wearable device. While only a couple examples are described above, one skilled in the art would understand that different input modalities are interchangeable along with different output modalities in response to the inputs.
[0042] Specific operations described above may occur as a result of specific hardware. The devices described are not limiting and features on these devices can be removed or additional features can be added to these devices. The different devices can include one or more analogous hardware components. For brevity, analogous devices and components are described herein. Any differences in the devices and components are described below in their respective sections.
[0043] As described herein, a processor (e.g., a central processing unit (CPU) or microcontroller unit (MCU)), is an electronic component that is responsible for executing instructions and controlling the operation of an electronic device (e.g., a wrist-wearable device, a head-wearable device, a handheld intermediary processing device (HIPD), a smart textilebased garment, or other computer system). There are various types of processors that may be used interchangeably or specifically required by examples described herein. For example, a processor may be (i) a general processor designed to perform a wide range of tasks, such as running software applications, managing operating systems, and performing arithmetic and logical operations; (ii) a microcontroller designed for specific tasks such as controlling electronic devices, sensors, and motors; (iii) a graphics processing unit (GPU) designed to accelerate the creation and rendering of images, videos, and animations (e.g., VR animations, such as three-dimensional modeling); (iv) a field-programmable gate array (FPGA) that can be programmed and reconfigured after manufacturing and/or customized to perform specific tasks, such as signal processing, cryptography, and machine learning; or (v) a digital signal processor (DSP) designed to perform mathematical operations on signals such as audio, video, and radio waves. One of skill in the art will understand that one or more processors of one or more electronic devices may be used in various examples described herein.
[0044] As described herein, controllers are electronic components that manage and coordinate the operation of other components within an electronic device (e.g., controlling inputs, processing data, and/or generating outputs). Examples of controllers can include (i) microcontrollers, including small, low-power controllers that are commonly used in embedded systems and Internet of Things (loT) devices; (ii) programmable logic controllers (PLCs) that may be configured to be used in industrial automation systems to control and monitor manufacturing processes; (iii) system-on-a-chip (SoC) controllers that integrate multiple components such as processors, memory, I/O interfaces, and other peripherals into a single chip; and/or (iv) DSPs. As described herein, a graphics module is a component or software module that is designed to handle graphical operations and/or processes and can include a hardware module and/or a software module.
[0045] As described herein, memory refers to electronic components in a computer or electronic device that store data and instructions for the processor to access and manipulate. The devices described herein can include volatile and non-volatile memory. Examples of memory can include (i) random access memory (RAM), such as DRAM, SRAM, DDR RAM or other random access solid state memory devices, configured to store data and instructions temporarily; (ii) read-only memory (ROM) configured to store data and instructions permanently (e.g., one or more portions of system firmware and/or boot loaders); (iii) flash memory, magnetic disk storage devices, optical disk storage devices, other non-volatile solid state storage devices, which can be configured to store data in electronic devices (e.g., universal serial bus (USB) drives, memory cards, and/or solid-state drives (SSDs)); and (iv) cache memory configured to temporarily store frequently accessed data and instructions. Memory, as described herein, can include structured data (e.g., SQL databases, MongoDB databases, GraphQL data, or JSON data). Other examples of memory can include (i) profile data, including user account data, user settings, and/or other user data stored by the user; (ii) sensor data detected and/or otherwise obtained by one or more sensors; (iii) media content data including stored image data, audio data, documents, and the like; (iv) application data, which can include data collected and/or otherwise obtained and stored during use of an application; and/or (v) any other types of data described herein.
[0046] As described herein, a power system of an electronic device is configured to convert incoming electrical power into a form that can be used to operate the device. A power system can include various components, including (i) a power source, which can be an alternating current (AC) adapter or a direct current (DC) adapter power supply; (ii) a charger input that can be configured to use a wired and/or wireless connection (which may be part of a peripheral interface, such as a USB, micro-USB interface, near-field magnetic coupling, magnetic inductive and magnetic resonance charging, and/or radio frequency (RF) charging); (iii) a power-management integrated circuit, configured to distribute power to various components of the device and ensure that the device operates within safe limits (e.g., regulating voltage, controlling current flow, and/or managing heat dissipation); and/or (iv) a battery configured to store power to provide usable power to components of one or more electronic devices.
[0047] As described herein, peripheral interfaces are electronic components (e.g., of electronic devices) that allow electronic devices to communicate with other devices or peripherals and can provide a means for input and output of data and signals. Examples of peripheral interfaces can include (i) USB and/or micro-USB interfaces configured for connecting devices to an electronic device; (ii) Bluetooth interfaces configured to allow devices to communicate with each other, including Bluetooth low energy (BLE); (iii) near-field communication (NFC) interfaces configured to be short-range wireless interfaces for operations such as access control; (iv) pogo pins, which may be small, spring-loaded pins configured to provide a charging interface; (v) wireless charging interfaces; (vi) global- positioning system (GPS) interfaces; (vii) Wi-Fi interfaces for providing a connection between a device and a wireless network; and (viii) sensor interfaces.
[0048] As described herein, sensors are electronic components (e.g., in and/or otherwise in electronic communication with electronic devices, such as wearable devices) configured to detect physical and environmental changes and generate electrical signals. Examples of sensors can include (i) imaging sensors for collecting imaging data (e.g., including one or more cameras disposed on a respective electronic device, such as a simultaneous localization and mapping (SLAM) camera); (ii) biopotential-signal sensors; (iii) IMUs for detecting, for example, angular rate, force, magnetic field, and/or changes in acceleration; (iv) heart rate sensors for measuring a user’s heart rate; (v) peripheral oxygen saturation (SpO2) sensors for measuring blood oxygen saturation and/or other biometric data of a user; (vi) capacitive sensors for detecting changes in potential at a portion of a user’s body (e.g., a sensor-skin interface) and/or the proximity of other devices or objects; (vii) sensors for detecting some inputs (e.g., capacitive and force sensors); and (viii) light sensors (e.g., ToF sensors, infrared light sensors, or visible light sensors), and/or sensors for sensing data from the user or the user’s environment. As described herein biopotential-signal-sensing components are devices used to measure electrical activity within the body (e.g., biopotentialsignal sensors). Some types of biopotential-signal sensors include (i) electroencephalography (EEG) sensors configured to measure electrical activity in the brain to diagnose neurological disorders; (ii) electrocardiography (ECG or EKG) sensors configured to measure electrical activity of the heart to diagnose heart problems; (iii) EMG sensors configured to measure the electrical activity of muscles and diagnose neuromuscular disorders; (iv) electrooculography (EOG) sensors configured to measure the electrical activity of eye muscles to detect eye movement and diagnose eye disorders.
[0049] As described herein, an application stored in memory of an electronic device (e.g., software) includes instructions stored in the memory. Examples of such applications include (i) games; (ii) word processors; (iii) messaging applications; (iv) media-streaming applications; (v) financial applications; (vi) calendars; (vii) clocks; (viii) web browsers; (ix) social media applications; (x) camera applications; (xi) web-based applications; (xii) health applications; (xiii) AR and MR applications; and/or (xiv) any other applications that can be stored in memory. The applications can operate in conjunction with data and/or one or more components of a device or communicatively coupled devices to perform one or more operations and/or functions.
[0050] As described herein, communication interface modules can include hardware and/or software capable of data communications using any of a variety of custom or standard wireless protocols (e.g, IEEE 802.15.4, Wi-Fi, ZigBee, 6L0WPAN, Thread, Z-Wave, Bluetooth Smart, ISA100.11a, WirelessHART, or MiWi), custom or standard wired protocols (e.g., Ethernet or HomePlug), and/or any other suitable communication protocol, including communication protocols not yet developed as of the filing date of this document. A communication interface is a mechanism that enables different systems or devices to exchange information and data with each other, including hardware, software, or a combination of both hardware and software. For example, a communication interface can refer to a physical connector and/or port on a device that enables communication with other devices (e.g., USB, Ethernet, HDMI, or Bluetooth). A communication interface can refer to a software layer that enables different software programs to communicate with each other (e.g., APIs and protocols such as HTTP and TCP/IP).
[0051] As described herein, a graphics module is a component or software module that is designed to handle graphical operations and/or processes and can include a hardware module and/or a software module.
[0052] As described herein, non-transitory computer-readable storage media are physical devices or storage medium that can be used to store electronic data in a non- transitory form (e.g., such that the data is stored permanently until it is intentionally deleted and/or modified).
[0053] The systems and methods disclosed herein provide different ways in which wearable devices can utilize artificial intelligence (Al) and/or an Al Agent (also referred to as a Al digital assistant or Al assistant). For example, a head-wearable device can retrieve information and use the information with an Al agent to generate responses and/or recommendations that are displayed at the head-wearable device and/or another communicatively device. The systems and method disclosed here can be used collaborate with other users (including wearers of other wearable devices), and interact with third party applications using built-in Al models. The systems and methods disclosed herein can utilize a user interactable Al agent to perform various tasks at the user’s request, as well as utilize the Al agent to monitor situations and provide user-specific assistance.
[0054] The systems and methods disclosed herein utilize Al agent to work with wearable devices and other devices (e.g., laptop, tablet, watches, desktops, phones, and other internet connected devices) within an ecosystem to accomplish tasks across multiple devices (e.g., XR systems described below in reference to Figures 7A-7C-2). For example, an Al agent can be configured to control an aspect of one or more of the other devices based on a request from the user. In some examples, the Al agent can also be invoked on different devices based on a determination that the user is interacting with a device other than a wearable device.
[0055] In some examples, the systems and methods disclosed herein can use an Al agent to augment a user experience. In particular, the Al agent can receive sensor data and/or other information captured by a wearable device, and use the sensor data and/or other information to generate and provide recommended actions and/or context-based responses. For example, a head-wearable device worn by the user can capture information corresponding to a field of view of the user 105 and/or a location of the user to generate and provide recommended actions and/or context-based responses. The systems and methods disclosed herein generate and provide tailored information to a user based on location and/or data received from one or more wearable devices (e.g., sensor data and/or image data of a wristwearable device, a head-wearable device, etc.).
[0056] The systems and methods disclosed herein utilize an Al agent to collate recorded information (e.g., camera photos and videos) across multiple wearable devices to produce unique media (e.g., a single video which stitches the multiple head-wearable devices video feed into a single viewing experience). In some examples, positional data of each communicatively coupled device (e.g., wearable device, such as a head-wearable device) can be used to determine how the media is presented.
[0057] The systems and methods disclosed herein utilize an Al agent to work with third-party applications through the use of an API. In other words, the user can use an Al agent implemented at a wearable device to perform a task of applications by utilizing the API to communicate with the Al agent. In some examples, the Al agent can be configured to interact with applications and graphical user interfaces (GUIs) without the use of an API. [0058] Figures 1A-1 N illustrate invocation of an artificially intelligent agent at one or more wearable devices for providing guidance based on an activity of a user. An Al guidance system 100 shown and described in reference to Figures 1A-1 N provides example orchestrated guidance provided to a user 105 visiting a museum. The Al guidance system 100 includes at least a wrist-wearable device 1 10 and a head-wearable device 120 donned by the user 105. The Al guidance system 100 can include other wearable devices worn by the user 105, such as smart textile-based garments (e.g., wearable bands, shirts, etc.), and/or other electronic devices, such as an HIPD 742, a computer 740 (e.g., a laptop), mobile devices 750 (e.g., smartphones, tablets), and/or other electronic devices described below in reference to Figures 7A-7C. The Al guidance system 100, the wearable devices, and the electronic devices can be communicatively coupled via a network (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN). The Al guidance system 100 further includes an Al agent 115 (represented by star symbols) that can be invoked by the user 105 via one or more devices of the Al guidance system 100 (e.g., a wearable device, such as a wrist-wearable device 110 and/or a head-wearable device 120). Alternatively or in addition, In some examples, the Al agent 1 15 can be invoked in accordance with a determination that an Al agent trigger condition is present (as discussed below).
[0059] As described below in reference to Figure 7A, the wrist-wearable device 1 10 (analogous to wrist-wearable device 726; Figures 7A-7C-2) can include a display 1 12, an imaging device 1 14 (e.g., a camera), a microphone, a speaker, input surfaces (e.g., touch input surfaces, mechanical inputs, etc.), and one or more sensors (e.g., biopotential sensors (e.g., EMG sensors), proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors, etc.). Similarly, the head-wearable device 120 (analogous to AR device 728 and MR device 732; Figures 7A-7C-2) can include another imaging device 122, an additional microphone, an additional speaker, additional input surfaces (e.g., touch input surfaces, mechanical inputs, etc.), and one or more additional sensors (e.g., biopotential sensors (e.g., EMG sensors), gaze trackers, proximity sensors, ToF sensors, sensors of an IMU, capacitive sensors, strain sensors, etc.). In some examples, the head-wearable device 120 includes a display.
[0060] Turning to Figure 1A, the wrist-wearable device 110 provides first example orchestrated guidance. While the user 105 is at the museum, the wrist-wearable device 1 10 and the head-wearable device 120 capture at least sensor data and image data via one or more sensors and/or imaging devices (e.g., imaging devices 1 14 and 122). In some examples, the head-wearable device 120 captures audio data. The Al guidance system 100 can determine, based on image data, sensor data, audio data, and/or any other data available to the Al guidance system 100, whether an Al agent trigger condition is satisfied and, in accordance with a determination that an Al agent trigger condition is satisfied, the Al guidance system 100 can provide the indication that an Al agent trigger condition is present. In response to an indication that an Al agent trigger condition is present, the Al guidance system 100 provides the Al agent 1 15, at least, image data, sensor data, audio data, and/or any other data captured by the devices of the Al guidance system 100. Alternatively or in addition, In some examples, the Al guidance system 100 provides the Al agent 1 15, at least, image data, sensor data, audio data, and/or any other data captured by the devices of the Al guidance system 100 in response to user invocation of the Al agent 1 15. The Al agent 1 15 can be invoked via touch inputs, voice commands, hand gestures detected by and/or received at the wristwearable device 1 10, the head-wearable device 120, and/or any other device of the Al guidance system 100.
[0061] The Al agent 115 can use, at least, the image data and/or the sensor data received from the Al guidance system 100 to determine a context-based activity. For example, the Al agent 1 15 can use the image data and/or the sensor data to determine that the user 115 is visiting or exploring the museum. In some examples, the Al agent 1 15 can also use audio data to determine a context-based activity. The context-based activity can be a physical activity (e.g. running, walking, etc.) and/or participation in an event (e.g., sightseeing, performing a hobby, cooking, driving, participating in a meeting, etc.). The Al agent 115 can further generate orchestrated guidance based on the context-based activity. The orchestrated guidance includes a recommended action for performing the context-based activity. The Al guidance system 100 can present the orchestrated guidance at a wearable device (e.g., the wrist-wearable device 1 10 and/or the head-wearable device 120) and/or any other communicatively coupled device.
[0062] For example, in Figure 1A, the Al agent 1 15 provides orchestrated guidance for the user 105’s museum visit, the orchestrated guidance including one or more recommended actions for facilitating the museum visit. The orchestrated guidance and the recommended actions are presented at a display 1 12 of the wrist-wearable device 1 10. In Figure 1 A, the wrist-wearable device 1 10 presents, via the display 112, the first orchestrated guidance 116 (e.g., “Welcome to the museum! Here are some things you can do!”) and the recommended actions (e.g. take tour user interface (Ul) element 1 18 and do-not-disturb Ul element 1 19) generated by the Al agent 1 15. In this way, the Al guidance system 100 can tailor the guided tour for the user 105.
[0063] Figure 1 B shows a field of view 125 of the user 105 via the head-wearable device 120. As shown in Figure 1 B, the orchestrated guidance generated by the Al agent 1 15 can also be presented via a display of the head-wearable device 120. For example, the field of view 125 of the user 105 includes a first orchestrated guidance Ul element 127 (e.g. “Welcome to the museum! Let’s take a look around”). While Figures 1A and 1 B show orchestrated guidance and recommended actions presented at displays of the wrist-wearable device 110 and/or the head-wearable device 120, In some examples, the orchestrated guidance and recommended actions can be presented via a speaker of wrist-wearable device 110, the head-wearable device 120, and/or another communicatively coupled device.
[0064] Figure 1 C show the user 105 providing a first user input 129 selecting a recommended action of the first orchestrated guidance 1 16. In particular, the user 105 performs a hand gesture (e.g. a pinch) to provide a first user input 129 selecting the do-not- disturb Ul element 1 19. In some examples, the first user input 129 selecting the do-not-disturb Ul element 119 causes the wrist-wearable device 110, the head-wearable device 120, and/or other devices of the Al guidance system 100 to initiate a do-not-disturb mode (or focus mode, away mode, etc.). While in the do-not-disturb mode, the Al guidance system 100 suppresses, at least, received notifications, calls, and/or messages. In some examples, the use 105 can provide a voice request and/or other input to the Al guidance system 100 to silence notifications and provide a summary of the notifications later.
[0065] Figure 1 D shows a confirmation message generated by the Al agent 1 15. The Al agent, in response to the first user input 129, generates a corresponding response or recommended action. For example, the field of view 125 of the user 105 includes a confirmation message Ul element 130 based on an accepted recommended action of the first orchestrated guidance 116.
[0066] Figure 1 E shows updates to the first orchestrated guidance 116 based on one or more user inputs. The orchestrated guidance generated by the Al agent 115 can include a subset of a plurality of recommended actions for performing the context-based activity. The orchestrated guidance, when presented at a wearable device, can include at least the subset of the plurality of recommended actions for performing the context-based activity. In some examples, one or more recommended actions of an orchestrated guidance are updated based on a user input selecting the one or more recommended actions. For example, the first orchestrated guidance 116 includes at least two Ul elements - take tour Ul element 118 and do-not-disturb Ul element 1 19 - and the Al agent 1 15 updates the first orchestrated guidance 1 16 to replace the do-not-disturb Ul element 119 with a view map Ul element 131 after detecting the first user input 129 selecting the do-not-disturb Ul element 1 19. Similarly, the second user input 133 selecting the take tour Ul element 1 18 cause the Al agent 115 to present updated first orchestrated guidance 116 and/or updated recommended actions. Alternatively, or in addition, In some examples, one or more recommended actions of an orchestrated guidance are updated based on the user 105 forgoing to select or ignoring one or more recommended actions.
[0067] In some examples, the Al agent 115 can determine that a context-based activity is one of a plurality of context based activities and, when generating the orchestrated guidance, determine a sequence for performing the plurality of context based activities (or context-based activities to be performed together and/or on parallel). For example, the context-based activity can be a first context-based activity of a plurality of context-based activities determined by the by the Al agent 1 15 (based on the sensor data, audio data, and/or image data), the orchestrated guidance can include a plurality of recommended actions for performing the plurality of context-based activities, and the recommended action is a first recommended action of the plurality of recommended actions, the first recommended action being configured to perform the first context-based activity.
[0068] In some examples, the Al agent 1 15 can determine when one or more contextbased activities are completed, identify similar context-based activities, provide alternate context-based activities (if one or more specific context-based activities cannot be performed or alternate suggestion are present). For example, the user 105 can have a schedule including at least two events - the museum visit (e.g., a first context-based activity) and a dinner (e.g., a second context-based activity) - and the orchestrated guidance determined by the Al agent 115 can include a first set of recommended actions for augmenting the user 105’s museum visit and a second set of recommended actions for augmenting the user 105’s dinner, the second set of recommended actions being presented to the user 105 in accordance with a determination that the museum visit has concluded (e.g, the user 105 leaves the museum, the user 105 terminates an augmented experience for the museum visit provided by the Al agent, the scheduled museum visit time elapses, etc.).
[0069] Figure 1 F shows a context-based response generated by the Al agent 115. The context-based response is generated in response to the second user input 133 selecting the take tour Ul element 1 18. In particular, the Al agent 1 15 generates a context-based response to facilitate the museum tour. For example, the user 105 can view a piece of art and the Al agent 1 15 can recognize the art and provide contextual information (or the contextbased response) to the user 105 (e.g., by presenting the information at a wrist-wearable device). The Al agent 1 15 can use the sensor data, audio data, and/or the image data to generate the context-based response. For example, in Figure 1 F, the Al agent 1 15 uses the sensor data, audio data, and/or the image data to identify that the statute 134 is an object of interest to the user 105 and generates a context-based response based on the statute 134. Identification of an object of interest is discussed below in reference to Figure 1 G.
[0070] The context-based response can be presented at the wrist-wearable device 110, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100. For example, the Al agent 1 15 presents a first context-based response Ul element 135 via a display of the head-wearable device 120, as shown in field of view 125 of the user 105.
[0071] Figure 1 G shows identification of an object of interest. In some examples, the Al guidance system 100 can identify an object of interest based on user gaze (determined by one or more eye tackers, sensors, and/or imaging devices of the head-wearable device 120 (e.g., gaze of user focused on an object for a predetermined amount of time (e.g., 10 seconds, 30 seconds, etc.))), direction of a field of view of the user 105 (determined by one or more sensors and/or imaging devices of the head-wearable device 120), pointing gestures performed by the user 105 (determined by one or more sensors and/or imaging devices of the wrist-wearable device 110 and/or the head-wearable device 120), voice commands, and/or other inputs provided by the user 105 to select an object of interest. For example, in Figures 1G, the user 105 provides a voice command 137 describing an object of interest. Alternatively, or in addition, the user 105 can perform a pointing gesture 138 to identify the object of interest and/or to augment or supplement the voice command 137. In other words, the Al guidance system 100 can use one or more inputs modalities to identify an object of interest. In this way, the Al guidance system 100 can provide the user 105 with a tailored guided tour of a venue or the museum based on user specific objects of interest (animate or inanimate) within the venue or the museum (e.g., artwork the user 105 spends time appreciating).
[0072] Figure 1 H shows one or more additional Ul elements associated with the orchestrated guidance. In some examples, the Al guidance system 100 can present a highlight and/or one or more animations to identify a target object or object of interest. The head-wearable device 120 can include a dimmable lens controlled by the Al guidance system 100 and can provide additional information to the user 105 (e.g., directing the user 105’s focus to certain objects within their field of view). For example, in Figure 1 H, the Al guidance system 100 cause selective dimming of a portion of a display of the head-wearable device 120 such that an animated dimming target Ul element 139 is presented to the user 105. The animated dimming target Ul element 139 can be used to draw the user 105’s attention to a portion of the field of view 125 such that the use 105 can confirm a selected object of interest or be notified of a portion of the field of view 125 being analyzed by the Al guidance system 100.
[0073] Figure 1 H further shows a second context-based response presented at the wrist-wearable device 1 10, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100. For example, the Al agent 115 presents a second context-based response Ul element 141 via a display of the head-wearable device 120, as shown in field of view 125 of the user 105. The second context-based response is based on the object of interest identified by the user 105 and highlighted by the Al guidance system 100. The context-based responses can also be provided as audio responses (or audio guidance) via speakers of the wrist-wearable device 1 10, the head-wearable device 120, and/or any device communicatively coupled to the Al guidance system 100.
[0074] Turning to Figure 11, updated orchestrated guidance is presented to the user 105. In particular, second orchestrated guidance 143 including a second set of recommended actions e.g., Ul elements 144, 145, and 146) are presented to the user 105 via one or more wearable devices. The second orchestrated guidance 143 and the second set of recommended actions can be based on the user’s current and/or past experiences at the museum and/or during the museum tour. For example, in accordance with a determination by the Al guidance system 100 that the user 105 has not previously viewed landmarks near the museum, the Al agent 1 15 can provide a recommended action to explore the unseen landmarks (e.g., as shown by explore landmarks Ul element 145). As further shown in Figure 11, the user 105 provides a third user input 147 selecting an end tour Ul element 146.
[0075] Figure 1 J shows a notification summary presented at a wearable device of the Al guidance system 100. In some examples, the Al guidance system 100, in accordance with a determination that the end tour Ul element 146 was selected, ceases the user 105’s participation in the context-based activity (e.g., the museum visit). The Al guidance system 100, in accordance with a determination that the museum visit has ended, causes the wearable devices or other communicatively coupled devices to cease the do-not-disturb mode. The Al guidance system 100, after detecting that the do-not-disturb mode ceased, generate, using the Al agent 115, a notification summary based on the notifications received while the wearable devices (or other devices of the Al guidance system 100) were in the do- not-disturb mode. In some examples, the summary can be a natural language summary provided by the Al agent 115 that summarized the received notifications. The notification summary can be presented via visual feedback (e.g., notification summary Ul element 140 presented via a communicatively coupled display), audio feedback (e.g., text-to-speech presented via a communicatively coupled speaker), and/or haptic feedback.
[0076] Figure 1 K shows further updated orchestrated guidance. In particular, Figure 1 K shows a third orchestrated guidance 153 and a third set of recommended actions (e.g., Ul elements 154 and 155) presented at a wearable device. The Al agent 115 determines the third orchestrated guidance 153 and the third set of recommended actions based on the notifications received while the wearable devices (or other devices of the Al guidance system 100) were in the do-not-disturb mode. For example, the third orchestrated guidance 153 and the third set of recommended actions provide the user 105 with options for responding to received messages and missed calls.
[0077] As further shown in Figure 1 K, the user 105 forgoes selecting the third set of recommended actions. Alternatively, the user 105 provides a touch input 157 at the headwearable device 120 to initiate a microphone of the head-wearable device 120 (or other communicatively coupled device) and provide a voice command 151 to the Al guidance system 100. The voice command provided to the Al guidance system 100 can be used by the Al agent to determine another context-based activity (e.g., organizing dinner plans). The Al agent 115 can generate for the other context-based activity additional orchestrated guidance recommended action for performing the other context-based activity. For example, the Al agent 115 can generate orchestrated guidance for organizing dinner plans and recommended actions.
[0078] Figure 1 L shows the Al guidance system 100 utilizing a web-agent to assist the user in the performance of the other context-based activity and/or determine recommended actions. In some examples, in response to a user input selecting the recommended action for performing the context-based activity, the Al guidance system 100 can perform, using the Al agent, a (web or application) search based on the recommended action. The Al guidance system 100 can further determine a task to perform based on the search, and presenting the task at the wearable device. For example, In some examples, the Al guidance system 100 receives a request from a user to cause an Al agent to perform a task (e.g., “find a restaurant for dinner tomorrow downtown and make a reservation for 4”) and, based on content of the request, the Al guidance system 100 can determine that traversal of one or more web pages is required to perform the task that fulfills the request from the user. Further, the Al guidance system 100, responsive to the request, can traverse, using a web-based Al agent, one or more web pages and/or applications and, after the traversing, process the collected data to generate, via the Al agent, the response for the user 105 (e.g., response identifying a restaurant for 4 people and a time for making reservations).
[0079] In some examples, the Al guidance system 100 use the web-agent to autonomously carry out requests made by the user 105 even when the request is not associated with an API. In some examples, the Al guidance system 100 will report back on progress made in fulfilling the request of the user 105. For example, the Al guidance system 100 can report to the user 105 restaurant availability, restaurant wait times, errors in booking, reservation confirmations, etc. For example, as shown in Figure 1 L, the Al agent 115 identifies a restaurant and a reservation time for organizing the user 105’s dinner plans, and the Al guidance system 100 presents the restaurant and the reservation time to the user 105 via the wearable device (e.g., response Ul element 159).
[0080] In some examples, the Al guidance system 100 can utilize the web-agent (application-agent and/or other computer implemented agent) to assist the user 105 in collecting additional information for fulfilling the request from the user 105. For example, the Al guidance system 100 can search information related to social media posts to identify restaurant recommendations and/or restaurants in proximity and provide the information related to the social media posts to the Al agent 1 15 for generating a response and/or providing recommended actions. In some examples, the information is determined through the use of an Al model that is configured to determine additional information from images/videos/audio to provide contextual information (e.g., a picture of a posted restaurant and use an Al to determine which restaurant the poster was at). In some examples, the Al guidance system 100 can provide the user 105 with information about a previously seen social media post. In some examples, the Al guidance system 100 can be used to find additional information on posts or other content the user 105 has previously viewed via one or more devices, thereby providing unique results specific to the user’s viewing history.
[0081] In some examples, the Al guidance system 100 can perform additional Al actions to assist the user 105 and/or augment the user 105’s experience. For example, the Al guidance system 100 can proactively provide or silence notifications based on user situations determined by the Al agent 1 15 (e.g. the Al guidance system 100 can detect ongoing activities of the user 105 based on sensor data, audio data, and/or image data, and determine situations would benefit from additional focus (e.g., productivity tasks, participation in events, etc.) and silence non-essential notifications until the situations are complete). Additionally, the Al guidance system 100 can also proactively display information that is determined to be essential to the user 105 and/or predicted to be useful to the user 105 based on the environment of the wearable devices and/or other devices of the Al guidance system 100. For example, a wearable device, such as the head-wearable device 120, can automatically display a menu of a restaurant (that is determined to be of interest to the user 105) when the user 105 is in proximity (e.g., 3 feet, 6 feet, etc.) of the restaurant such that the user 105 does not have to perform an additional search (e.g. navigate a search engine to find the menu). In some examples, the Al guidance system 100 operations can occur without the need of user input (e.g. .touch inputs, voice commands, etc.).
[0082] Figure 1 M illustrate orchestrated guidance based on the restaurant identified by the Al guidance system 100. In particular, the Al guidance system 100 presents via the wearable devices a fourth orchestrated guidance 162 and a fourth set of recommended actions (e.g., Ul elements 163, 164, and 165). In Figure 1 M, the user 105 provides another voice command 161 to the Al guidance system 100 for performing an action corresponding to the orchestrated guidance for organizing dinner plans. The user 105 performs a pinch and hold gesture to initiate a microphone of the head-wearable device 120 (or other communicatively coupled device) and provide the other voice command 161 to the Al guidance system 100.
[0083] Figure 1 N shows the Al guidance system 100 providing confirmation of a completed task and generating an event for the user 105. For example, the Al guidance system 100 causes presentation of a task completion Ul element 167 via a display of the headwearable device 120. Additionally, the Al guidance system 100 also presents a calendar Ul element 169 showing an event or calendar invite generated by the Al agent 1 15.
[0084] The examples provided above are non-limiting. The Al guidance system 100 can be used to augment user experience of other activities. For example, the Al guidance system 100 can be used for a cooking context-based activity and the Al guidance system 100 can be used by the user 105 to find a recipe, make a dish based on the recipe, present guidance on preparation of the dish based on the recipe (e.g., step-by-step instructions, illustration, and/or video). Similar to the process described above, the Al guidance system 100 can use sensor data, audio data, and/or image data of wearable devices and/or other devices to determine a current step of the recipe and/or progress made by the user 105. For example, the user 105 can query the Al guidance system 100 on the next step of the recipe, and the Al guidance system 100 can provide tailored instructions to the user 105. In some examples, the Al guidance system 100 can provide information about steps of the recipe, how much time is left, determinations of food preparedness based on sensor data, audio data, image data, etc.
[0085] In another example, the Al guidance system 100 can augment a user experience of a game application. For example, a user can query the Al guidance system 100 to perform a task in game, and the Al guidance system 100 can leverage the one or more sensors of the wearable devices (e.g., the head-wearable device 120) and/or other devices in communication with the Al guidance system 100 to satisfy the request of the user 105. For example, the Al guidance system 100 can provide natural language responses to guide a user 105 within an augmented reality environment by using IMU data and image data (e.g., the device can state “There is a monster behind you, watch out!”). In some examples, the request to the Al guidance system 100 can initiate the game without the need for the user 105 to open the application themselves. In some examples, the Al guidance system 100 could output audio spatially to the user to help them identify where an interactable object is in a game.
[0086] In yet another example, the Al guidance system 100 can augment a user experience of a sports event or sports application. For example, the user 105 can ask the Al guidance system 100 a question about an ongoing Formula 1 race to understand the positions of the drivers - e.g., “compare the pace between two drivers.” The Al guidance system 100 can be configured to use live data from the application or the sports stream to provide the appropriate response. For sports that are heavily data driven, there is a lot of data that is not provided to the user 105, but the Al guidance system 100 can access any available data (e.g., microphone communications of one driver, tire data, lap times, showing different cameras of different drivers including selecting specific cameras on each car, etc.).
[0087] Figures 2A-2L illustrate context-based responses generated by an artificially intelligent agent based on activities performed by a user. Similar to Figures 1A-1 N, the operations shown in Figure 2A-2L can be performed by any XR systems described below in reference to Figures 7A-7C. For example, the operations of Figures 2A-2L can be performed by wearable devices, such as a wrist-wearable device 1 10 and/or a head-wearable device 120. The operations of Figures 2A-2L are performed by an Al assistive system 200 including at least a wrist-wearable device 1 10 and a head-wearable device 120 donned by the user 105 and/or other electronic devices described below in reference to Figures 7A-7C. The Al assistive system 200 can include the Al agent 115. The Al assistive system 200 is analogous to the Al guidance system 100 shown and described in reference to Figures 1A-1 N. In some examples, the Al assistive system 200 and the Al guidance system 100 are the same. Alternatively, In some examples, the Al assistive system 200 and the Al guidance system 100 are distinct system implemented at any XR systems described below in reference to Figures 7A-7C. Operations of the Al assistive system 200 and the Al guidance system 100 can be performed in parallel, sequentially, concurrently, and/or in a predetermined order.
[0088] In some examples, the Al assistive system 200 can augment the user 105’s experience in performing a physical activity and/or user experience while using a fitness application. The Al assistive system 200 can assist the use in the performance of different physical or fitness activities. The Al assistive system 200 can operate as a virtual coach and emulate a coach’s voice, provide specific instructions, and/or provide feedback to the user 105. For example, one or more sensors of the head-wearable device 120 and/or communicative coupled devices to determine whether the user 105 is performing the physical activity correctly. In accordance with a determination that the user 105 is not performing the exercise correctly, the Al assistive system 200 can provide guidance to the user 105 to improve performance of the exercise.
[0089] In Figures 2A-2L, the user 105 is participating in an activity with at least one other user 205. In some examples, the activity is physical exercise. For example, the user 105 and the at least one other user 205 are at a gym and start performing an exercise (e.g. a run). The Al assistive system 200, in response to an indication that the user 105 of a wearable device, such as the wrist-wearable device 1 10 and/or the head-wearable device 120, is participating in an activity, obtains data associated with an on-going activity performed by the user 105 of the wearable device. In some examples, the indication can be provided in response to a user input. For example, first user input 207 at the head-wearable device 120 initiating a workout. Alternatively, the Al assistive system 200 can generate the indication based on sensor data, audio data, and/or image data captured by one or more devices of the Al assistive system 200. For example, the Al assistive system 200 can detect that the user 105 is engaging in a physical activity, such as running, cycling, weightlifting, skiing, etc., and generate the indication that the user 105 is participating in an activity. In some examples, the Al assistive system 200 can generate the indication based on audio cues or context. For example, the user comment “ready for the run?” can be used to initiate and identify the activity. [0090] The Al assistive system 200 generates, using the Al agent 1 15, a contextbased response based, in part, on the data associated with the on-going activity performed by the user 105 of the wearable device and presents, at the wearable device, the context-based response. For example, as shown in Figure 2B, the Al agent 115 can generate a workout Ul 21 1 including activity information and a first context-based response (represented by first context-based response Ul element 209), and cause presentation of the workout Ul 21 1 and the first context-based response Ul element 209 at the head-wearable device 120. In some examples, the context-based response is presented within a portion of a field of view 212 of the user 105. In some examples, the context-based response and/or the workout Ul 211 are presented such that they are always visible to the user 105. For example, the context-based response and/or the workout Ul 21 1 can be positioned at a portion of the display of the wearable device reserved for the context-based response and/or the workout Ul 21 1. Alternatively, or in addition, the context-based response and/or the workout Ul 211 can be configured such that they are always overlayed over other applications and/or Uls.
[0091] In some examples, the context-based response is a coaching response to assist the user 105 on performance of the activity. For example, in Figure 2B, the first contextbased response Ul element 209 prompts the user 105 if they would like help with their workout. In some examples, the context-based response can include navigation instructions.
[0092] In some examples, the workout Ul 211 includes activity information, such as activity information Ul element 212 and activity route 217 (or activity map). In some examples, the workout Ul 21 1 includes biometric data to allow the user 105 to easily track their workout. For example, the workout Ul 211 can include real-time statistics including, but not limited to, speed, pace, splits, total distance, total duration, map, segments, elevation, gradient, heart rate, cadence, persona records (or PRs), challenges, and segment comparisons. In some examples, the Al assistive system 200 operates in conjunction with the wearable devices to automatically select information about the physical activity to present within the user interface elements. In some examples, the workout Ul 211 includes one or more quick access applications 215 that allow the user 105 to initiate one or more applications.
[0093] The Al assistive system 200 can present and/or share data rich overlay Uls that can include image data (e.g., Figures 2L and 2C) and/or other data about activities that the user 105 is performing. The Al assistive system 200 allows the user 105 to connect and engage with their communities in more interesting and engaging ways, by curating informative overlays to captured activities. For example, by providing the user 105 with capabilities for sharing personal states about physical activities that the user is performing, the Al assistive system 200 allows the user 105 to elevate and showcase their efforts and progress.
[0094] In some examples, the Al assistive system 200 can provide visual feedback to the user 105 via frames of the head-wearable device 120. For example, the head-wearable device 120 includes one or more indicators for assisting the user 105 in performance of the activity. For example, Figure 2B shows an interior portion 219 (e.g., face-facing portion of the frames) of the head-wearable device 120, the interior portion including a first light emitter portion 221 and a second light emitter portion 223. The first and the second light emitter portions 221 and 223 can be light-emitting diodes (LEDs). The Al assistive system 200 can use the first light emitter portion 221 and the second light emitter portion 223 to provide directions to the user (e.g., turn the first light emitter portion 221 on and the second light emitter portion 223 off to direct the user to the left, turn on both the first and the second light emitter portions 221 and 223 to direct the user to go forward; etc.). In some examples, the first and the second light emitter portions 221 and 223 can turn different colors, illuminate in different patterns and/or frequencies, and/or illuminate with different brightness to provide the user 105 with biometric information (e.g., green to indicate that the heart rate of the user 105 is in a first target threshold, yellow to indicate that the heart rate of the user 105 is in a second target threshold, red to indicate that the heart rate of the user 105 is in a third target threshold, etc.) [0095] In Figure 2C, the user 105 responds to the first context-based response via a voice command 225. In particular, the user 105 requests that the Al assistive system 200 assist the user 105 in setting a PR. The user 105 can provide different types of request to the Al assistive system 200. For example, the user 105 can provide the voice command 225 requesting that the Al agent 1 15 notify the user 105 when their heart rate is above a predefined threshold (e.g., heart rate goes above 165 BPM). The Al assistive system 200 can provides a series of visual and/or audio response to the user 105 based on the voice command 225 or other user request. The visual and/or audio response can be encouragement, suggestions, instructions, updates to biometric data, etc. In some examples, the Al assistive system 200 can provide the audio response in distinct vocal personalities and/or other characteristics, which may be based on the type of physical activity the user is performing (e.g., a personified Al agent). For example, the Al assistive system 200 can use the voice of a famous motivational runner in accordance with detecting that the user 105 is running a 10K.
[0096] In Figure 2D, the Al assistive system 200 generates, via the Al agent 115, second context-based response and updates to the workout Ul 211. For example, the Al assistive system 200 can generate a response to the voice command 225 and present the response to the user 105 via a wearable device (e.g., the second context-based response Ul element 227 presented within field of view 212). Additionally, the workout Ul 211 can be updated to show changes to biometric data (e.g., changes to calories burned, heart rate, etc.), workout completion, split times, etc.
[0097] Turning to Figure 2E, the user 105 provides the Al assistive system 200 a request to live stream their activity. The Al assistive system 200 can allow the user 105 to enable a live stream using wearable devices (e.g., the head-wearable device 120 and/or the wrist-wearable device 1 10) and/or other communicatively coupled device capture and transmit image data, audio data, and/or sensor data. For example, as shown in Figure 2E, the user 105 can provide another voice command 229 requesting that the Al assistive system 200 initiate a stream to capture their run. In some examples, the user 105 can initiate the live stream via a touch input at the wrist-wearable device 1 10 and/or the head-wearable device 120. In some examples, the user 105 can perform a gesture to select one or more Ul element for selecting a particular functionality. For example, the user can perform a pinch gesture to select the streaming Ul element 234.
[0098] In Figure 2F, the Al assistive system 200 provides a third context-based response confirming the initiation of the stream to the user 105 (e.g., third context-based response Ul element 231). The Al assistive system 200 can further present a streaming Ul 233 at the head-wearable device 120 and/or another streaming Ul 237 at the wrist-wearable device (or other communicatively coupled display). In some examples, the Al assistive system 200 can present static holographic elements 235 that provide simple information and/or images to the user 105. For example, the static holographic elements 235 can includes battery information, simplified notifications corresponding to stream interactions, and/or other Al agent 115 information (such as a camera view finder) can be presented.
[0099] The Al assistive system 200 can initiate the live stream on one or more platforms associated with the user 105. In some examples, the Al assistive system 200 can automatically select the streaming platform for the user 105 (e.g., based on user behavior). Alternatively, or in addition, the user 105 can provide a user input (e.g., voice command, touch input, gesture, etc.) identifying a streaming platform and/or selecting from one or more suggested streaming platforms identified by the Al assistive system 200. In some examples, the Al assistive system 200 notifies one or more followers of the user 105 that the live stream has been initiated. In other words, In some examples, the Al assistive system 200 can perform a complimentary operations to a requested operation of the user 105, which may be based on data about the user’s interaction history with the respective social platforms.
[00100] In some examples, the streaming Ul 233 and the other streaming Ul 237 include a chat of the live stream. Alternative, or in addition, the streaming Ul 233 and the other streaming Ul 237 can present the broadcasted stream (e.g., captured and transmitted image data, audio data, sensor data, and/or other transmitted data). In some examples, the user 105 can toggle information presented via the streaming Ul 233 and/or the other streaming Ul 237. For example, the user 105 can select one or more Ul elements within the streaming Ul 233 and/or the other streaming Ul 237 to toggle the presented information. Additionally, the user 105 can select a share Ul element to share additional content or information. In some examples, the Al assistive system 200 can apply one or more overlays and/or Ul elements to the streamed data such that the one or more overlays and/or Ul elements are viewable by devices receiving the streamed data. For example, the streamed image data can include information on the user’s current activity (e.g., current progress, percentage complete, and/or other information shared by the user 105). The Al assistive system 200 can provides automatic user interactions by automatically engaging the user 105 and/or with other communicatively coupled devices with streamed data.
[00101] Figure 2G and 2H shows the Al assistive system 200 connecting the user 105 with the at least one other user 205. In some examples, the Al assistive system 200, in accordance with a determination that the activity is a group activity performed with at least one contact ofthe user 105 (e.g., a friend or connection of the user 105 ), obtains from an electronic device associated with the at least one contact of the user 105 additional data associated with a respective on-going activity performed by the at least one contact of the user 105. The context-based response can further be based on the additional data associated with the respective on-going activity performed by the at least one contact of the user. For example, as shown in Figure 2G, the Al assistive system 200 presents via a display of the headwearable device 120 a context-based response (e.g., a fourth context-based response 239) prompting the user 105 if they would like to connect with a contact (e.g., the at least one contact 205), as well as an updated workout Ul 21 1 including a pin 241 or flag of a position of the at least one contact 205 relative to a current position of the user 105. Figure 2H further shows the user 105 providing a user input (e.g., yet another voice command 243) requesting that data be shared with the at least one contact 205.
[00102] In some examples, the Al assistive system 200 provides a plurality communication modalities in which the user 105 can quickly connect with friends and/or contacts. The Al assistive system 200 can be used to contact a single contact participating in a group activity or all contacts participating in the group activity. In some examples, the Al assistive system 200 can include one or more communication channels. For example, the Al assistive system 200 can include a walkie-talkie feature to quickly and effortlessly connect with one or more contacts. In some examples, the Al assistive system 200 can identify one or more participants in a group activity based on proximity data of one or more devices adjacent to wearable devices of the Al assistive system 200. Alternatively, or in addition, In some examples, the Al assistive system 200 can identify one or more participants in a group activity based on electronic devices attempting to communicatively couple with the wearable devices and/or other devices of the Al assistive system 200. In some examples, the Al assistive system 200 can identify one or more participants in a group activity based on the user 105’s contact list and/or by reviewing recent group conversations about an event or activity. In some examples, the Al assistive system 200 uses natural language systems to invoke a conversation with a group and quickly communicate with the group. For example, the user may invoke a conversation generally without specifying the recipients and based on what the user 105 asks, the Al assistive system 200 can determine the appropriate audience (e.g., asking “where is everyone?” when the user is at a food festival with friends). [00103] Figure 2I and 2J show a perspective of the at least one contact 205. In particular, Figures 2I and 2J show another Al assistive system (analogous to the Al assistive system 200) implemented on one or more wearable devices or other devices of the at least one contact 205. In Figure 2I, the other Al assistive system presents via a speaker of a headwearable device 253 of the at least one contact 205 a context-based response 245 prompting the at least one contact 205 if they would like to connect with the user 105. The at least one contact 205 further provides a voice command confirming that they would like to connect with the user 105.
[00104] Figure 2J shows a field of view 246 of the at least one contact 205 as viewed by the head-wearable device 253. The field of view 246 of the at least one contact 205 includes a first workout Ul 249 tracking the at least one contact 205’s workout and a second workout Ul 247 including shared workout information from the user 105. The first workout Ul 249 further includes a respective pin 250 identifying the location of the user 105 relative to the at least one contact 205. Figure 2J further shows the at least one contact 205 providing the other Al assistive system a request. For example, the request 251 from the at least one contact 205 asks the other Al assistive system to send an encouraging message to user 105. In some examples, the Al assistive system 200 of the user 105 can receive the encouraging message and automatically cause presentation of the visual and/or audio message. In some examples, the encouraging message can include a haptic feedback response. In some examples, the Al assistive system 200 presents the encouraging message after determining that the user 105 has achieved a particular milestone related to the performance of the activity. In some examples, users are able to unlock pre-recorded praise from the Al assistive system 200 (e.g., personified Al agents) and/or pre-recorded audio by professional athletes related to the physical activities that the user is performing.
[00105] Figure 2J further shows one or more indicators 255 on the head-wearable device 253 of the at least one contact 205. The indicators 255 of the head-wearable device 253 of the at least one contact 205 can be one or more light-emitters (e.g., LEDs). Similar to the first and second light emitter portions 221 and 223, the indicators 255 can communicate information to the at least one contact 205. For example, the indicators 255 can illuminate in different colors, patterns and/or frequencies, and/or brightness to convey information to the at least one contact 205. For example, the indicators 255 can illuminate to notify the at least one contact 205 when they are within target activity thresholds, performing an activity at a predetermined pace or speed, etc. In some examples, the indicators 255 provides a persistent indication to the at least one contact 205 based on whether a particular condition satisfies a predefined threshold. For example, based on the at least one contact 205 providing a user input activating the indicators 255, the indicators 255 can remain active until disabled. [00106] In some examples, the head-wearable device 253 are a pair of low-cost headwearable device that do not include a display and opt for presenting information via audio outputs and/or haptic feedback to the at least one contact 205. Alternatively, In some examples, the head-wearable device 253 can include low fidelity display that is configured to provide glanceable information. In some examples, this information may be text and glyphs (e.g., emoji’s, gifs, or low-resolution images) only, as opposed to media rich images (e.g., video or color images). In some examples, the low-fidelity display can be configured to display a single color (e.g., green) or grayscale. In some examples, the head-wearable device 253 can include an outward facing projector configured for displaying information. For example, the head wearable device 253 can be configured to display a text message onto a wearer’s hand or other surface. In some examples, the head-wearable device can project user interfaces such that a wearer can interact with a desktop-like user interface without needing to bring a laptop with them.
[00107] While these head-wearable devices are shown as having different features it is envisioned that a single head-wearable device could be configured to use all or a subset of these information presenting modalities.
[00108] As described above, the Al assistive system 200 can include different modalities for presenting and/or sharing information. While numerous modalities are discussed, it is envisioned that an operating system would be configured to present the information based on the device, and the developer would only need to specify the content to be presented and not the specific modality. In this way software can be produced to work across head-wearable devices with different capabilities (e.g., information output modalities). All of these devices described are configured to work with Al models for presenting information to users.
[00109] Figure 2K and 2L show additional data collected and/or shared during the performance of an activity (or group activity). For example, Figures 2K and 2L show image data collected during the performance of the group activity, shared image data between the members of the group activity, and/or synchronization of the image data. In Figure 2K, the other Al assistive system presents via a speaker or a display of the head-wearable device 253 of the at least one contact 205 another context-based response 257 prompting the at least one contact 205 if they would like receive and synchronize image data shared by the user 105. The at least one contact 205 further provides a voice command confirming that they would like to connect receive and synchronize image data shared by the user 105.
[00110] In other words, the Al assistive system 200 include sharing operations for creating and sharing user interfaces that include imaging data captured by the intelligent autocapture assistive operations. The Al assistive system 200 provides user interfaces that include image data that is captured while a user performing a physical activity (e.g., a fitness activity, such as performing a bike ride). In addition to the image data, the user interfaces also include user interface elements generated based on other data, different than the image data, related to the user’s performance of the respective physical activity. In some examples, the Al assistive system 200 is configured to allow users to tag captured media with personal metadata (e.g., real-time statistics). For example, the user interfaces may include engaging montages of captured images and other content about the performance of the physical activity. As shown in Figure 2L, an image sync Ul 259 can be configured display captured image data, combined image data (e.g., combined first image data 261 and second image data 263), and/or image montages. In some examples, the image sync Ul 259 can be presented at other devices of the Al assistive system 200. In some examples, the Al assistive system 200, in accordance with a determination that a plurality of video streams are (i) captured within a predefined amount of time of each other and (ii) within a predefined distance of each other, prepares a collated video of two or more of the plurality of video streams in a time- synchronized fashion.
[00111] Figures 3A-3D illustrate example user interfaces and additional features available at the Al assistive system 200. Figures 3A and 3B show a map application and directions provided via the Al assistive system 200. Figures 3C and 3D show automatic image capture capabilities of the Al assistive system 200.
[00112] In Figures 3A and 3B, the Al assistive system 200 presents a map Ul 307. The map Ul 307 can include one or more Ul elements providing directions to the user 105. For example, the map Ul 307 can include a next step Ul element 309 including the next directions to take, as well as a path highlight 305 (which can be overlaid over the next path in the directions). In some examples, the user 105 can toggle between application via one or more user inputs. For example, the user 105 can cause presentation of the map Ul 307, via a wearable device of the Al assistive system 200, in response to user selection of the map application Ul element 308. Additionally, or alternatively, In some examples, the Al assistive system 200 presents context-based responses 305 providing directions to the user 105.
[00113] Figure 3B shows a map settings Ul 313. The map settings Ul 313 can be presented in response to user input 31 1 (selecting the downward arrow). The map settings Ul 313 provides one or more options for allowing the user 105 to select settings for voiced directions (e.g., on, off, and/or a particular voice), visual direction indicators (e.g., path highlights, next step Ul elements, etc.), view (e.g., setting 2D, 3D, and/or street views), location sharing (e.g., privacy setting for sharing location, automatic sharing of location, etc.), etc.
[00114] Turning to Figures 3C and 3D, the Al assistive system 200 presents an image capture Ul 317. The image capture Ul 317 can include one or more Ul elements for showing captured image data and/or options 323 for modifying, sharing, and/or dismissing the captured image data. For example, the image capture Ul 317 can include first and second image data 319 and 321 captured during the activity of the user 105. In some examples, the user 105 can toggle between application via one or more user inputs. For example, the user 105 can cause presentation of the image capture Ul 317, via a wearable device of the Al assistive system 200, in response to user selection of the image application Ul element 318. Additionally, or alternatively, In some examples, the Al assistive system 200 presents contextbased responses 315 providing information on the automatically captured image data.
[00115] Figure 3D shows a capture settings Ul 327. The capture settings Ul 327 can be presented in response to user input 325 (selecting the downward arrow). The capture settings Ul 327 provides one or more options for allowing the user 105 to select settings for capture triggers (e.g., triggers that cause the automatic capture of image data, such as changes in movement, instant spikes in acceleration, activity milestones (e.g., hitting a baseball with the baseball bat), changes in vibration, etc.), capture settings (e.g., image capture setting such as resolution, format, frames per second, etc.), tagging options (e.g., settings identifying people and/or objects to be tagged), sharing options (e.g., privacy setting for sharing image data, identifying images that can be shared, frequency at which image data is shared, etc.), etc. In some examples, the Al assistive system 200 is configured to perform sharing operations based on the user input in accordance with determining that the user has already enabled the automatic image-capture operations. In some examples, the Al assistive system 200 can perform automatic smoothing functions on image data.
[00116] Figures 4A and 4B illustrate example sequences of user interactions with personalized assistive systems (e.g., the Al guidance system 100 and/or the Al assistive system 200; Figures 1 A-2L). The legend in the top right of Figures 4A and 4B indicates types of interactions and input modes for each respective segment of the timeline flow. The task icon 401 indicates a productivity-based interaction, media-play icon 405 indicates media and/or an “edutainment” interaction, the messaging icon 407 indicates a communicationbased interaction, the information icon 409 indicates an information-based interaction, the solid line 41 1 indicates a touch input, the double line 413 indicates a wake word input, the triple line 415 indicates an Al chat session.
[00117] The interaction sequences of Figures 4A and 4B can be performed by a user that is wearing a head-worn device 120 (e.g., AR device 728) while the user of the device is performing a sequence of daily activities. The head-worn device 120 (Figures 1A-3D) includes or is in electronic communication with an assistive system for assisting in interactions with the head-worn device 120 to cause operations to be performed. For example, the head-worn device 120 may provide information (e.g., information related to data collected about a physical activity that a user is performing, an alert about an incoming message) without explicit user input to do so. [00118] In accordance with some examples, the user can perform voice commands to cause operations to be performed at the head-worn device 120. For example, as shown in block 402, the user can provide a voice command to turn on do-not-disturb (DND) at their head-worn device 120, with an option set for VIP exceptions, which would allow for certain users’ messages or other requests may be allowed. In some examples, the assistive system, in accordance with receiving the request to turn on do not disturb, determines a set of potential operation commands that the request may correspond to.
[00119] As shown in block 404, the assistive system can determine to check one or more messenger threads accessible via the head-worn device 120 to determine a bike ride location for a bike ride that the user is participating in. In some examples, the assistive system performs the operations in response to a question by the user that does not directly provide instructions to search the user’s messages for the bike ride location. In other words, in accordance with some examples, the assistive system is capable of performing a set of operations based on a general prompt provided by the user.
[00120] As shown in block 406, the head-worn device 120 can automatically begin providing real-time navigation (e.g., via the assistive system or a different navigational application) to the user based on determining that the user is performing a bike ride along a particular navigational route. That is, the assistive system may be capable of determining when a user is performing an activity that can be enhanced by content from a different application stored in memory or otherwise in electronic communication with the head-worn device 120 (e.g., an application stored on the user’s smart phone).
[00121] As shown in block 408, the head-worn device 120 can provide message readouts from a group message for fellow cyclists to keep the user informed about updates in the chat while the user is performing the physical activity.
[00122] As shown in block 410, the head-worn device 120 can provide capabilities for the user to send and receive voice messages to other members of the cycling group chat.
[00123] As shown in block 412, the head-worn device 120 can cause the userto receive a text message (e.g., an audio readout of the text message) based on a determination that the message sender is from a user that qualifies under the VIP exceptions for the do not disturb setting that was instantiated at block 402. That is, In some examples, the assistive system can determine whether a particular received message should be provided for audio readout to the user based on settings of a different application.
[00124] As shown in block 414, the head-worn device 120 can cause a different group thread (e.g., a noisy group thread) to be silenced, such that audio readouts are not provided by the particular messaging thread. As shown in block 416, the assistive system can unmute and catch up on soccer group thread in messenger after ride. As shown in block 418, the assistive system can allow the user to message soccer group thread in messenger in response to a received message from the group thread. As shown in block 420, the assistive system can allow a user to record a voice note about new commitments to soccer group, which may be provided to the user by the assistive system based on a prompt inquiring about the user’s availability for a particular event and/or time.
[00125] As shown in block 422, the assistive system can allow the user to look up local family events happening this weekend (e.g., by providing a general prompt about the user’s availability). In some examples, the assistive system can provide the information to the user about the family events based on a different event that has occurred at the head-worn device 120 (e.g., receiving a different message from a different user about the user’s availability to participate in a cycling event).
[00126] As shown in block 424, the user can receive a summary of a specific family event, for example, in accordance to provide an input in response to receiving the information about local family events happening that weekend. As shown in block 426, the user can provide an input (e.g., “Hey Al assistant, repeat that on my phone”) to cause a previous audio message from the assistive system to be provided at a different electronic device (e.g., “Play last Al response on phone speaker for child to hear”). As shown in block 428, the user can also share feedback from the assistive system (e.g, an Al response) with another user (e.g., the user’s partner) on a different application, different than the application that is providing the assistive system (e.g., a messaging application).
[00127] As shown in block 430, the user can receive a real-time game notification from sports app. As shown in block 432, the user can cause the assistive system to provide on- demand translation for audio or textual content in another language. In some examples, the on-demand translation can be provided automatically based on a user request to read out content that is not in the user’s native language. As shown in block 434, the user can request slower speed translation. As shown in block 436, the user can receive voice messages from the cycling group on messenger. As shown in block 438, the user can mute a noisy messenger group chat, which the assistive system may be configured to automatically recognize based on a frequency that electronic messages are being received by the head-worn device 120 or another electronic device in electronic communication with the head-worn device 120. As shown in block 440, the user can check messages.
[00128] As shown following block 440, the assistive system can provide a notification to the user about a geographic landmark that the user is in proximity too (e.g., as determined by a navigational application on the user’s phone (e.g., “Location Update: At Farmer’s Market”). As shown in block 444, the assistive system can be configured to provide new recipe ideas for a new ingredient (e.g., an ingredient purchased at the farmer’s market). In some examples, the suggestions can be provided in accordance with receiving purchase confirmation at the head-wearable device about a purchase that the user made at the farmers’ market.
[00129] Figure 4B illustrates another timeline view of another interaction sequence with a head-worn device 120 (e.g., AR device 700) while a user of the device is performing a sequence of daily activities. As shown in Figure 4B, the user can engage in an Al chat session (as indicated by the red segment) to perform various activities to start their day (e.g., block 446 to check the local time while traveling, block 448 to set an alarm to leave for the airport later, block 450 to check the weather to decide what to wear, block 452 to check the calendar for a time and/or location of the next event, block 454 to look up local business address and hours, block 456 to message a colleague, and block 458 to listen to news on a podcast). In some examples, once a user activates another application that persistently provides audio feedback (e.g., a podcast), the assistive system can be configured to automatically stop the Al chat session.
[00130] After stropping the Al chat session, the user can perform a sequence of touch inputs, which may be used to cause the assistive operations to perform various functions, including those related to the audio outputs of the assistive system (e.g., block 460 to receive a text message reply from a colleague, block 462 to replay to the text message, block 464 to resume a podcast, block 466 to book a rideshare to an upcoming event, block 468 to receive a notification about the arrival of the rideshare, block 470 to check status of the rideshare, block 472 to call the rideshare to clarify pickup location, block 474 to listen to a music playlist while chatting, block 476 to receive an alarm to leave for the airport, block 478 to check a status of a flight, block, 480 to receive a reminder to buy a gift before departure of the flight, block 482 to call a partner on a messaging application, and block 484 to listen to meditation for the user’s flight anxiety). In some examples, the touch inputs provided by the user corresponding to one or more of blocks are based on universal gestures corresponding to universal inputs at the AR device 728, while one or more other blocks may correspond to user inputs provided to contextual input prompts (e.g., in response to an assistive prompt provided by the head-worn device 120).
[00131] Thus, as shown in Figures 4A and 4B, the systems described herein allow users to interact with an assistive system provided at the head-worn device 120 to allow for increased efficiency and effectiveness of the user’s interactions with the head-worn device 120. For example, the assistive system can allow for the user to use the head-worn device 120 as a tool to help level up their efficiencies, including by allowing for multi-tasking and productivity on the go. The assistive systems and devices described herein also allow the user to interact with the assistive system relatively inconspicuously, allowing for them to perform actions without distracting others around them. [00132] Figures 5 and 6 illustrates flow diagrams of methods of generating Al contextbased response and actions. Operations (e.g., steps) of the methods 500 and 600 can be performed by one or more processors (e.g., central processing unit and/or MCU) of an system XR system (e.g., XR systems of Figures 7A-7C-2). At least some of the operations shown in Figures 5 and 6 correspond to instructions stored in a computer memory or computer-readable storage medium (e.g., storage, RAM, and/or memory). Operations of the methods 500 and 600 can be performed by a single device alone or in conjunction with one or more processors and/or hardware components of another communicatively coupled device (e.g., wristwearable device 1 10 and a head-wearable device 120) and/or instructions stored in memory or computer-readable medium of the other device communicatively coupled to the system. In some examples, the various operations of the methods described herein are interchangeable and/or optional, and respective operations of the methods are performed by any of the aforementioned devices, systems, or combination of devices and/or systems. For convenience, the method operations will be described below as being performed by particular component or device, but should not be construed as limiting the performance of the operation to the particular device in all examples.
[00133] (A1) Figure 5 shows a flow chart of a method 500 for generating orchestrated guidance based on an activity of a user. The method 500 occurs at a wrist-wearable device 110, head-wearable device 120, and/or other wearable device including one or more sensors, imaging devices, displays, and/or other components described herein. The method 500 includes in response to an indication received at a wearable device that an artificial intelligence (Al) agent trigger condition is present, providing (502) an Al agent sensor data obtained by the wearable device. For example, as shown and described in reference to Figures 1A-1 N, a wrist-wearable device 1 10 and/or a head-wearable device 120 of a user can use image data, location data, audio data, and/or other data to detect the presence of an Al agent trigger condition. Non-limiting examples of Al agent trigger conditions include user queries, objects of interest, locations of interest, people of interest, time of day, user invocation, etc.
[00134] The method 500 includes determining (504), by the Al agent, a context-based activity based on the sensor data obtained by the wearable device. The context-based activity is an interpretation of a particular activity, action, and/or event with which the user is engaged. For example, as shown and described in reference to Figures 1A-1 N, the context-based activity is a museum visit or museum tour. Non-limiting examples of context-based activities include shopping, driving, sightseeing, traveling, exploring, cooking, gardening, tours, social meetings, productivity based tasks (e.g., working, note takings, etc.), exercising, etc. The method 500 includes generating (506), by the Al agent, orchestrated guidance based on the context-based activity and presenting (508) the orchestrated guidance at the wearable device. [00135] The orchestrated guidance includes a recommended action for performing the context-based activity. The orchestrated guidance can be a single recommended action, a sequence of recommended actions, and/or or concurrent (and/or parallel) recommended actions for performing the context-based activity. For example, as shown and described in reference to Figures 1A-1 N, the orchestrated guidance can be one or more recommended actions for facilitating the user’s museum tour, such as a recommended action for placing the user devices on “do not disturb,’’ a recommended action for initiating a guided tour, recommended actions for exploring museum exhibits, presentation of a summary collating missed notifications and/or messages while the user was engaged in the tour, and recommended actions for responding to the missed notifications and/or messages. The orchestrated guidance can be number of recommended actions for assisting the user in performance of the context-based activity - e.g., actions to be performed, during, or after the context-based activity.
[00136] (A2) In some examples of A1 , the context-based activity is a first context-based activity, the sensor data is first sensor data, the orchestrated guidance is first orchestrated guidance, the recommended action is a first recommended action, and the method 500 further includes, in accordance with a determination that the first recommended action for performing the first context-based activity was performed (or was ignored), providing the Al agent second sensor data obtained by the wearable device, determining, by the Al agent, a second contextbased activity based on the second sensor data obtained by the wearable device, generating, by the Al agent, second orchestrated guidance based on the second context-based activity and presenting the second orchestrated guidance at the wearable device. The second orchestrated guidance including a second recommended action for performing the second context-based activity. In other words, the method can build on different recommended actions and/or orchestrated guidance. For example, as shown and described in reference to Figures 1A-1 N, the user can accept one or more recommended actions (e.g., Figures 1A-1J) and/or cause the Al agent to generate new recommended actions (e.g., Figures 1 K-1 N - initiating a new context-based activity of searching for a restaurant).
[00137] (A3) In some examples of any one of A1-A2, the context-based activity is a first context-based activity of a plurality of context-based activities determined by the by the Al agent based on the sensor data, the orchestrated guidance includes a plurality of recommended actions for performing the plurality of context-based activities, and the recommended action is a first recommended action of the plurality of recommended actions, the first recommended action being configured to perform the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions. In other words, any number of context-based activities can be determined for a user and respective orchestrated guidance (and associated recommended actions) can be determined for the context-based activities and presented to the user.
[00138] (A4) In some examples of A3, generating the orchestrated guidance includes determining a subset of the plurality of recommended actions for performing the first contextbased activity, and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions and the subset of the plurality of recommended actions for performing the first context-based activity. In other words, a plurality recommended of actions associated with a context-based activity can be presented to the user. For example, as shown and described in reference to at least Figure 1A, at least two recommended actions are presented to the user in accordance with a determination that the user is visiting a museum.
[00139] (A5) In some examples of any one of A3-A4, generating the orchestrated guidance includes determining a sequence of context-based activities of the plurality of context-based activities to be performed, including a second context-based activity to follow the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action and the second recommended action of the plurality of recommended actions for performing the plurality of context-based activities. For example, as shown and described in reference to at least Figures 1A-1 E, a string of recommended actions are presented to the user and the recommended actions are updated based oner inputs selecting one or more user inputs. Similarly, deviations from the recommended actions are shown and described in reference to at least Figures 1 K- 1 N.
[00140] (A6) In some examples of any one of A1-A5, the method 500 includes, in response to a user input selecting the recommended action for performing the context-based activity, causing the wearable device to initiate a do-not-disturb mode (or focus mode, away mode, etc.). While in the do-not-disturb mode, the wearable device suppresses, at least, received notifications, and in response to an indication that participation in the context-based activity ceased causing the wearable device to cease the do-not-disturb mode; generating, by the Al agent, a notification summary based on the notifications received while the wearable device was in the do-not-disturb mode; and presenting the notification summary at the wearable device. Examples of the do-not-disturb mode and the notification summary are shown and described in reference to at least Figures 1A-1 J.
[00141] (A7) In some examples of any one of A1-A6, the method 500 includes, in response to a user input selecting the recommended action for performing the context-based activity, performing, by the Al agent, a search based on the recommended action, determining a task to perform based on the search, and presenting the task at the wearable device. An example search request provided by a user is shown and described in reference to at least Figures 1 K and 1 L.
[00142] (A8) In some examples of any one of A1-A7, presenting the orchestrated guidance at the wearable device includes, at least one of causing presentation of a user interface element associated with the orchestrated guidance at a communicatively coupled display, and causing presentation of audible guidance associated with the orchestrated guidance at a communicatively coupled speaker. An examples of one or more user interface elements associated with the orchestrated guidance and audible guidance are shown and described in reference to at least Figure 1 H.
[00143] (A9) In some examples of any one of A1-A8, the context-based activity is to be performed at a physical activity. For example, as described above, the context-based activity can be an exercise and a recommended action is performance of a particular routine or exercise (detected by the wearable device or another communicatively coupled device).
[00144] (B1) In accordance with some examples, a method includes receiving sensor data from one or more sensors of a head-wearable device and in response to receiving the data from the one or more sensors of the head-wearable device, processing the data, via an Al agent, to analyze the sensor data to identify a task performed or to be performed by a user, and causing the Al agent to provide guidance associated with performance of the task. For example, a head-wearable device 120 can cause performance of the operations shown and described in reference to Figures 1A-1 N.
[00145] (B2) In some examples B1 , the causing occurs in response to a selection at a wrist-wearable device of a user interface element that indicates that a guided tour is available. For example as shown and described in reference to Figures 1A and 1 B, user interface elements corresponding to a guided tour can be presented at a head-wearable device 120 and/or a wrist-wearable device 1 10.
[00146] (B3) In some examples of any one of B1 and B2, the sensor data from the one or more sensors is one or more of microphone data, camera data, movement data, and positioning data. In other words, sensor data captured by the wrist-wearable device, the headwearable device, and/or any other communicatively couple device can be used by the Al agent.
[00147] (B4) In some examples of any one of B1-B3, the method further includes, after causing the Al agent to provide guidance associated with the task, receiving additional sensor data from the one or more sensors of the head-wearable device, in response to receiving the additional sensor data from the one or more sensors of the head-wearable device, processing the additional sensor data, via the Al agent, to identify an additional task performed or to be performed by the user, and causing the Al agent to provide guidance associated with the additional task. In other words, the Al agent can determine subsequent tasks based on additional data received.
[00148] (B5) In some examples of B4, the additional task is related to the task.
[00149] (C1) In accordance with some examples, a method includes receiving a request at an Al agent to (i) forgo immediate output of incoming notifications and (ii) provide a summary of the incoming notifications at a later time, receiving a plurality of notifications, providing the notifications to a large language model (LLM), producing, using the LLM, a summary of the plurality of notifications, and providing a natural language summary, via an output modality of a head-wearable device, at the later time. Examples of summarized notifications are shown and described in reference to Figure 1 J.
[00150] (D1) In accordance with some examples, a method includes receiving a request from a user interacting with an Al agent, the request requiring traversing content on a website using the Al agent. The method also includes, in response to receiving the request, traversing, using an computer-implemented agent associated with the Al agent, one or more graphical user interfaces associated with the website to collect data needed to formulate a response to the request from the user, and after the traversing, processing the data collected by the computer-implemented agent associated with the Al agent to generate the response and providing the response to the user. For example, as shown and described in reference to Figure 1 K-1 N, the Al agent can utilize a web agent to search webpages and/or perform a web search to complete a user request and provide a corresponding response. In some examples, the web-based Al agent is distinct from the Al agent that received the task request. In some examples, different training data used to train that Al agent and the web-based agent. In some examples, the traversing the one or more web pages includes obtaining data needed to formulate a response to the request from the user. In some examples, surface Ul element related to progress of the Al agent in performing the traversal is presented (e.g., an Al agent symbol moving or spinning to show progress). In some examples, the web-based agent can be used to inquire about a contact (e.g., ask about a particular person that may be a contact of the user - e.g., “What kind of trip would Mike go on?”).
[00151] In some examples of A1-D1 , the context-based activities are further determined based on stored user data (e.g., use data about the user’s previous experiences and/or interests to curate the information about the guided tour). For example, if the user previously participated in an experience that was relevant to an aspect of the guided tour (e.g., Figures 1A-1 N), the Al agent may cause information about the previous event to surface or otherwise be integrated into the guided tour.
[00152] In some examples of A1-D1 , a head-wearable device is a display-less AR headset. In some examples, the input/output interface of the head-wearable device only includes one or more speakers. In some examples, the operations of the head-wearable device can be performed by a set of earbuds or other head-worn speaker device.)
[00153] In some examples of A1 -D1 , a user interface associated with the orchestrated set of guidance instructions is provided by the Al agent via a Lo-Fi display, the Lo-Fi being a glanceable display that presents notifications, live activities, Al agent information, and messages.
[00154] In some examples of A1 -D1 , a user interface associated with the orchestrated set of guidance instructions is provided by the Al agent via a projector display, the projector display configured to project information at a hand of the user (e.g., at a palm or other body part of the user).
[00155] In some examples of A1-D1 , a non-textual user interface element is presented at the head-wearable device (e.g., an audio message, an arrow or similar symbol), and the non-textual user interface element is configured to direct a user of the head-wearable device toward a physical landmark as part of the orchestrated set of guidance instructions.
[00156] In some examples of A1-D1 , the user can select objects within a field of view of the user (e.g., captured by one or more sensors of a wearable device, such as an imaging device) to receive additional information on the selected object.
[00157] In some examples of A1-D1 , the Al agent may cause some notifications to be muted during the guided tour, and then provide with an Al-generated summary of the conversations later so that the user can quickly catch up without reviewing many different messages right away.
[00158] (E1) Figure 6 shows a flow chart of a method 600 for facilitating performance of a physical activity performed by user. The method 600 occurs at a wrist-wearable device 110, a head-wearable device 120, and/or other wearable device including one or more sensors, imaging devices, displays, and/or other components described herein. The method 600 includes, in response to an indication that a user of a head-wearable device is participating in an activity, obtaining (602) data associated with an on-going activity performed by the user of the head-wearable device. The method 600 includes generating (604), by an Al agent, a context-based response based, in part, on the data associated with the on-going activity performed by the user of the head-wearable device, and presenting (606), at the headwearable device, context-based response. The context-based response is presented within a portion of a field of view of the user. For example, as shown and described in reference to Figures 2A-2H, a head-wearable device 120 can present different context-based responses to the user based on a physical activity being performed.
[00159] (E2) In some examples of E1 , the method 600 includes, in accordance with a determination that the activity is a group activity performed with at least one contact of the user, obtaining, from an electronic device associated with the at least one contact of the user, additional data associated with a respective on-going activity performed by the at least one contact of the user. The context-based response is further based on the additional data associated with the respective on-going activity performed by the at least one contact of the user. For example, as shown and described in reference to Figures 2I-2L, an Al agent can detect other contacts performing an activity with a user and share information between the users.
[00160] (E3) In some examples of E2, the data associated with the on-going activity performed by the user of the head-wearable device and the additional data associated with the respective on-going activity performed by the at least one contact of the user includes respective image data and/or audio data, and the context-based response is an image response including a combination of the respective image data. For example, as shown and described in reference to Figure 2L, image data captured between the wearable devices can be synchronized, combined into a single image, and/or combined into an image collage.
[00161] (E4) In some examples of E3, the respective image data includes a plurality of video streams from a plurality of respective head-wearable devices, and generating, by the Al agent, the context-based response includes in accordance with a determination that the plurality of video streams are (i) captured within a predefined amount of time of each other and (ii) within a predefined distance of each other, preparing a collated video of two or more of the plurality of video streams in a time-synchronized fashion. In some examples, the method includes providing to each of the respective head-wearable devices the collated video. At least one aspect of the collated video provided to each of the respective head-wearable devices is tailored to that respective head-wearable device.
[00162] (E5) In some examples of any one of E1-E4, the activity is a physical exercise; and the context-based response is a coaching response to assist the user on performance of the physical exercise. For example, as shown and described in reference to Figures 2A-2H, an Al agent can coach a user through an exercise.
[00163] (E6) In some examples of any one of E1-E5, the activity is outdoor physical activity (e.g., running, biking, hiking, etc.), and the context-based response is a navigation instructions. For example, as shown in at least Figures 3A and 3B, the Al agent can provide navigation instructions to the user.
[00164] (E7) In some examples of any one of E1 -E6, the activity is participation in a note-taking session (e.g., a meeting, class, lecture, etc.), and the context-based response is a request to generate notes. While the primary example shown in Figures 2A-2L is an exercise, the Al agent can be used with other activities performed by the user.
[00165] (F1) In accordance with some examples, a method is performed at a headwearable device including (i) one or more cameras, and (ii) a display component configured to display digital content. The method includes determining that a user wearing the head- wearable device is performing a physical activity and, in accordance with determining that the user wearing the head-wearable device is performing the physical activity, automatically, without additional user input, initializing assistive operations based on data provided by the one or more cameras of the head-wearable device. The method also includes, while the assistive operations are being performed based on image data from the one or more cameras of the head-wearable device, identifying, based on the assistive operations, that at least a portion of a respective field of view of a respective camera of the one or more cameras satisfies automatic-image-capture criteria for automatically capturing an image. The method further includes, based on the identifying, causing the respective camera to capture an image automatically, without further user input. For example, as shown and described in reference to Figure 3A, a wearable device can automatically capture image data.
[00166] (F2) In some examples of F1 , the method further includes detecting a user input directed to a universal action button on a peripheral portion of the head-wearable device. The assistive operations are initialized based on the user input being detected while the user is performing the physical activity. For example, as shown and described in reference to Figure 2A, the user can perform a tap gesture at a wearable device, such as the head-wearable device, to initiate the Al agent and/or other operations.
[00167] (G1) In accordance with some examples, a method includes receiving (i) performance data corresponding to a physical activity that a user of a head-wearable device is performing, and (ii) capturing image data by the head-wearable device during performance of the physical activity. The method also includes causing presentation, at a display component of the head-wearable device, a user interface element that includes one or more representations of the performance data, and responsive to provided user preferences, automatically sharing a field of view of the user in conjunction with sharing the user interface element as a composite user interface element to one or more other electronic devices. For example, as shown and described in reference to Figures 2G-2L, information captured by wearable devices can be shared between users.
[00168] (G2) In some examples of G1 , the performance data is received from a software application different than another software application that is performing operations at the head-wearable device for capturing the image data. For example, the information can be received from a streaming application and/or other application.
[00169] (H1) In accordance with some examples, a method includes determining that a user of a head-wearable device is beginning performance of a physical activity while data about the physical activity is configured to be obtained by the head-wearable device of the user and, in accordance with the determining that the user of the head-wearable device is beginning performance of the physical activity, identifying an assistive module that uses one or more specialized artificial-intelligence models. The method also includes causing interactive content to be provided to the user via the assistive module based on the data obtained about the physical activity that the user is performing. For example, as shown and described in reference to Figures 2A-2D, information captured by wearable devices can be used to assist the user in performance of the activity.
[00170] (H2) In some examples of H1 , the method further includes generating an audio message using an artificial intelligence model of the assistive module performing operations during performance of the physical activity by the user and determining based on data obtained about performance of the physical activity by the user, that one or more messageproviding criteria are satisfied. The method also includes, in accordance with the determining that the one or more message-providing criteria are satisfied, generating, using an Al model, a message related to the performance of the physical activity, and providing the generated electronic message to the user via one or of (i) a microphone of the head-wearable device, and (ii) a display component within a frame of the head-wearable device.
[00171] (11) In accordance with some examples, a method includes, at a head-worn device including a user interface for providing user interface elements to a user based on physical activities that the user is performing, receiving an update about a location of a user, based on a physical activity that the user is performing, and in accordance with receiving the indication, presenting a navigational user interface to the user providing navigation to the user based on an identified activity that the user is performing while wearing a head-worn device. For example, as shown and described above in reference to Figure 3A, navigation instructions can be provided to the user.
[00172] (J1) In accordance with some examples, a system that includes one or more wrist wearable devices and a pair of augmented-reality glasses, and the system is configured to perform operations corresponding to any of A1-I1.
[00173] (K1) In accordance with some examples, a non-transitory computer readable storage medium including instructions that, when executed by a computing device in communication with a pair of augmented-reality glasses, cause the computer device to perform operations corresponding to any of A1- 11.
[00174] (L1) In accordance with some examples, a means for performing or causing performance of operations corresponding to any of A1- 11.
[00175] (M1) In accordance with some examples, a wearable device (a head-wearable device and/or a wrist-wearable device) configured to perform or cause performance of operations corresponding to any of A1- 11.
[00176] (N1) In accordance with some examples, an intermediary processing device
(e.g., configured to offload processing operations for a wrist-wearable device and/or a head- worn device (e.g. augmented-reality glasses)) configured to perform or cause performance operations corresponding to any of A1- 11. [00177] Figures 7A - 7C-2 illustrate example XR systems that include AR and MR systems. Figure 7A shows a first XR system 700a and first example user interactions using a wrist-wearable device 726, a head-wearable device (e.g., AR device 728), and/or a HIPD 742. Figure 7B shows a second XR system 700b and second example user interactions using a wrist-wearable device 726, AR device 728, and/or an HIPD 742. Figures 7C-1 and 7C-2 show a third MR system 700c and third example user interactions using a wrist-wearable device 726, a head-wearable device (e.g., an MR device such as a VR device), and/or an HIPD 742. As the skilled artisan will appreciate upon reading the descriptions provided herein, the aboveexample AR and MR systems (described in detail below) can perform various functions and/or operations.
[00178] The wrist-wearable device 726, the head-wearable devices, and/or the HIPD 742 can communicatively couple via a network 725 (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN). Additionally, the wrist-wearable device 726, the head-wearable device, and/or the HIPD 742 can also communicatively couple with one or more servers 730, computers 740 (e.g., laptops, computers), mobile devices 750 (e.g., smartphones, tablets), and/or other electronic devices via the network 725 (e.g., cellular, near field, Wi-Fi, personal area network, wireless LAN). Similarly, a smart textile-based garment, when used, can also communicatively couple with the wrist-wearable device 726, the head-wearable device(s), the HIPD 742, the one or more servers 730, the computers 740, the mobile devices 750, and/or other electronic devices via the network 725 to provide inputs.
[00179] Turning to Figure 7A, a user 702 is shown wearing the wrist-wearable device 726 and the AR device 728 and having the HIPD 742 on their desk. The wrist-wearable device 726, the AR device 728, and the HIPD 742 facilitate user interaction with an AR environment. In particular, as shown by the first AR system 700a, the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 cause presentation of one or more avatars 704, digital representations of contacts 706, and virtual objects 708. As discussed below, the user 702 can interact with the one or more avatars 704, digital representations of the contacts 706, and virtual objects 708 via the wrist-wearable device 726, the AR device 728, and/or the HIPD 742. In addition, the user 702 is also able to directly view physical objects in the environment, such as a physical table 729, through transparent lens(es) and waveguide(s) of the AR device 728. Alternatively, an MR device could be used in place of the AR device 728 and a similar user experience can take place, but the user would not be directly viewing physical objects in the environment, such as table 729, and would instead be presented with a virtual reconstruction of the table 729 produced from one or more sensors of the MR device (e.g., an outward facing camera capable of recording the surrounding environment).
[00180] The user 702 can use any of the wrist-wearable device 726, the AR device 728 (e.g., through physical inputs at the AR device and/or built-in motion tracking of a user’s extremities), a smart-textile garment, externally mounted extremity tracking device, the HIPD 742 to provide user inputs, etc. For example, the user 702 can perform one or more hand gestures that are detected by the wrist-wearable device 726 (e.g., using one or more EMG sensors and/or IMUs built into the wrist-wearable device) and/or AR device 728 (e.g., using one or more image sensors or cameras) to provide a user input. Alternatively, or additionally, the user 702 can provide a user input via one or more touch surfaces of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742, and/or voice commands captured by a microphone of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742. The wrist-wearable device 726, the AR device 728, and/or the HIPD 742 include an artificially intelligent digital assistant to help the user in providing a user input (e.g., completing a sequence of operations, suggesting different operations or commands, providing reminders, confirming a command). For example, the digital assistant can be invoked through an input occurring at the AR device 728 (e.g., via an input at a temple arm of the AR device 728). In some examples, the user 702 can provide a user input via one or more facial gestures and/or facial expressions. For example, cameras of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 can track the user 702’s eyes for navigating a user interface.
[00181] The wrist-wearable device 726, the AR device 728, and/or the HIPD 742 can operate alone or in conjunction to allow the user 702 to interact with the AR environment. In some examples, the HIPD 742 is configured to operate as a central hub or control center for the wrist-wearable device 726, the AR device 728, and/or another communicatively coupled device. For example, the user 702 can provide an input to interact with the AR environment at any of the wrist-wearable device 726, the AR device 728, and/or the HIPD 742, and the HIPD 742 can identify one or more back-end and front-end tasks to cause the performance of the requested interaction and distribute instructions to cause the performance of the one or more back-end and front-end tasks at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742. In some examples, a back-end task is a background-processing task that is not perceptible by the user (e.g., rendering content, decompression, compression, application-specific operations), and a front-end task is a user-facing task that is perceptible to the user (e.g., presenting information to the user, providing feedback to the user). The HIPD 742 can perform the back-end tasks and provide the wrist-wearable device 726 and/or the AR device 728 operational data corresponding to the performed back-end tasks such that the wrist-wearable device 726 and/or the AR device 728 can perform the front-end tasks. In this way, the HIPD 742, which has more computational resources and greater thermal headroom than the wrist-wearable device 726 and/or the AR device 728, performs computationally intensive tasks and reduces the computer resource utilization and/or power usage of the wrist-wearable device 726 and/or the AR device 728. [00182] In the example shown by the first AR system 700a, the HIPD 742 identifies one or more back-end tasks and front-end tasks associated with a user request to initiate an AR video call with one or more other users (represented by the avatar 704 and the digital representation of the contact 706) and distributes instructions to cause the performance of the one or more back-end tasks and front-end tasks. In particular, the HIPD 742 performs back- end tasks for processing and/or rendering image data (and other data) associated with the AR video call and provides operational data associated with the performed back-end tasks to the AR device 728 such that the AR device 728 performs front-end tasks for presenting the AR video call (e.g., presenting the avatar 704 and the digital representation of the contact 706).
[00183] In some examples, the HIPD 742 can operate as a focal or anchor point for causing the presentation of information. This allows the user 702 to be generally aware of where information is presented. For example, as shown in the first AR system 700a, the avatar 704 and the digital representation of the contact 706 are presented above the HIPD 742. In particular, the HIPD 742 and the AR device 728 operate in conjunction to determine a location for presenting the avatar 704 and the digital representation of the contact 706. In some examples, information can be presented within a predetermined distance from the HIPD 742 (e.g., within five meters). For example, as shown in the first AR system 700a, virtual object 708 is presented on the desk some distance from the HIPD 742. Similarto the above example, the HIPD 742 and the AR device 728 can operate in conjunction to determine a location for presenting the virtual object 708. Alternatively, In some examples, presentation of information is not bound by the HIPD 742. More specifically, the avatar 704, the digital representation of the contact 706, and the virtual object 708 do not have to be presented within a predetermined distance of the HIPD 742. While an AR device 728 is described working with an HIPD, an MR headset can be interacted with in the same way as the AR device 728.
[00184] User inputs provided at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 are coordinated such that the user can use any device to initiate, continue, and/or complete an operation. For example, the user 702 can provide a user input to the AR device 728 to cause the AR device 728 to present the virtual object 708 and, while the virtual object 708 is presented by the AR device 728, the user 702 can provide one or more hand gestures via the wrist-wearable device 726 to interact and/or manipulate the virtual object 708. While an AR device 728 is described working with a wrist-wearable device 726, an MR headset can be interacted with in the same way as the AR device 728.
[00185] Figure 7A illustrates an interaction in which an artificially intelligent virtual assistant can assist in requests made by a user 702. The Al virtual assistant can be used to complete open-ended requests made through natural language inputs by a user 702. For example, in Figure 7A the user 702 makes an audible request 744 to summarize the conversation and then share the summarized conversation with others in the meeting. In addition, the Al virtual assistant is configured to use sensors of the XR system (e.g., cameras of an XR headset, microphones, and various other sensors of any of the devices in the system) to provide contextual prompts to the user for initiating tasks.
[00186] Figure 7A also illustrates an example neural network 752 used in Artificial Intelligence applications. Uses of Artificial Intelligence (Al) are varied and encompass many different aspects of the devices and systems described herein. Al capabilities cover a diverse range of applications and deepen interactions between the user 702 and user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726). The Al discussed herein can be derived using many different training techniques. While the primary Al model example discussed herein is a neural network, other Al models can be used. Nonlimiting examples of Al models include artificial neural networks (ANNs), deep neural networks (DNNs), convolution neural networks (CNNs), recurrent neural networks (RNNs), large language models (LLMs), long short-term memory networks, transformer models, decision trees, random forests, support vector machines, k-nearest neighbors, genetic algorithms, Markov models, Bayesian networks, fuzzy logic systems, and deep reinforcement learnings, etc. The Al models can be implemented at one or more of the user devices, and/or any other devices described herein. For devices and systems herein that employ multiple Al models, different models can be used depending on the task. For example, for a natural-language artificially intelligent virtual assistant, an LLM can be used and for the object detection of a physical environment, a DNN can be used instead.
[00187] In another example, an Al virtual assistant can include many different Al models and based on the user’s request, multiple Al models may be employed (concurrently, sequentially or a combination thereof). For example, an LLM-based Al model can provide instructions for helping a user follow a recipe and the instructions can be based in part on another Al model that is derived from an ANN, a DNN, an RNN, etc. that is capable of discerning what part of the recipe the user is on (e.g., object and scene detection).
[00188] As Al training models evolve, the operations and experiences described herein could potentially be performed with different models other than those listed above, and a person skilled in the art would understand that the list above is non-limiting.
[00189] A user 702 can interact with an Al model through natural language inputs captured by a voice sensor, text inputs, or any other input modality that accepts natural language and/or a corresponding voice sensor module. In another instance, input is provided by tracking the eye gaze of a user 702 via a gaze tracker module. Additionally, the Al model can also receive inputs beyond those supplied by a user 702. For example, the Al can generate its response further based on environmental inputs (e.g., temperature data, image data, video data, ambient light data, audio data, GPS location data, inertial measurement (i.e., user motion) data, pattern recognition data, magnetometer data, depth data, pressure data, force data, neuromuscular data, heart rate data, temperature data, sleep data) captured in response to a user request by various types of sensors and/or their corresponding sensor modules. The sensors’ data can be retrieved entirely from a single device (e.g., AR device 728) or from multiple devices that are in communication with each other (e.g., a system that includes at least two of an AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726, etc.). The Al model can also access additional information (e.g., one or more servers 730, the computers 740, the mobile devices 750, and/or other electronic devices) via a network 725.
[00190] A non-limiting list of Al-enhanced functions includes but is not limited to image recognition, speech recognition (e.g., automatic speech recognition), text recognition (e.g., scene text recognition), pattern recognition, natural language processing and understanding, classification, regression, clustering, anomaly detection, sequence generation, content generation, and optimization. In some examples, Al-enhanced functions are fully or partially executed on cloud-computing platforms communicatively coupled to the user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726) via the one or more networks. The cloud-computing platforms provide scalable computing resources, distributed computing, managed Al services, interference acceleration, pre-trained models, APIs and/or other resources to support comprehensive computations required by the Al- enhanced function.
[00191] Example outputs stemming from the use of an Al model can include natural language responses, mathematical calculations, charts displaying information, audio, images, videos, texts, summaries of meetings, predictive operations based on environmental factors, classifications, pattern recognitions, recommendations, assessments, or other operations. In some examples, the generated outputs are stored on local memories of the user devices (e.g., the AR device 728, an MR device 732, the HIPD 742, the wrist-wearable device 726), storage options of the external devices (servers, computers, mobile devices, etc.), and/or storage options of the cloud-computing platforms.
[00192] The Al-based outputs can be presented across different modalities (e.g., audiobased, visual-based, haptic-based, and any combination thereof) and across different devices of the XR system described herein. Some visual-based outputs can include the displaying of information on XR augments of an XR headset, user interfaces displayed at a wrist-wearable device, laptop device, mobile device, etc. On devices with or without displays (e.g., HIPD 742), haptic feedback can provide information to the user 702. An Al model can also use the inputs described above to determine the appropriate modality and device(s) to present content to the user (e.g., a user walking on a busy road can be presented with an audio output instead of a visual output to avoid distracting the user 702). [00193] Figure 7B shows the user 702 wearing the wrist-wearable device 726 and the AR device 728 and holding the HIPD 742. In the second AR system 700b, the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 are used to receive and/or provide one or more messages to a contact of the user 702. In particular, the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 detect and coordinate one or more user inputs to initiate a messaging application and prepare a response to a received message via the messaging application.
[00194] In some examples, the user 702 initiates, via a user input, an application on the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 that causes the application to initiate on at least one device. For example, in the second AR system 700b the user 702 performs a hand gesture associated with a command for initiating a messaging application (represented by messaging user interface 712); the wrist-wearable device 726 detects the hand gesture; and, based on a determination that the user 702 is wearing the AR device 728, causes the AR device 728 to present a messaging user interface 712 of the messaging application. The AR device 728 can present the messaging user interface 712 to the user 702 via its display (e.g., as shown by user 702’s field of view 710). In some examples, the application is initiated and can be run on the device (e.g., the wrist-wearable device 726, the AR device 728, and/or the HIPD 742) that detects the user input to initiate the application, and the device provides another device operational data to cause the presentation of the messaging application. For example, the wrist-wearable device 726 can detect the user input to initiate a messaging application, initiate and run the messaging application, and provide operational data to the AR device 728 and/or the HIPD 742 to cause presentation of the messaging application. Alternatively, the application can be initiated and run at a device other than the device that detected the user input. For example, the wrist-wearable device 726 can detect the hand gesture associated with initiating the messaging application and cause the HIPD 742 to run the messaging application and coordinate the presentation of the messaging application.
[00195] Further, the user 702 can provide a user input provided at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to continue and/or complete an operation initiated at another device. For example, after initiating the messaging application via the wrist-wearable device 726 and while the AR device 728 presents the messaging user interface 712, the user 702 can provide an input at the HIPD 742 to prepare a response (e.g., shown by the swipe gesture performed on the HIPD 742). The user 702’s gestures performed on the HIPD 742 can be provided and/or displayed on another device. For example, the user 702’s swipe gestures performed on the HIPD 742 are displayed on a virtual keyboard of the messaging user interface 712 displayed by the AR device 728. [00196] In some examples, the wrist-wearable device 726, the AR device 728, the HIPD 742, and/or other communicatively coupled devices can present one or more notifications to the user 702. The notification can be an indication of a new message, an incoming call, an application update, a status update, etc. The user 702 can select the notification via the wristwearable device 726, the AR device 728, or the HIPD 742 and cause presentation of an application or operation associated with the notification on at least one device. For example, the user 702 can receive a notification that a message was received at the wrist-wearable device 726, the AR device 728, the HIPD 742, and/or other communicatively coupled device and provide a user input at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to review the notification, and the device detecting the user input can cause an application associated with the notification to be initiated and/or presented at the wrist-wearable device 726, the AR device 728, and/or the HIPD 742.
[00197] While the above example describes coordinated inputs used to interact with a messaging application, the skilled artisan will appreciate upon reading the descriptions that user inputs can be coordinated to interact with any number of applications including, but not limited to, gaming applications, social media applications, camera applications, web-based applications, financial applications, etc. For example, the AR device 728 can present to the user 702 game application data and the HIPD 742 can use a controller to provide inputs to the game. Similarly, the user 702 can use the wrist-wearable device 726 to initiate a camera of the AR device 728, and the user can use the wrist-wearable device 726, the AR device 728, and/or the HIPD 742 to manipulate the image capture (e.g., zoom in or out, apply filters) and capture image data.
[00198] While an AR device 728 is shown being capable of certain functions, it is understood that an AR device can be an AR device with varying functionalities based on costs and market demands. For example, an AR device may include a single output modality such as an audio output modality. In another example, the AR device may include a low-fidelity display as one of the output modalities, where simple information (e.g., text and/or low-fidelity images/video) is capable of being presented to the user. In yet another example, the AR device can be configured with face-facing light emitting diodes (LEDs) configured to provide a user with information, e.g., an LED around the right-side lens can illuminate to notify the wearerto turn right while directions are being provided or an LED on the left-side can illuminate to notify the wearer to turn left while directions are being provided. In another example, the AR device can include an outward-facing projector such that information (e.g., text information, media) may be displayed on the palm of a user’s hand or other suitable surface (e.g., a table, whiteboard). In yet another example, information may also be provided by locally dimming portions of a lens to emphasize portions of the environment in which the user’s attention should be directed. Some AR devices can present AR augments either monocularly or binocularly (e.g., an AR augment can be presented at only a single display associated with a single lens as opposed presenting an AR augmented at both lenses to produce a binocular image). In some instances an AR device capable of presenting AR augments binocularly can optionally display AR augments monocularly as well (e.g., for power-saving purposes or other presentation considerations). These examples are non-exhaustive and features of one AR device described above can be combined with features of another AR device described above. While features and experiences of an AR device have been described generally in the preceding sections, it is understood that the described functionalities and experiences can be applied in a similar manner to an MR headset, which is described below in the proceeding sections.
[00199] Turning to Figures 7C-1 and 7C-2, the user 702 is shown wearing the wristwearable device 726 and an MR device 732 (e.g., a device capable of providing either an entirely VR experience or an MR experience that displays object(s) from a physical environment at a display of the device) and holding the HIPD 742. In the third AR system 700c, the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 are used to interact within an MR environment, such as a VR game or other MR/VR application. While the MR device 732 presents a representation of a VR game (e.g., first MR game environment 720) to the user 702, the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 detect and coordinate one or more user inputs to allow the user 702 to interact with the VR game.
[00200] In some examples, the user 702 can provide a user input via the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 that causes an action in a corresponding MR environment. For example, the user 702 in the third MR system 700c (shown in Figure 7C-1) raises the HIPD 742 to prepare for a swing in the first MR game environment 720. The MR device 732, responsive to the user 702 raising the HIPD 742, causes the MR representation of the user 722 to perform a similar action (e.g., raise a virtual object, such as a virtual sword 724). In some examples, each device uses respective sensor data and/or image data to detect the user input and provide an accurate representation of the user 702’s motion. For example, image sensors (e.g., SLAM cameras or other cameras) of the HIPD 742 can be used to detect a position of the HIPD 742 relative to the user 702’s body such that the virtual object can be positioned appropriately within the first MR game environment 720; sensor data from the wrist-wearable device 726 can be used to detect a velocity at which the user 702 raises the HIPD 742 such that the MR representation of the user 722 and the virtual sword 724 are synchronized with the user 702’s movements; and image sensors of the MR device 732 can be used to represent the user 702’s body, boundary conditions, or real-world objects within the first MR game environment 720. [00201] In Figure 7C-2, the user 702 performs a downward swing while holding the HIPD 742. The user 702’s downward swing is detected by the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 and a corresponding action is performed in the first MR game environment 720. In some examples, the data captured by each device is used to improve the user’s experience within the MR environment. For example, sensor data of the wrist-wearable device 726 can be used to determine a speed and/or force at which the downward swing is performed and image sensors of the HIPD 742 and/or the MR device 732 can be used to determine a location of the swing and how it should be represented in the first MR game environment 720, which, in turn, can be used as inputs for the MR environment (e.g., game mechanics, which can use detected speed, force, locations, and/or aspects of the user 702’s actions to classify a user’s inputs (e.g., user performs a light strike, hard strike, critical strike, glancing strike, miss) or calculate an output (e.g., amount of damage)).
[00202] Figure 7C-2 further illustrates that a portion of the physical environment is reconstructed and displayed at a display of the MR device 732 while the MR game environment 720 is being displayed. In this instance, a reconstruction of the physical environment 746 is displayed in place of a portion of the MR game environment 720 when object(s) in the physical environment are potentially in the path of the user (e.g., a collision with the user and an object in the physical environment are likely). Thus, this example MR game environment 720 includes (i) an immersive VR portion 748 (e.g., an environment that does not have a corollary counterpart in a nearby physical environment) and (ii) a reconstruction of the physical environment 746 (e.g., table 750 and cup 752). While the example shown here is an MR environment that shows a reconstruction of the physical environment to avoid collisions, other uses of reconstructions of the physical environment can be used, such as defining features of the virtual environment based on the surrounding physical environment (e.g., a virtual column can be placed based on an object in the surrounding physical environment (e.g., a tree)).
[00203] While the wrist-wearable device 726, the MR device 732, and/or the HIPD 742 are described as detecting user inputs, In some examples, user inputs are detected at a single device (with the single device being responsible for distributing signals to the other devices for performing the user input). For example, the HIPD 742 can operate an application for generating the first MR game environment 720 and provide the MR device 732 with corresponding data for causing the presentation of the first MR game environment 720, as well as detect the user 702’s movements (while holding the HIPD 742) to cause the performance of corresponding actions within the first MR game environment 720. Additionally or alternatively, In some examples, operational data (e.g., sensor data, image data, application data, device data, and/or other data) of one or more devices is provided to a single device (e.g., the HIPD 742) to process the operational data and cause respective devices to perform an action associated with processed operational data.
[00204] In some examples, the user 702 can wear a wrist-wearable device 726, wear an MR device 732, wear smart textile-based garments 738 (e.g., wearable haptic gloves), and/or hold an HIPD 742 device. In this example, the wrist-wearable device 726, the MR device 732, and/or the smart textile-based garments 738 are used to interact within an MR environment (e.g., any AR or MR system described above in reference to Figures 7A-7B). While the MR device 732 presents a representation of an MR game (e.g., second MR game environment 720) to the user 702, the wrist-wearable device 726, the MR device 732, and/or the smart textile-based garments 738 detect and coordinate one or more user inputs to allow the user 702 to interact with the MR environment.
[00205] In some examples, the user 702 can provide a user input via the wrist-wearable device 726, an HIPD 742, the MR device 732, and/or the smart textile-based garments 738 that causes an action in a corresponding MR environment. In some examples, each device uses respective sensor data and/or image data to detect the user input and provide an accurate representation of the user 702’s motion. While four different input devices are shown (e.g., a wrist-wearable device 726, an MR device 732, an HIPD 742, and a smart textile-based garment 738) each one of these input devices entirely on its own can provide inputs for fully interacting with the MR environment. For example, the wrist-wearable device can provide sufficient inputs on its own for interacting with the MR environment. In some examples, if multiple input devices are used (e.g., a wrist-wearable device and the smart textile-based garment 738) sensor fusion can be utilized to ensure inputs are correct. While multiple input devices are described, it is understood that other input devices can be used in conjunction or on their own instead, such as but not limited to external motion-tracking cameras, other wearable devices fitted to different parts of a user, apparatuses that allow for a user to experience walking in an MR environment while remaining substantially stationary in the physical environment, etc.
[00206] As described above, the data captured by each device is used to improve the user’s experience within the MR environment. Although not shown, the smart textile-based garments 738 can be used in conjunction with an MR device and/or an HIPD 742.
[00207] While some experiences are described as occurring on an AR device and other experiences are described as occurring on an MR device, one skilled in the art would appreciate that experiences can be ported over from an MR device to an AR device, and vice versa.
[00208] Some definitions of devices and components that can be included in some or all of the example devices discussed are defined here for ease of reference. A skilled artisan will appreciate that certain types of the components described may be more suitable for a particular set of devices, and less suitable for a different set of devices. But subsequent reference to the components defined here should be considered to be encompassed by the definitions provided.
[00209] Example devices and systems, including electronic devices and systems, will be discussed. Such example devices and systems are not intended to be limiting, and one of skill in the art will understand that alternative devices and systems to the example devices and systems described herein may be used to perform the operations and construct the systems and devices that are described herein.
[00210] As described herein, an electronic device is a device that uses electrical energy to perform a specific function. It can be any physical object that contains electronic components such as transistors, resistors, capacitors, diodes, and integrated circuits. Examples of electronic devices include smartphones, laptops, digital cameras, televisions, gaming consoles, and music players, as well as the example electronic devices discussed herein. As described herein, an intermediary electronic device is a device that sits between two other electronic devices, and/or a subset of components of one or more electronic devices and facilitates communication, and/or data processing and/or data transfer between the respective electronic devices and/or electronic components.
[00211] The foregoing descriptions of Figures 7A-7C-2 provided above are intended to augment the description provided in reference to Figures 1A-6. While terms in the following description may not be identical to terms used in the foregoing description, a person having ordinary skill in the art would understand these terms to have the same meaning.
[00212] Any data collection performed by the devices described herein and/or any devices configured to perform or cause the performance of the different examples described above in reference to any of the Figures, hereinafter the “devices,” is done with user consent and in a manner that is consistent with all applicable privacy laws. Users are given options to allow the devices to collect data, as well as the option to limit or deny collection of data by the devices. A user is able to opt in or opt out of any data collection at any time. Further, users are given the option to request the removal of any collected data.
[00213] It will be understood that, although the terms “first,” “second,” etc. may be used herein to describe various elements, these elements should not be limited by these terms. These terms are only used to distinguish one element from another.
[00214] The terminology used herein is for the purpose of describing particular examples only and is not intended to be limiting of the claims. As used in the description of the examples and the appended claims, the singular forms “a,” “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will also be understood that the term “and/or” as used herein refers to and encompasses any and all possible combinations of one or more of the associated listed items. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
[00215] As used herein, the term “if’ can be construed to mean “when” or “upon” or “in response to determining” or “in accordance with a determination” or “in response to detecting,” that a stated condition precedent is true, depending on the context. Similarly, the phrase “if it is determined [that a stated condition precedent is true]” or “if [a stated condition precedent is true]” or “when [a stated condition precedent is true]” can be construed to mean “upon determining” or “in response to determining” or “in accordance with a determination” or “upon detecting” or “in response to detecting” that the stated condition precedent is true, depending on the context.
[00216] The foregoing description, for purpose of explanation, has been described with reference to specific examples. However, the illustrative discussions above are not intended to be exhaustive or to limit the claims to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The examples were chosen and described in order to best explain principles of operation and practical applications, to thereby enable others skilled in the art.

Claims

Claims
1. A method, comprising: in response to an indication that an artificial intelligence (Al) agent trigger condition is present, providing an Al agent sensor data obtained by a wearable device; determining, by the Al agent, a context-based activity based on the sensor data obtained by the wearable device; generating, by the Al agent, orchestrated guidance based on the context-based activity, wherein the orchestrated guidance includes a recommended action for performing the context-based activity; and presenting the orchestrated guidance at the wearable device.
2. The method of claim 1 , wherein: the context-based activity is a first context-based activity; the sensor data is first sensor data; the orchestrated guidance is first orchestrated guidance; the recommended action is a first recommended action; and the method further comprises: in accordance with a determination that the first recommended action for performing the first context-based activity was performed, providing the Al agent second sensor data obtained by the wearable device, determining, by the Al agent, a second context-based activity based on the second sensor data obtained by the wearable device, generating, by the Al agent, second orchestrated guidance based on the second context-based activity, wherein the second orchestrated guidance includes a second recommended action for performing the second context-based activity, and presenting the second orchestrated guidance at the wearable device.
3. The method of claim 1 or 2, wherein: the context-based activity is a first context-based activity of a plurality of contextbased activities determined by the by the Al agent based on the sensor data; the orchestrated guidance includes a plurality of recommended actions for performing the plurality of context-based activities; and the recommended action is a first recommended action of the plurality of recommended actions, the first recommended action being configured to perform the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions.
4. The method of claim 3, wherein: generating the orchestrated guidance includes determining a subset of the plurality of recommended actions for performing the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action of the plurality of recommended actions and the subset of the plurality of recommended actions for performing the first context-based activity.
5. The method of claim 3 or 4, wherein: generating the orchestrated guidance includes determining a sequence of contextbased activities of the plurality of context-based activities to be performed, including a second context-based activity to follow the first context-based activity; and presenting the orchestrated guidance at the wearable device includes presenting at least the first recommended action and the second recommended action of the plurality of recommended actions for performing the plurality of context-based activities.
6. The method of any preceding claim, further comprising: in response to a user input selecting the recommended action for performing the context-based activity, causing the wearable device to initiate a do-not-disturb mode, wherein, while in the do-not-disturb mode, the wearable device suppresses, at least, received notifications; and in response to an indication that participation in the context-based activity ceased: causing the wearable device to cease the do-not-disturb mode, generating, by the Al agent, a notification summary based on the notifications received while the wearable device was in the do-not-disturb mode, and presenting the notification summary at the wearable device.
7. The method of any preceding claim, further comprising: in response to a user input selecting the recommended action for performing the context-based activity, performing, by the Al agent, a search based on the recommended action; determining a task to perform based on the search; and presenting the task at the wearable device.
8. The method of any preceding claim, wherein presenting the orchestrated guidance at the wearable device includes, at least one of: causing presentation of a user interface element associated with the orchestrated guidance at a communicatively coupled display, and causing presentation of audible guidance associated with the orchestrated guidance at a communicatively coupled speaker.
9. The method of any preceding claim, wherein the context-based activity is to be performed at a physical activity.
10. Awearable device, comprising: a display; one or more sensors; and one or more programs, wherein the one or more programs are stored in memory and configured to be executed by one or more processors to cause the one or more processors to carry out the method of any preceding claim.
11. A computer-readable storage medium including executable instructions that, when executed by one or more processors, cause the one or more processors to carry out the method of any of claims 1 to 9.
12. A computer program product comprising instructions that, when executed by one or more processors of a computer, cause the one or more processors to carry out the method of any of claims 1 to 9.
PCT/US2025/030023 2024-05-17 2025-05-19 Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers Pending WO2025240965A1 (en)

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US202463649289P 2024-05-17 2024-05-17
US63/649,289 2024-05-17
US19/209,723 US20250355916A1 (en) 2024-05-17 2025-05-15 Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers
US19/209,723 2025-05-15

Publications (1)

Publication Number Publication Date
WO2025240965A1 true WO2025240965A1 (en) 2025-11-20

Family

ID=96020029

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2025/030023 Pending WO2025240965A1 (en) 2024-05-17 2025-05-19 Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers

Country Status (1)

Country Link
WO (1) WO2025240965A1 (en)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150213371A1 (en) * 2012-08-14 2015-07-30 Sri International Method, system and device for inferring a mobile user's current context and proactively providing assistance
DK202070658A1 (en) * 2020-06-01 2022-01-13 Apple Inc Suggesting executable actions in response to detecting events
US20220272055A1 (en) * 2021-02-25 2022-08-25 Google Llc Inferring assistant action(s) based on ambient sensing by assistant device(s)

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20150213371A1 (en) * 2012-08-14 2015-07-30 Sri International Method, system and device for inferring a mobile user's current context and proactively providing assistance
DK202070658A1 (en) * 2020-06-01 2022-01-13 Apple Inc Suggesting executable actions in response to detecting events
US20220272055A1 (en) * 2021-02-25 2022-08-25 Google Llc Inferring assistant action(s) based on ambient sensing by assistant device(s)

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
ANONYMOUS: "Use Driving Focus on your iPhone to concentrate on the road - Apple Support (UK)", 2 May 2024 (2024-05-02), pages 1 - 6, XP093303543, Retrieved from the Internet <URL:https://web.archive.org/web/20240502214622/https://support.apple.com/en-gb/108384> *
ANONYMOUS: "View and respond to notifications on iPhone - Apple Support (UK)", 19 September 2023 (2023-09-19), pages 1 - 3, XP093303545, Retrieved from the Internet <URL:https://web.archive.org/web/20230919150538/https://support.apple.com/en-gb/guide/iphone/iph6534c01bc/16.0/ios/16.0> *

Similar Documents

Publication Publication Date Title
US20240085978A1 (en) Enhancing physical notebooks in extended reality
US9563272B2 (en) Gaze assisted object recognition
US20170206064A1 (en) Persistent companion device configuration and deployment platform
US20160151917A1 (en) Multi-segment social robot
KR20180129886A (en) Persistent companion device configuration and deployment platform
HK1217847A1 (en) Apparatus and methods for providing a persistent companion device
WO2016011159A1 (en) Apparatus and methods for providing a persistent companion device
US12411542B2 (en) Electronic devices using object recognition and/or voice recognition to provide personal and health assistance to users
US12267482B2 (en) Controlling and editing presentation of volumetric content
US20230403460A1 (en) Techniques for using sensor data to monitor image-capture trigger conditions for determining when to capture images using an imaging device of a head- wearable device, and wearable devices and systems for performing those techniques
US20250131609A1 (en) Generating image scenarios based on events
US20250252700A1 (en) Wearable Device Including An Artificially Intelligent Assistant For Generating Responses Based On Shared Contextual Data, And Systems And Methods Of Use Thereof
US20250166665A1 (en) Systems and methods for automated digital editing
US20250060975A1 (en) Intermediary device for offloading processing operations for an artificial-reality system and for enabling interactions and input paradigms, and systems and methods of use thereof
US20250355916A1 (en) Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers
US20250245885A1 (en) Systems and methods for generating and distributing instant avatar stickers
WO2025240965A1 (en) Wearable devices including artificially intelligent systems for generating and presenting guidance to wearers
US20250298642A1 (en) Command recommendation system and user interface element generator, and methods of use thereof
WO2018183812A1 (en) Persistent companion device configuration and deployment platform
US20250390523A1 (en) Gaze-activated information retrieval and systems and methods of use thereof
US20250265788A1 (en) Techniques For Determining Tasks Based On Data From A Sensor Set Of An Extended-Reality System Using A Sensor Set Agnostic Contrastively-Trained Learning Model, And Systems And Methods Of Use Thereof
US20250278902A1 (en) Techniques for coordinating artificial-reality interactions using augmented-reality interaction guides for performing interactions with physical objects within a user&#39;s physical surroundings, and systems and methods for using such techniques
US20250117119A1 (en) Language augmented video editing, and systems and methods of use thereof
US20250390198A1 (en) Techniques for presenting augmented-reality content in multiple presentation modes based on real-world context, and devices, systems, and methods thereof
WO2025169014A2 (en) Wearable device including an artificially intelligent assistant for generating responses based on shared contextual data, and systems and methods of use thereof