WO2024064373A1 - Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments - Google Patents
Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments Download PDFInfo
- Publication number
- WO2024064373A1 WO2024064373A1 PCT/US2023/033530 US2023033530W WO2024064373A1 WO 2024064373 A1 WO2024064373 A1 WO 2024064373A1 US 2023033530 W US2023033530 W US 2023033530W WO 2024064373 A1 WO2024064373 A1 WO 2024064373A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- user interface
- displaying
- user
- detecting
- input
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/163—Wearable computers, e.g. on a belt
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1686—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being an integrated camera
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F1/00—Details not covered by groups G06F3/00 - G06F13/00 and G06F21/00
- G06F1/16—Constructional details or arrangements
- G06F1/1613—Constructional details or arrangements for portable computers
- G06F1/1633—Constructional details or arrangements of portable computers not specific to the type of enclosures covered by groups G06F1/1615 - G06F1/1626
- G06F1/1684—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675
- G06F1/1694—Constructional details or arrangements related to integrated I/O peripherals not covered by groups G06F1/1635 - G06F1/1675 the I/O peripheral being a single or a set of motion sensors for pointer control or gesture input obtained by sensing movements of the portable computer
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04812—Interaction techniques based on cursor appearance or behaviour, e.g. being affected by the presence of displayed objects
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04845—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range for image manipulation, e.g. dragging, rotation, expansion or change of colour
Definitions
- the present disclosure relates generally to computer systems that are in communication with a display generation component and, optionally, one or more input devices that provide computer-generated experiences, including, but not limited to, electronic devices that provide virtual reality and mixed reality experiences via a display.
- Example augmented reality environments include at least some virtual elements that replace or augment the physical world.
- Input devices such as cameras, controllers, joysticks, touch-sensitive surfaces, and touch-screen displays for computer systems and other electronic computing devices are used to interact with virtual/augmented reality environments.
- virtual elements include virtual objects, such as digital images, video, text, icons, and control elements such as buttons and other graphics.
- Some methods and interfaces for interacting with environments that include at least some virtual elements are cumbersome, inefficient, and limited.
- environments that include at least some virtual elements e.g., applications, augmented reality environments, mixed reality environments, and virtual reality environments
- systems that provide insufficient feedback for performing actions associated with virtual objects systems that require a series of inputs to achieve a desired outcome in an augmented reality environment, and systems in which manipulation of virtual objects are complex, tedious, and error-prone, create a significant cognitive burden on a user, and detract from the experience with the virtual/augmented reality environment.
- these methods take longer than necessary, thereby wasting energy of the computer system. This latter consideration is particularly important in battery-operated devices.
- the computer system is a desktop computer with an associated display.
- the computer system is a portable device (e.g., a notebook computer, tablet computer, or handheld device).
- the computer system is a personal electronic device (e.g., a wearable electronic device, such as a watch, or a head-mounted device).
- the computer system has a touchpad.
- the computer system has one or more cameras.
- the computer system has a touch- sensitive display (also known as a “touch screen” or “touch-screen display”).
- the computer system has one or more eye-tracking components. In some embodiments, the computer system has one or more hand-tracking components. In some embodiments, the computer system has one or more output devices in addition to the display generation component, the output devices including one or more tactile output generators and/or one or more audio output devices. In some embodiments, the computer system has a graphical user interface (GUI), one or more processors, memory and one or more modules, programs or sets of instructions stored in the memory for performing multiple functions.
- GUI graphical user interface
- the user interacts with the GUI through a stylus and/or finger contacts and gestures on the touch-sensitive surface, movement of the user’ s eyes and hand in space relative to the GUI (and/or computer system) or the user’s body as captured by cameras and other movement sensors, and/or voice inputs as captured by one or more audio input devices.
- the functions performed through the interactions optionally include image editing, drawing, presenting, word processing, spreadsheet making, game playing, telephoning, video conferencing, e-mailing, instant messaging, workout support, digital photographing, digital videoing, web browsing, digital music playing, note taking, and/or digital video playing. Executable instructions for performing these functions are, optionally, included in a transitory and/or non-transitory computer readable storage medium or other computer program product configured for execution by one or more processors.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying, via the first display generation component, a first object in a first view of a three- dimensional environment, wherein the first object includes at least a first portion of the first object and a second portion of the first object.
- the method includes, while displaying the first object, detecting, via the one or more input devices, a first gaze input that meets first criteria, wherein the first criteria require that the first gaze input is directed to the first portion of the first object in order for the first criteria to be met.
- the method includes, in response to detecting that the first gaze input meets the first criteria, displaying a first control element that corresponds to a first operation associated with the first object, wherein the first control element was not displayed prior to detecting that the first gaze input met the first criteria.
- the method includes, while displaying the first control element, detecting, via the one or more input devices, a first user input directed to the first control element, and, in response to detecting the first user input directed to the first control element, performing the first operation with respect to the first object.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying, via the first display generation component, a first user interface object and a first control element that is associated with performing a first operation with respect to the first user interface object, in a first view of a three-dimensional environment, wherein the first control element is spaced apart from the first user interface object in the first view of the three-dimensional environment, and wherein the first control element is displayed with a first appearance.
- the method includes, while displaying the first control element with the first appearance, detecting, via the one or more input devices, a first gaze input that is directed to the first control element.
- the method includes, in response to detecting the first gaze input that is directed to the first control element, updating an appearance of the first control element from the first appearance to a second appearance that is different from the first appearance.
- the method includes, while displaying the first control element with the second appearance, detecting, via the one or more input devices, a first user input directed to the first control element.
- the method includes, in response to detecting the first user input directed to the first control element, and in accordance with a determination that the first user input meets first criteria, updating the appearance of the first control element from the second appearance to a third appearance that is different from the first appearance and the second appearance and that indicates that additional movement associated with the first user input will cause the first operation associated with the first control element to be performed.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes concurrently displaying, via the first display generation component, a first application window and a first title bar of the first application window, wherein the first title bar of the first application window is separate from the first application window on a first side of the first application window and displays a respective identifier of the first application window.
- the method includes, while displaying the first application window with the first title bar separate from the first application window, detecting, via the one or more input devices, that a user’s attention is directed to the first title bar.
- the method includes, in response to detecting that the user’s attention is directed to the first title bar, in accordance with a determination that the user’s attention meets first criteria with respect to the first title bar, expanding the first title bar to display one or more first selectable controls for interacting with a first application corresponding to the first application window, wherein the one or more first selectable controls were not displayed prior to expanding the first title bar.
- a method is performed at a computer system that is in communication with a first display generation component having a first display area and one or more input devices.
- the method includes displaying, via the first display generation component, a first application window of a first application at a first window position in the first display area.
- the method includes, in accordance with a determination that the first application is accessing one or more sensors of the computer system, displaying a first indicator at a first indicator position in the first display area with a first spatial relationship to the first application window as an indication that the first application is accessing the one or more sensors of the computer system.
- the method includes, while displaying the first indicator at the first indicator position in the first display area with the first spatial relationship to the first application window, detecting a first user input that corresponds to a request to move the first application window of the first application to a second window position in the first display area, the second window position being different from the first window position.
- the method includes, in response to detecting the first user input that corresponds to the request to move the first application window of the first application from the first window position to the second window position in the first display area, displaying the first application window of the first application at the second window position in the first display area; and in accordance with a determination that the first application is accessing the one or more sensors of the computer system, displaying the first indicator at a second indicator position in the first display area that is different from the first indicator position in the first display area, wherein the second indicator position in the first display area has the first spatial relationship to the first application window displayed at the second window position.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying a user interface, wherein displaying the user interface includes concurrently displaying a content region with first content, a first user interface object, and a second user interface object in the user interface, wherein: respective content in the content region is constrained to having an appearance in which a respective parameter is within a first range of values, the first user interface object is displayed with an appearance in which the respective parameter has a value that is outside of the first range of values, and the second user interface object is displayed with an appearance in which the respective parameter has a value that is outside of the first range of values.
- the method includes, while concurrently displaying the first content, the first user interface object, and the second user interface object, updating the user interface, including: changing the first content to second content while the respective content in the content region continues to be constrained to having an appearance in which the respective parameter is within the first range of values, updating the first user interface object in appearance and continuing to display the first user interface object with an appearance in which the respective parameter has a value that is outside of the first range of values, and updating the second user interface object in appearance and continuing to display the second user interface object with an appearance in which the respective parameter has a value that is outside of the first range of values.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying, via the first display generation component, a first view of a three-dimensional environment that corresponds to a first viewpoint of a user.
- the method further includes, while displaying, via the first display generation component, the first view of the three- dimensional environment that corresponds to the first viewpoint of the user, detecting a first event that corresponds to a request to display a first virtual object in the first view of the three-dimensional environment.
- the method further includes, in response to detecting the first event that corresponds to a request to display the first virtual object in the first view of the three-dimensional environment, displaying, in the first view of the three-dimensional environment, the first virtual object at a first position in the three-dimensional environment, wherein the first virtual object is displayed with a first object management user interface that corresponds to the first virtual object, and wherein the first object management user interface has a first appearance relative to the first virtual object at the first position in the three- dimensional environment.
- the method includes, detecting, via the one or more input devices, a first user input that corresponds to a request to move the first virtual object in the three- dimensional environment.
- the method further includes, in response to detecting the first user input that corresponds to a request to move the first virtual object in the three-dimensional environment: displaying, in the first view of the three-dimensional environment, the first virtual object at a second position, different from the first position, in the three-dimensional environment, wherein the first virtual object is displayed concurrently with the first object management user interface at the second position in the three-dimensional environment, and wherein the first object management user interface has a second appearance relative to the first virtual object, different from the first appearance.
- a method is performed at a computer system that is in communication with one or more display generation components and one or more input devices.
- the method includes, while displaying, via the one or more display generation components, a user interface of a first application concurrently with a close affordance associated with the user interface of the first application, detecting a first input directed to the close affordance.
- the method includes, in response to detecting the first input, in accordance with a determination that the first input is a first type of input, displaying a first option to close applications other than the first application.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying, via the first display generation component, a first object at a first position in a first view of a three-dimensional environment.
- the method further includes, while displaying, via the first display generation component, the first object at the first position in the first view of the three-dimensional environment, displaying a first set of one or more control objects, wherein a respective control object of the first set of one or more control objects corresponds to a respective operation applicable to the first object.
- the method includes, detecting, via the one or more input devices, a first user input that corresponds to a request to move the first object in the three-dimensional environment.
- the method further includes, in response to detecting the first user input that corresponds to a request to move the first object in the three- dimensional environment: moving the first object from the first position to a second position and, while moving the first object from the first position to the second position, visually deemphasizing relative to the first object, at least one of the first set of one or more control objects that corresponds to the respective operation that is applicable to the first object.
- a method is performed at a computer system that is in communication with a display generation component and one or more input devices.
- the method includes, while displaying, via the display generation component, a first application user interface at a first location in a three-dimensional environment, detecting, at a first time via the one or more input devices, a first input corresponding to a request to close the first application user interface.
- the method includes, in response to detecting the first input corresponding to a request to close the first application user interface: closing the first application user interface, including ceasing to display the first application user interface in the three-dimensional environment; and, in accordance with a determination that respective criteria are met, displaying a home menu user interface at a respective home menu position that is determined based on the first location of the first application user interface in the three-dimensional environment.
- a method is performed at a computer system that is in communication with a first display generation component and one or more input devices.
- the method includes displaying, via the display generation component, a first user interface object, wherein the first user interface object includes first content.
- the method includes, while displaying the first user interface object including the first content via the display generation component, detecting, via the one or more input devices, a first user input that is directed to the first user interface object.
- the method further includes, in response to detecting the first user input that is directed to the first user interface object: in accordance with a determination that the first user input corresponds to a request to resize the first user interface object, resizing the first user interface object in accordance with the first user input, wherein resizing the first user interface object in accordance with the first user input includes: one or more temporary resizing operations, including, in accordance with a determination that a characteristic refresh rate of the first content within the first user interface object is a first refresh rate, scaling the first user interface object with the first content by a first amount of scaling before the first content is updated within the first user interface object in accordance with a first updated size of the first user interface object that is specified by the first user input; and in accordance with a determination that the characteristic refresh rate of the first content within the first user interface object is a second refresh rate different from the first refresh rate, scaling the first user interface object with the first content by a second amount of scaling different from the first amount of scaling before the first content is updated within the
- the method includes, after the one or more temporary resizing operations, displaying the first user interface object at the first updated size that is specified by the first user input, and updating the first content within the first user interface object in accordance with the first updated size that is specified by the first user input.
- a method is performed at a first computer system that is in communication with one or more display generation components and one or more input devices.
- the method includes displaying, via the one or more display generation components, a first application window at a first scale.
- the method includes, while displaying the first application window at the first scale, detecting, via the one or more input devices, a first gesture directed toward the first application window.
- the method includes, in response to detecting the first gesture, in accordance with a determination that the first gesture is directed toward a respective portion of the first application window that is not associated with an application-specific response to the first gesture, changing a respective scale of the first application window from the first scale to a second scale that is different from the first scale.
- Figure 1 A is a block diagram illustrating an operating environment of a computer system for providing extended reality (XR) experiences in accordance with some embodiments.
- XR extended reality
- Figures IB- IP are examples of a computer system for providing XR experiences in the operating environment of Figure 1A.
- Figure 2 is a block diagram illustrating a controller of a computer system that is configured to manage and coordinate an XR experience for the user in accordance with some embodiments.
- Figure 3 is a block diagram illustrating a display generation component of a computer system that is configured to provide a visual component of the XR experience to the user in accordance with some embodiments.
- Figure 4 is a block diagram illustrating a hand tracking unit of a computer system that is configured to capture gesture inputs of the user in accordance with some embodiments.
- Figure 5 is a block diagram illustrating an eye tracking unit of a computer system that is configured to capture gaze inputs of the user in accordance with some embodiments.
- Figure 6 is a flow diagram illustrating a glint-assisted gaze tracking pipeline in accordance with some embodiments.
- Figures 7A-7R illustrate example techniques for conditionally displaying controls for an application window and updating visual properties of the controls in response to user interactions, in accordance with some embodiments.
- Figures 7S-7AD3 illustrate example techniques for displaying a title bar with information for an application, in accordance with some embodiments.
- Figures 7AE-7AH illustrate example techniques for displaying privacy indicators, in accordance with some embodiments.
- Figures 7AI-7AP illustrate example techniques for displaying some user interface elements with values for a parameter constrained to a first range, and displaying other user interface elements with values for the parameter that are outside the first range, in accordance with some embodiments.
- Figures 7AQ-7BB illustrate example techniques for displaying controls for a virtual object and automatically updating the displayed controls as the virtual object is moved within the AR/VR environment, in accordance with some embodiments.
- Figures 7BC-7BL illustrate example techniques for closing different applications or groups of applications in a three-dimensional environment in response to different inputs, in accordance with some embodiments.
- Figures 7BM1-7CH illustrate example techniques for displaying a home menu user interface after closing one or more application user interfaces, in accordance with some embodiments.
- Figure 8 is a flow diagram of methods of conditionally displaying controls for applications, in accordance with various embodiments.
- Figure 9 is a flow diagram of methods of updating visual properties of controls in response to user interactions, in accordance with various embodiments.
- Figure 10 is a flow diagram of methods of displaying a title bar with information for an application, in accordance with various embodiments.
- Figure 11 is a flow diagram of methods of displaying privacy indicators, in accordance with various embodiments.
- Figure 12 is a flow diagram of methods of displaying some user interface elements with values for a parameter constrained to a first range, and displaying other user interface elements with values for the parameter that are outside the first range, in accordance with various embodiments.
- Figure 13 is a flow diagram of methods of displaying controls for a virtual object and automatically updating the displayed controls as the virtual object is moved within the AR/VR environment, in accordance with various embodiments.
- Figure 14 is a flow diagram of methods of closing different applications or groups of applications in a three-dimensional environment in response to different inputs, in accordance with various embodiments.
- Figure 15 is a flow diagram of methods of displaying controls for an object and automatically reducing a prominence of the controls while the object is moved within the AR/VR environment, in accordance with various embodiments.
- Figure 16 is a flow diagram of methods of displaying a home menu user interface after closing one or more application user interfaces, in accordance with various embodiments.
- Figures 17A1-17T illustrate example techniques for scaling a user interface object, during a resize operation of the user interface object, by different scaling factors based on a refresh rate of content in the user interface object, in accordance with various embodiments.
- Figure 18 is a flow diagram of methods for scaling a user interface object, during a resize operation of the user interface object, by different scaling factors based on a refresh rate of content in the user interface object, in accordance with various embodiments.
- Figures 19A-19S illustrate example techniques for scaling a user interface object based on respective gestures, in accordance with various embodiments.
- Figure 20 is a flow diagram of methods for scaling a user interface object based on respective gestures, in accordance with various embodiments.
- the present disclosure relates to user interfaces for providing an extended reality (XR) experience to a user, in accordance with some embodiments.
- XR extended reality
- a computer system displays an application window.
- the computer system In response to detecting the user’s gaze is directed to a respective portion of the application window, the computer system displays a respective control associated with the respective portion of the application window.
- a control In response to detecting a user’s gaze directed to an area of the control, without requiring additional user input, enables the user to access a particular control to perform an operation by shifting the user’s gaze, without cluttering the user interface with display of all the available controls.
- a computer system displays a control for an application with a first appearance in response to the user gazing at the control.
- the computer system updates display of the control to be displayed with a second appearance after detecting the user is interacting with the control, such as performing a gesture, to perform an operation.
- Automatically updating an appearance of the control when the user is gazing at the control and further updating the appearance of the move control when the user is interacting with the control provides the user with improved visual feedback of the user’s interactions.
- a computer system displays an application window for a first application.
- the computer system concurrently displays, with the application window, a title bar for the application window.
- the title bar dynamically increases in size, and displays additional controls in the title bar, in response to detecting a user’s attention is directed to the title bar. Dynamically increasing the title bar in size to provide the user with access to additional controls, reduces a number of inputs required to access additional controls for the application window and provides visual feedback about a state of the device.
- a computer system displays an application window in a display area.
- the computer system determines if an application is accessing sensitive user data and displays a privacy indicator near an application window of the application that is accessing sensitive user data.
- the computer system detects a user input to move the application window in the display area. Continuing to provide a privacy indicator for an application window, even as the application window is repositioned to be displayed at a different location in the display area, improves security and privacy of the system by providing real-time information about the specific application window that is accessing the sensitive user data and maintaining the information as the application window moves within the display area.
- a computer system displays a user interface, wherein displaying the user interface includes concurrently displaying a content region with first content, a first user interface object, and a second user interface object in the user interface.
- the respective content in the content region is constrained to having an appearance in which a respective parameter is within a first range of values
- the first user interface object is displayed with an appearance in which the respective parameter has a value that is outside of the first range of values
- the second user interface object is displayed with an appearance in which the respective parameter has a value that is outside of the first range of values.
- Updating the user interface includes changing the first content to second content while the respective content in the content region continues to be constrained to having an appearance in which the respective parameter is within the first range of values; updating the first user interface object in appearance and continuing to display the first user interface object with an appearance in which the respective parameter has a value that is outside of the first range of values; and updating the second user interface object in appearance and continuing to display the second user interface object with an appearance in which the respective parameter has a value that is outside of the first range of values.
- a computer system in response to detecting a first event that corresponds to a request to display a first virtual object in a three-dimensional environment, displays the first virtual object at first position in the three-dimensional environment with a first object management user interface that corresponds to the first virtual object, and wherein the first object management user interface has a first appearance relative to the first virtual object at the first position.
- the computer system In response to detecting a first user input that corresponds to a request to move the first virtual object in the three-dimensional environment, the computer system displays, in the first view of the three-dimensional environment, the first virtual object at a second position concurrently with the first object management user interface at the second position in the three-dimensional environment, the first object management user interface having a second appearance relative to the first virtual object.
- a computer system displays a user interface of a first application concurrently with a close affordance associated with the user interface of the first application.
- the computer system displays a first option to close applications other than the first application. Providing different options for closing one or more user interfaces reduces the number of inputs needed to display one or more user interfaces of interest.
- a computer system while displaying, a first object at a first position in a first view of a three-dimensional environment, displays a first set of one or more control objects, wherein a respective control object of the first set of one or more control objects corresponds to a respective operation applicable to the first object.
- the computer system In response to detecting a first user input that corresponds to a request to move the first object in the three-dimensional environment, the computer system: moves the first object from the first position to a second position and, while moving the first object from the first position to the second position, visually deemphasizes relative to the first object, at least one of the first set of one or more control objects that corresponds to the respective operation that is applicable to the first object.
- Automatically updating a control for a virtual object in response to detecting that the object is moving within the AR/VR environment, without requiring additional user input, enables the user to continue to access the control to perform an operation, and provides improved visual feedback by indicating the controls are not available for interaction while the object is being moved.
- a computer system displays a first application user interface at a first location in a three-dimensional environment.
- the computer system closes the first application user interface and displays a home menu user interface at a respective home menu position that is determined based on the first location of the first application user interface in the three-dimensional environment.
- Automatically displaying the home menu user interface allows the user to continue navigating through one or more collections of selectable representations of the home menu user interface when no application user interface remains in a viewport of the three-dimensional environment, without displaying additional controls.
- a computer system in response to detecting a user input corresponding to a request to resize a first user interface object, resizes the first user interface object in accordance with the user input, including performing one or more temporary resizing operations, including: in accordance with a determination that a characteristic refresh rate of the first content is a first refresh rate, scaling the first user interface object with the first content by a first amount; and in accordance with a determination that the characteristic refresh rate of the first content is a second refresh rate, scaling the first user interface object with the first content by a second amount.
- the computer system displays the first user interface object at the first updated size that is specified by the first user input, and updates the first content within the first user interface object.
- a computer system displays a first application window at a first scale.
- the computer system while displaying the first application window at the first scale, detects a first gesture directed toward the first application window.
- the computer system changes a respective scale of the first application window from the first scale to a second scale that is different from the first scale.
- Figures 1 A-6 provide a description of example computer systems for providing XR experiences to users.
- Figures 7A-7R illustrate example techniques for conditionally displaying controls for an application window and updating visual properties of the controls in response to user interactions, in accordance with some embodiments.
- Figure 8 is a flow diagram of methods of conditionally displaying controls for an application window, in accordance with various embodiments.
- Figure 9 is a flow diagram of methods of updating visual properties of the controls in response to user interactions, in accordance with various embodiments.
- the user interfaces in Figures 7A-7R are used to illustrate the processes in Figures 8 and 9.
- Figures 7S-7AD3 illustrate example techniques for displaying a title bar with information for an application, in accordance with some embodiments.
- Figure 10 is a flow diagram of methods of displaying a title bar with information for an application, in accordance with various embodiments.
- the user interfaces in Figures 7S-7AD3 are used to illustrate the processes in Figure 10.
- Figures 7T (e.g., Figures 7T1, 7T2 and 7T3)-7AH illustrate example techniques for displaying privacy indicators, in accordance with some embodiments.
- Figure 11 is a flow diagram of methods of displaying privacy indicators, in accordance with various embodiments.
- the user interfaces in Figures 7T-7AH are used to illustrate the processes in Figure 11.
- Figures 7AI-7AP illustrate example techniques for displaying some user interface elements with values for a parameter constrained to a first range, and displaying other user interface elements with values for the parameter that are outside the first range, in accordance with some embodiments.
- Figure 12 is a flow diagram of methods for displaying some user interface elements with values for a parameter constrained to a first range, and displaying other user interface elements with values for the parameter that are outside the first range, in accordance with various embodiments.
- the user interfaces in Figures 7AI-7AP are used to illustrate the processes in Figure 12.
- Figures 7AQ-7BB illustrate example techniques for displaying controls for a virtual object and automatically updating the displayed controls as the virtual object is moved within the AR/VR environment and for displaying controls for an object and automatically reducing a prominence of the controls while the object is moved within the AR/VR environment, in accordance with some embodiments.
- Figure 13 is a flow diagram of methods of displaying controls for a virtual object and automatically updating the displayed controls as the virtual object is moved within the AR/VR environment, in accordance with various embodiments.
- Figure 15 is a flow diagram of methods of displaying controls for an object and automatically reducing a prominence of the controls while the object is moved within the AR/VR environment, in accordance with various embodiments.
- the user interfaces in Figures 7AQ-7BB are used to illustrate the processes in Figures 13 and 15.
- Figures 7BC-7BL illustrate example techniques for closing different applications or groups of applications in a three-dimensional environment in response to different inputs.
- Figure 14 is a flow diagram of methods of closing different applications or groups of applications in a three-dimensional environment in response to different inputs, in accordance with various embodiments.
- Figures 7BC-7BL are used to illustrate the processes in Figure 14.
- Figures 7BM-7CH illustrate example techniques for displaying a home menu user interface after closing one or more application user interfaces, in accordance with some embodiments.
- Figure 16 is a flow diagram of methods of displaying a home menu user interface after closing one or more application user interfaces, in accordance with various embodiments.
- the user interfaces in Figures 7BM-7CH are used to illustrate the processes in Figure 16.
- Figures 17A1-17T illustrate example techniques for scaling a user interface object, during a resize operation of the user interface object, by different scaling factors based on a refresh rate of content in the user interface object, in accordance with various embodiments.
- Figure 18 is a flow diagram of methods for scaling a user interface object, during a resize operation of the user interface object, by different scaling factors based on a refresh rate of content in the user interface object, in accordance with various embodiments.
- the user interfaces in Figures 17A1-17T are used to illustrate the processes in Figure 18.
- Figures 19A-19S illustrate example techniques for scaling a user interface object based on respective gestures, in accordance with various embodiments.
- Figure 20 is a flow diagram for scaling a user interface object based on respective gestures, in accordance with various embodiments.
- the user interfaces in 19A-19S are used to illustrate the processes in Figure 20.
- the processes described below enhance the operability of the devices and make the user-device interfaces more efficient (e.g., by helping the user to provide proper inputs and reducing user mistakes when operating/interacting with the device) through various techniques, including by providing improved visual feedback to the user, reducing the number of inputs needed to perform an operation, providing additional control options without cluttering the user interface with additional displayed controls, performing an operation when a set of conditions has been met without requiring further user input, improving privacy and/or security, providing a more varied, detailed, and/or realistic user experience while saving storage space, and/or additional techniques. These techniques also reduce power usage and improve battery life of the device by enabling the user to use the device more quickly and efficiently. Saving on battery power, and thus weight, improves the ergonomics of the device.
- These techniques also enable real-time communication, allow for the use of fewer and/or less precise sensors resulting in a more compact, lighter, and cheaper device, and enable the device to be used in a variety of lighting conditions. These techniques reduce energy usage, thereby reducing heat emitted by the device, which is particularly important for a wearable device where a device well within operational parameters for device components can become uncomfortable for a user to wear if it is producing too much heat.
- system or computer readable medium contains instructions for performing the contingent operations based on the satisfaction of the corresponding one or more conditions and thus is capable of determining whether the contingency has or has not been satisfied without explicitly repeating steps of a method until all of the conditions upon which steps in the method are contingent have been met.
- a system or computer readable storage medium can repeat the steps of a method as many times as are needed to ensure that all of the contingent steps have been performed.
- the XR experience is provided to the user via an operating environment 100 that includes a computer system 101.
- the computer system 101 includes a controller 110 (e.g., processors of a portable electronic device or a remote server), a display generation component 120 (e.g., a head-mounted device (HMD), a display, a projector, a touch-screen, etc.), one or more input devices 125 (e.g., an eye tracking device 130, a hand tracking device 140, other input devices 150), one or more output devices 155 (e.g., speakers 160, tactile output generators 170, and other output devices 180), one or more sensors 190 (e.g., image sensors, light sensors, depth sensors, tactile sensors, orientation sensors, proximity sensors, temperature sensors, location sensors, motion sensors, velocity sensors, etc.), and optionally one or more peripheral devices 195 (e.g., home appliances, wearable devices, etc.).
- Physical environment refers to a physical world that people can sense and/or interact with without aid of electronic systems.
- Physical environments such as a physical park, include physical articles, such as physical trees, physical buildings, and physical people. People can directly sense and/or interact with the physical environment, such as through sight, touch, hearing, taste, and smell.
- Extended reality In contrast, an extended reality (XR) environment refers to a wholly or partially simulated environment that people sense and/or interact with via an electronic system.
- XR extended reality
- an XR system may detect a person’s head turning and, in response, adjust graphical content and an acoustic field presented to the person in a manner similar to how such views and sounds would change in a physical environment.
- adjustments to characteristic(s) of virtual object(s) in an XR environment may be made in response to representations of physical motions (e.g., vocal commands).
- a person may sense and/or interact with an XR object using any one of their senses, including sight, sound, touch, taste, and smell.
- a person may sense and/or interact with audio objects that create a 3D or spatial audio environment that provides the perception of point audio sources in 3D space.
- audio objects may enable audio transparency, which selectively incorporates ambient sounds from the physical environment with or without computer-generated audio.
- a person may sense and/or interact only with audio objects.
- Examples of XR include virtual reality and mixed reality.
- a virtual reality (VR) environment refers to a simulated environment that is designed to be based entirely on computer-generated sensory inputs for one or more senses.
- a VR environment comprises a plurality of virtual objects with which a person may sense and/or interact.
- virtual objects For example, computer-generated imagery of trees, buildings, and avatars representing people are examples of virtual objects.
- a person may sense and/or interact with virtual objects in the VR environment through a simulation of the person’s presence within the computer-generated environment, and/or through a simulation of a subset of the person’s physical movements within the computer-generated environment.
- a mixed reality (MR) environment In contrast to a VR environment, which is designed to be based entirely on computer-generated sensory inputs, a mixed reality (MR) environment refers to a simulated environment that is designed to incorporate sensory inputs from the physical environment, or a representation thereof, in addition to including computer-generated sensory inputs (e.g., virtual objects).
- MR mixed reality
- a mixed reality environment is anywhere between, but not including, a wholly physical environment at one end and virtual reality environment at the other end.
- computer-generated sensory inputs may respond to changes in sensory inputs from the physical environment.
- some electronic systems for presenting an MR environment may track location and/or orientation with respect to the physical environment to enable virtual objects to interact with real objects (that is, physical articles from the physical environment or representations thereof). For example, a system may account for movements so that a virtual tree appears stationary with respect to the physical ground.
- mixed realities include augmented reality and augmented virtuality.
- Augmented reality refers to a simulated environment in which one or more virtual objects are superimposed over a physical environment, or a representation thereof.
- an electronic system for presenting an AR environment may have a transparent or translucent display through which a person may directly view the physical environment.
- the system may be configured to present virtual objects on the transparent or translucent display, so that a person, using the system, perceives the virtual objects superimposed over the physical environment.
- a system may have an opaque display and one or more imaging sensors that capture images or video of the physical environment, which are representations of the physical environment. The system composites the images or video with virtual objects, and presents the composition on the opaque display.
- a person, using the system indirectly views the physical environment by way of the images or video of the physical environment, and perceives the virtual objects superimposed over the physical environment.
- a video of the physical environment shown on an opaque display is called “pass-through video,” meaning a system uses one or more image sensor(s) to capture images of the physical environment, and uses those images in presenting the AR environment on the opaque display.
- a system may have a projection system that projects virtual objects into the physical environment, for example, as a hologram or on a physical surface, so that a person, using the system, perceives the virtual objects superimposed over the physical environment.
- An augmented reality environment also refers to a simulated environment in which a representation of a physical environment is transformed by computer-generated sensory information.
- a system may transform one or more sensor images to impose a select perspective (e.g., viewpoint) different than the perspective captured by the imaging sensors.
- a representation of a physical environment may be transformed by graphically modifying (e.g., enlarging) portions thereof, such that the modified portion may be representative but not photorealistic versions of the originally captured images.
- a representation of a physical environment may be transformed by graphically eliminating or obfuscating portions thereof.
- Augmented virtuality refers to a simulated environment in which a virtual or computer-generated environment incorporates one or more sensory inputs from the physical environment.
- the sensory inputs may be representations of one or more characteristics of the physical environment.
- an AV park may have virtual trees and virtual buildings, but people with faces photorealistically reproduced from images taken of physical people.
- a virtual object may adopt a shape or color of a physical article imaged by one or more imaging sensors.
- a virtual object may adopt shadows consistent with the position of the sun in the physical environment.
- a view of a three-dimensional environment is visible to a user.
- the view of the three-dimensional environment is typically visible to the user via one or more display generation components (e.g., a display or a pair of display modules that provide stereoscopic content to different eyes of the same user) through a virtual viewport that has a viewport boundary that defines an extent of the three-dimensional environment that is visible to the user via the one or more display generation components.
- display generation components e.g., a display or a pair of display modules that provide stereoscopic content to different eyes of the same user
- the region defined by the viewport boundary is smaller than a range of vision of the user in one or more dimensions (e.g., based on the range of vision of the user, size, optical properties or other physical characteristics of the one or more display generation components, and/or the location and/or orientation of the one or more display generation components relative to the eyes of the user). In some embodiments, the region defined by the viewport boundary is larger than a range of vision of the user in one or more dimensions (e.g., based on the range of vision of the user, size, optical properties or other physical characteristics of the one or more display generation components, and/or the location and/or orientation of the one or more display generation components relative to the eyes of the user).
- the viewport and viewport boundary typically move as the one or more display generation components move (e.g., moving with a head of the user for a head mounted device or moving with a hand of a user for a handheld device such as a tablet or smartphone).
- a viewpoint of a user determines what content is visible in the viewport, a viewpoint generally specfies a location and a direction relative to the three-dimensional environment, and as the viewpoint shifts, the view of the three-dimensional environment will also shift in the viewport.
- a viewpoint is typically based on a location and direction of the head, face, and/or eyes of a user to provide a view of the three- dimensional environment that is perceptually accurate and provides an immersive experience when the user is using the head-mounted device.
- the viewpoint shifts as the handheld or stationed device is moved and/or as a position of a user relative to the handheld or stationed device changes (e.g., a user moving toward, away from, up, down, to the right, and/or to the left of the device).
- portions of the physical environment that are visible (e.g., displayed, and/or projected) via the one or more display generation components are based on a field of view of one or more cameras in communication with the display generation components which typically move with the display generation components (e.g., moving with a head of the user for a head mounted device or moving with a hand of a user for a handheld device such as a tablet or smartphone) because the viewpoint of the user moves as the field of view of the one or more cameras moves (and the appearance of one or more virtual objects displayed via the one or more display generation components is updated based on the viewpoint of the user (e.g., displayed positions and poses of the virtual objects are updated based on the movement of the viewpoint of the user)).
- the viewpoint of the user e.g., displayed positions and poses of the virtual objects are updated based on the movement of the viewpoint of the user
- portions of the physical environment that are visible (e.g., optically visible through one or more partially or fully transparent portions of the display generation component) via the one or more display generation components are based on a field of view of a user through the partially or fully transparent portion(s) of the display generation component (e.g., moving with a head of the user for a head mounted device or moving with a hand of a user for a handheld device such as a tablet or smartphone) because the viewpoint of the user moves as the field of view of the user through the partially or fully transparent portions of the display generation components moves (and the appearance of one or more virtual objects is updated based on the viewpoint of the user).
- a field of view of a user e.g., moving with a head of the user for a head mounted device or moving with a hand of a user for a handheld device such as a tablet or smartphone
- a representation of a physical environment can be partially or fully obscured by a virtual environment.
- the amount of virtual environment that is displayed is based on an immersion level for the virtual environment (e.g., with respect to the representation of the physical environment). For example, increasing the immersion level optionally causes more of the virtual environment to be displayed, replacing and/or obscuring more of the physical environment, and reducing the immersion level optionally causes less of the virtual environment to be displayed, revealing portions of the physical environment that were previously not displayed and/or obscured.
- one or more first background objects are visually de-emphasized (e.g., dimmed, blurred, and/or displayed with increased transparency) more than one or more second background objects, and one or more third background objects cease to be displayed.
- a level of immersion includes an associated degree to which the virtual content displayed by the computer system (e.g., the virtual environment and/or the virtual content) obscures background content (e.g., content other than the virtual environment and/or the virtual content) around/behind the virtual content, optionally including the number of items of background content displayed and/or the visual characteristics (e.g., colors, contrast, and/or opacity) with which the background content is displayed, the angular range of the virtual content displayed via the display generation component (e.g., 60 degrees of content displayed at low immersion, 120 degrees of content displayed at medium immersion, or 180 degrees of content displayed at high immersion), and/or the proportion of the field of view displayed via the display generation component that is consumed by the virtual content (e.g., 33% of the field of view consumed by the virtual content at low immersion, 66% of the field of view consumed by the virtual content at medium immersion, or 100% of the field of view consumed by the virtual content at high immersion).
- the virtual content displayed by the computer system e.g., the virtual
- the background content is included in a background over which the virtual content is displayed (e.g., background content in the representation of the physical environment).
- the background content includes user interfaces (e.g., user interfaces generated by the computer system corresponding to applications), virtual objects (e.g., files or representations of other users generated by the computer system) not associated with or included in the virtual environment and/or virtual content, and/or real objects (e.g., pass-through objects representing real objects in the physical environment around the user that are visible such that they are displayed via the display generation component and/or a visible via a transparent or translucent component of the display generation component because the computer system does not obscure/prevent visibility of them through the display generation component).
- user interfaces e.g., user interfaces generated by the computer system corresponding to applications
- virtual objects e.g., files or representations of other users generated by the computer system
- real objects e.g., pass-through objects representing real objects in the physical environment around the user that are visible such that they are displayed via the display generation component and
- the background, virtual and/or real objects are displayed in an unobscured manner.
- a virtual environment with a low level of immersion is optionally displayed concurrently with the background content, which is optionally displayed with full brightness, color, and/or translucency.
- the background, virtual and/or real objects are displayed in an obscured manner (e.g., dimmed, blurred, or removed from display).
- a respective virtual environment with a high level of immersion is displayed without concurrently displaying the background content (e.g., in a full screen or fully immersive mode).
- a virtual environment displayed with a medium level of immersion is displayed concurrently with darkened, blurred, or otherwise de-emphasized background content.
- the visual characteristics of the background objects vary among the background objects. For example, at a particular immersion level, one or more first background objects are visually de-emphasized (e.g., dimmed, blurred, and/or displayed with increased transparency) more than one or more second background objects, and one or more third background objects cease to be displayed.
- a null or zero level of immersion corresponds to the virtual environment ceasing to be displayed and instead a representation of a physical environment is displayed (optionally with one or more virtual objects such as application, windows, or virtual three- dimensional objects) without the representation of the physical environment being obscured by the virtual environment.
- Adjusting the level of immersion using a physical input element provides for quick and efficient method of adjusting immersion, which enhances the operability of the computer system and makes the user-device interface more efficient.
- Viewpoint-locked virtual object A virtual object is viewpoint-locked when a computer system displays the virtual object at the same location and/or position in the viewpoint of the user, even as the viewpoint of the user shifts (e.g., changes).
- the viewpoint of the user is locked to the forward facing direction of the user’s head (e.g., the viewpoint of the user is at least a portion of the field-of-view of the user when the user is looking straight ahead); thus, the viewpoint of the user remains fixed even as the user’s gaze is shifted, without moving the user’s head.
- the viewpoint of the user is the augmented reality view that is being presented to the user on a display generation component of the computer system.
- a viewpoint-locked virtual object that is displayed in the upper left corner of the viewpoint of the user, when the viewpoint of the user is in a first orientation (e.g., with the user’s head facing north) continues to be displayed in the upper left corner of the viewpoint of the user, even as the viewpoint of the user changes to a second orientation (e.g., with the user’s head facing west).
- the location and/or position at which the viewpoint-locked virtual object is displayed in the viewpoint of the user is independent of the user’s position and/or orientation in the physical environment.
- the viewpoint of the user is locked to the orientation of the user’s head, such that the virtual object is also referred to as a “head-locked virtual object.”
- a virtual object is environment-locked (alternatively, “world-locked”) when a computer system displays the virtual object at a location and/or position in the viewpoint of the user that is based on (e.g., selected in reference to and/or anchored to) a location and/or object in the three-dimensional environment (e.g., a physical environment or a virtual environment).
- a location and/or object in the three-dimensional environment e.g., a physical environment or a virtual environment.
- an environment-locked virtual object that is locked onto a tree that is immediately in front of a user is displayed at the center of the viewpoint of the user.
- the viewpoint of the user shifts to the right (e.g., the user’s head is turned to the right) so that the tree is now left-of-center in the viewpoint of the user (e.g., the tree’s position in the viewpoint of the user shifts)
- the environment-locked virtual object that is locked onto the tree is displayed left-of-center in the viewpoint of the user.
- the location and/or position at which the environment- locked virtual object is displayed in the viewpoint of the user is dependent on the position and/or orientation of the location and/or object in the environment onto which the virtual object is locked.
- the computer system uses a stationary frame of reference (e.g., a coordinate system that is anchored to a fixed location and/or object in the physical environment) in order to determine the position at which to display an environment- locked virtual object in the viewpoint of the user.
- a stationary frame of reference e.g., a coordinate system that is anchored to a fixed location and/or object in the physical environment
- An environment-locked virtual object can be locked to a stationary part of the environment (e.g., a floor, wall, table, or other stationary object) or can be locked to a moveable part of the environment (e.g., a vehicle, animal, person, or even a representation of portion of the users body that moves independently of a viewpoint of the user, such as a user’s hand, wrist, arm, or foot) so that the virtual object is moved as the viewpoint or the portion of the environment moves to maintain a fixed relationship between the virtual object and the portion of the environment.
- a stationary part of the environment e.g., a floor, wall, table, or other stationary object
- a moveable part of the environment e.g., a vehicle, animal, person, or even a representation of portion of the users body that moves independently of a viewpoint of the user, such as a user’s hand, wrist, arm, or foot
- a virtual object that is environment-locked or viewpoint-locked exhibits lazy follow behavior which reduces or delays motion of the environment-locked or viewpoint-locked virtual object relative to movement of a point of reference which the virtual object is following.
- the computer system when exhibiting lazy follow behavior the computer system intentionally delays movement of the virtual object when detecting movement of a point of reference (e.g., a portion of the environment, the viewpoint, or a point that is fixed relative to the viewpoint, such as a point that is between 5- 300cm from the viewpoint) which the virtual object is following.
- the virtual object when the point of reference (e.g., the portion of the environement or the viewpoint) moves with a first speed, the virtual object is moved by the device to remain locked to the point of reference but moves with a second speed that is slower than the first speed (e.g., until the point of reference stops moving or slows down, at which point the virtual object starts to catch up to the point of reference).
- the device when a virtual object exhibits lazy follow behavior the device ignores small amounts of movment of the point of reference (e.g., ignoring movement of the point of reference that is below a threshold amount of movement such as movement by 0-5 degrees or movement by 0-50 cm).
- a distance between the point of reference and the virtual object increases (e.g., because the virtual object is being displayed so as to maintain a fixed or substantially fixed position relative to a viewpoint or portion of the environment that is different from the point of reference to which the virtual object is locked) and when the point of reference (e.g., the portion of the environment or the viewpoint to which the virtual object is locked) moves by a second amount that is greater than the first amount, a distance between the point of reference and the virtual object initially increases (e.g., because the virtual object is being displayed so as to maintain a fixed or substantially fixed position relative to a viewpoint or portion of the environment that is different from the point of reference to which the virtual object is locked) and then decreases as the amount of movement of the point of reference increases above a threshold (e.g., a “lazy follow” threshold) because the virtual object is moved by the computer system to maintain a fixed
- a threshold e.g., a “lazy follow” threshold
- the virtual object maintaining a substantially fixed position relative to the point of reference includes the virtual object being displayed within a threshold distance (e.g., 1, 2, 3, 5, 15, 20, 50 cm) of the point of reference in one or more dimensions (e.g., up/down, left/right, and/or forward/backward relative to the position of the point of reference).
- a threshold distance e.g. 1, 2, 3, 5, 15, 20, 50 cm
- Hardware There are many different types of electronic systems that enable a person to sense and/or interact with various XR environments. Examples include headmounted systems, projection-based systems, heads-up displays (HUDs), vehicle windshields having integrated display capability, windows having integrated display capability, displays formed as lenses designed to be placed on a person’s eyes (e.g., similar to contact lenses), headphones/earphones, speaker arrays, input systems (e.g., wearable or handheld controllers with or without haptic feedback), smartphones, tablets, and desktop/laptop computers.
- a head-mounted system may have one or more speaker(s) and an integrated opaque display.
- a head-mounted system may be configured to accept an external opaque display (e.g., a smartphone).
- the head-mounted system may incorporate one or more imaging sensors to capture images or video of the physical environment, and/or one or more microphones to capture audio of the physical environment.
- a head-mounted system may have a transparent or translucent display.
- the transparent or translucent display may have a medium through which light representative of images is directed to a person’s eyes.
- the display may utilize digital light projection, OLEDs, LEDs, uLEDs, liquid crystal on silicon, laser scanning light source, or any combination of these technologies.
- the medium may be an optical waveguide, a hologram medium, an optical combiner, an optical reflector, or any combination thereof.
- the transparent or translucent display may be configured to become opaque selectively.
- Projection-based systems may employ retinal projection technology that projects graphical images onto a person’s retina. Projection systems also may be configured to project virtual objects into the physical environment, for example, as a hologram or on a physical surface.
- the controller 110 is configured to manage and coordinate an XR experience for the user.
- the controller 110 includes a suitable combination of software, firmware, and/or hardware. The controller 110 is described in greater detail below with respect to Figure 2.
- the controller 110 is a computing device that is local or remote relative to the scene 105 (e.g., a physical environment). For example, the controller 110 is a local server located within the scene 105.
- the controller 110 is a remote server located outside of the scene 105 (e.g., a cloud server, central server, etc.).
- the controller 110 is communicatively coupled with the display generation component 120 (e.g., an HMD, a display, a projector, a touch-screen, etc.) via one or more wired or wireless communication channels 144 (e.g., BLUETOOTH, IEEE 802.1 lx, IEEE 802.16x, IEEE 802.3x, etc.).
- the display generation component 120 e.g., an HMD, a display, a projector, a touch-screen, etc.
- wired or wireless communication channels 144 e.g., BLUETOOTH, IEEE 802.1 lx, IEEE 802.16x, IEEE 802.3x, etc.
- the controller 110 is included within the enclosure (e.g., a physical housing) of the display generation component 120 (e.g., an HMD, or a portable electronic device that includes a display and one or more processors, etc.), one or more of the input devices 125, one or more of the output devices 155, one or more of the sensors 190, and/or one or more of the peripheral devices 195, or share the same physical enclosure or support structure with one or more of the above.
- the display generation component 120 e.g., an HMD, or a portable electronic device that includes a display and one or more processors, etc.
- the display generation component 120 is configured to provide the XR experience (e.g., at least a visual component of the XR experience) to the user.
- the display generation component 120 includes a suitable combination of software, firmware, and/or hardware. The display generation component 120 is described in greater detail below with respect to Figure 3.
- the functionalities of the controller 110 are provided by and/or combined with the display generation component 120.
- the display generation component 120 provides an XR experience to the user while the user is virtually and/or physically present within the scene 105.
- the display generation component is worn on a part of the user’s body (e.g., on his/her head, on his/her hand, etc.).
- the display generation component 120 includes one or more XR displays provided to display the XR content.
- the display generation component 120 encloses the field- of-view of the user.
- the display generation component 120 is a handheld device (such as a smartphone or tablet) configured to present XR content, and the user holds the device with a display directed towards the field-of-view of the user and a camera directed towards the scene 105.
- the handheld device is optionally placed within an enclosure that is worn on the head of the user.
- the handheld device is optionally placed on a support (e.g., a tripod) in front of the user.
- the display generation component 120 is an XR chamber, enclosure, or room configured to present XR content in which the user does not wear or hold the display generation component 120.
- Many user interfaces described with reference to one type of hardware for displaying XR content e.g., a handheld device or a device on a tripod
- could be implemented on another type of hardware for displaying XR content e.g., an HMD or other wearable computing device.
- a user interface showing interactions with XR content triggered based on interactions that happen in a space in front of a handheld or tripod mounted device could similarly be implemented with an HMD where the interactions happen in a space in front of the HMD and the responses of the XR content are displayed via the HMD.
- a user interface showing interactions with XR content triggered based on movement of a handheld or tripod mounted device relative to the physical environment could similarly be implemented with an HMD where the movement is caused by movement of the HMD relative to the physical environment (e.g., the scene 105 or a part of the user’s body (e.g., the user’s eye(s), head, or hand)).
- Figures 1 A-1P illustrate various examples of a computer system that is used to perform the methods and provide audio, visual and/or haptic feedback as part of user interfaces described herein.
- the computer system includes one or more display generation components (e.g., first and second display assemblies l-120a, l-120b and/or first and second optical modules 11.1. l-104a and 11.1.
- l-104b for displaying virtual elements and/or a representation of a physical environment to a user of the computer system, optionally generated based on detected events and/or user inputs detected by the computer system.
- User interfaces generated by the computer system are optionally corrected by one or more corrective lenses 11.3.2-216 that are optionally removably attached to one or more of the optical modules to enable the user interfaces to be more easily viewed by users who would otherwise use glasses or contacts to correct their vision. While many user interfaces illustrated herein show a single view of a user interface, user interfaces in a HMD are optionally displayed using two optical modules (e.g., first and second display assemblies 1- 120a, l-120b and/or first and second optical modules l l.l.
- the computer system includes one or more external displays (e.g., display assembly 1-108) for displaying status information for the computer system to the user of the computer system (when the computer system is not being worn) and/or to other people who are near the computer system, optionally generated based on detected events and/or user inputs detected by the computer system.
- external displays e.g., display assembly 1-108
- the computer system includes one or more audio output components (e.g., electronic component 1-112) for generating audio feedback, optionally generated based on detected events and/or user inputs detected by the computer system.
- the computer system includes one or more input devices for detecting input such as one or more sensors (e.g., one or more sensors in sensor assembly 1- 356, and/or Figure II) for detecting information about a physical environment of the device which can be used (optionally in conjunction with one or more illuminators such as the illuminators described in Figure II) to generate a digital passthrough image, capture visual media corresponding to the physical environment (e.g., photos and/or video), or determine a pose (e.g., position and/or orientation) of physical objects and/or surfaces in the physical environment so that virtual objects ban be placed based on a detected pose of physical objects and/or surfaces.
- a pose e.g., position and/or orientation
- the computer system includes one or more input devices for detecting input such as one or more sensors for detecting hand position and/or movement (e.g., one or more sensors in sensor assembly 1-356, and/or Figure II) that can be used (optionally in conjunction with one or more illuminators such as the illuminators 6-124 described in Figure II) to determine when one or more air gestures have been performed.
- one or more sensors for detecting hand position and/or movement e.g., one or more sensors in sensor assembly 1-356, and/or Figure II
- one or more illuminators such as the illuminators 6-124 described in Figure II
- the computer system includes one or more input devices for detecting input such as one or more sensors for detecting eye movement (e.g., eye tracking and gaze tracking sensors in Figure II) which can be used (optionally in conjunction with one or more lights such as lights 11.3.2-110 in Figure 10) to determine attention or gaze position and/or gaze movement which can optionally be used to detect gaze-only inputs based on gaze movement and/or dwell.
- one or more sensors for detecting eye movement e.g., eye tracking and gaze tracking sensors in Figure II
- lights e.g., lights 11.3.2-110 in Figure 10
- a combination of the various sensors described above can be used to determine user facial expressions and/or hand movements for use in generating an avatar or representation of the user such as an anthropomorphic avatar or representation for use in a real-time communication session where the avatar has facial expressions, hand movements, and/or body movements that are based on or similar to detected facial expressions, hand movements, and/or body movements of a user of the device.
- Gaze and/or attention information is, optionally, combined with hand tracking information to determine interactions between the user and one or more user interfaces based on direct and/or indirect inputs such as air gestures or inputs that use one or more hardware input devices such as one or more buttons (e.g., first button 1-128, button 11.1.1-114 , second button 1-132, and or dial or button 1-328), knobs (e.g., first button 1-128, button 11.1.1-114, and/or dial or button 1-328), digital crowns (e.g., first button 1-128 which is depressible and twistable or rotatable, button 11.1.1-114, and/or dial or button 1-328), trackpads, touch screens, keyboards, mice and/or other input devices.
- buttons e.g., first button 1-128, button 11.1.1-114 , second button 1-132, and or dial or button 1-328
- knobs e.g., first button 1-128, button 11.1.1-114, and/or dial or button 1-328
- digital crowns e.g.
- buttons are optionally used to perform system operations such as recentering content in three-dimensional environment that is visible to a user of the device, displaying a home user interface for launching applications, starting real-time communication sessions, or initiating display of virtual three-dimensional backgrounds.
- Knobs or digital crowns are optionally rotatable to adjust parameters of the visual content such as a level of immersion of a virtual three-dimensional environment (e.g., a degree to which virtual-content occupies the viewport of the user into the three-dimensional environment) or other parameters associated with the three-dimensional environment and the virtual content that is displayed via the optical modules (e.g., first and second display assemblies l-120a, l-120b and/or first and second optical modules 11.1.1- 104a and 11.1. l-104b).
- the optical modules e.g., first and second display assemblies l-120a, l-120b and/or first and second optical modules 11.1.1- 104a and 11.1. l-104b).
- FIG. IB illustrates a front, top, perspective view of an example of a head- mountable display (HMD) device 1-100 configured to be donned by a user and provide virtual and altered/mixed reality (VR/AR) experiences.
- the HMD 1-100 can include a display unit 1-102 or assembly, an electronic strap assembly 1-104 connected to and extending from the display unit 1-102, and a band assembly 1-106 secured at either end to the electronic strap assembly 1-104.
- the electronic strap assembly 1-104 and the band 1-106 can be part of a retention assembly configured to wrap around a user’s head to hold the display unit 1-102 against the face of the user.
- the band assembly 1-106 can include a first band 1- 116 configured to wrap around the rear side of a user’s head and a second band 1-117 configured to extend over the top of a user’s head.
- the second strap can extend between first and second electronic straps l-105a, 1 -105b of the electronic strap assembly 1-104 as shown.
- the strap assembly 1-104 and the band assembly 1-106 can be part of a securement mechanism extending rearward from the display unit 1-102 and configured to hold the display unit 1-102 against a face of a user.
- the securement mechanism includes a first electronic strap l-105a including a first proximal end 1-134 coupled to the display unit 1-102, for example a housing 1-150 of the display unit 1-102, and a first distal end 1-136 opposite the first proximal end 1-134.
- the securement mechanism can also include a second electronic strap 1 -105b including a second proximal end 1-138 coupled to the housing 1-150 of the display unit 1-102 and a second distal end 1-140 opposite the second proximal end 1-138.
- the securement mechanism can also include the first band 1-116 including a first end 1-142 coupled to the first distal end 1-136 and a second end 1-144 coupled to the second distal end 1-140 and the second band 1-117 extending between the first electronic strap l-105a and the second electronic strap 1 - 105b .
- the straps l-105a-b and band 1-116 can be coupled via connection mechanisms or assemblies 1-114.
- the second band 1-117 includes a first end 1-146 coupled to the first electronic strap l-105a between the first proximal end 1-134 and the first distal end 1-136 and a second end 1-148 coupled to the second electronic strap 1-105b between the second proximal end 1-138 and the second distal end 1-140.
- the first and second electronic straps l-105a-b include plastic, metal, or other structural materials forming the shape the substantially rigid straps 1- 105a-b.
- the first and second bands 1-116, 1-117 are formed of elastic, flexible materials including woven textiles, rubbers, and the like. The first and second bands 1-116, 1-117 can be flexible to conform to the shape of the user’ head when donning the HMD 1-100.
- one or more of the first and second electronic straps 1- 105a-b can define internal strap volumes and include one or more electronic components disposed in the internal strap volumes.
- the first electronic strap l-105a can include an electronic component 1-112.
- the electronic component 1-112 can include a speaker.
- the electronic component 1-112 can include a computing component such as a processor.
- the housing 1-150 defines a first, front-facing opening 1-152.
- the front-facing opening is labeled in dotted lines at 1-152 in FIG. IB because the display assembly 1-108 is disposed to occlude the first opening 1-152 from view when the HMD 1-100 is assembled.
- the housing 1-150 can also define a rear-facing second opening 1- 154.
- the housing 1-150 also defines an internal volume between the first and second openings 1-152, 1-154.
- the HMD 1-100 includes the display assembly 1-108, which can include a front cover and display screen (shown in other figures) disposed in or across the front opening 1-152 to occlude the front opening 1-152.
- the display screen of the display assembly 1-108 has a curvature configured to follow the curvature of a user’s face.
- the display screen of the display assembly 1-108 can be curved as shown to compliment the user’s facial features and general curvature from one side of the face to the other, for example from left to right and/or from top to bottom where the display unit 1-102 is pressed.
- the housing 1-150 can define a first aperture 1-126 between the first and second openings 1-152, 1-154 and a second aperture 1-130 between the first and second openings 1-152, 1-154.
- the HMD 1-100 can also include a first button 1-128 disposed in the first aperture 1-126 and a second button 1-132 disposed in the second aperture 1-130.
- the first and second buttons 1-128, 1-132 can be depressible through the respective apertures 1-126, 1-130.
- the first button 1-126 and/or second button 1- 132 can be twistable dials as well as depressible buttons.
- the first button 1-128 is a depressible and twistable dial button and the second button 1-132 is a depressible button.
- FIG. 1C illustrates a rear, perspective view of the HMD 1-100.
- the HMD 1- 100 can include a light seal 1-110 extending rearward from the housing 1-150 of the display assembly 1-108 around a perimeter of the housing 1-150 as shown.
- the light seal 1-110 can be configured to extend from the housing 1-150 to the user’s face around the user’s eyes to block external light from being visible.
- the HMD 1-100 can include first and second display assemblies l-120a, l-120b disposed at or in the rearward facing second opening 1-154 defined by the housing 1-150 and/or disposed in the internal volume of the housing 1-150 and configured to project light through the second opening 1-154.
- each display assembly l-120a-b can include respective display screens l-122a, l-122b configured to project light in a rearward direction through the second opening 1-154 toward the user’s eyes.
- the display assembly 1-108 can be a front-facing, forward display assembly including a display screen configured to project light in a first, forward direction and the rear facing display screens 1- 122a-b can be configured to project light in a second, rearward direction opposite the first direction.
- the light seal 1-110 can be configured to block light external to the HMD 1-100 from reaching the user’s eyes, including light projected by the forward facing display screen of the display assembly 1-108 shown in the front perspective view of FIG. IB.
- the HMD 1-100 can also include a curtain 1-124 occluding the second opening 1-154 between the housing 1-150 and the rear-facing display assemblies 1- 120a-b.
- the curtain 1-124 can be elastic or at least partially elastic.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIGS. IB and 1C can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. ID - IF and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. ID - IF can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIGS. IB and 1C.
- FIG. ID illustrates an exploded view of an example of an HMD 1-200 including various portions or parts thereof separated according to the modularity and selective coupling of those parts.
- the HMD 1-200 can include a band 1-216 which can be selectively coupled to first and second electronic straps l-205a, l-205b.
- the first securement strap l-205a can include a first electronic component 1-212a and the second securement strap l-205b can include a second electronic component 1-212b.
- the first and second straps l-205a-b can be removably coupled to the display unit 1- 202.
- the HMD 1-200 can include a light seal 1-210 configured to be removably coupled to the display unit 1-202.
- the HMD 1-200 can also include lenses 1-218 which can be removably coupled to the display unit 1-202, for example over first and second display assemblies including display screens.
- the lenses 1-218 can include customized prescription lenses configured for corrective vision.
- each part shown in the exploded view of FIG. ID and described above can be removably coupled, attached, reattached, and changed out to update parts or swap out parts for different users.
- bands such as the band 1-216, light seals such as the light seal 1-210, lenses such as the lenses 1-218, and electronic straps such as the straps l-205a-b can be swapped out depending on the user such that these parts are customized to fit and correspond to the individual user of the HMD 1-200.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. ID can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. IB, 1C, and IE - IF and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. IB, 1C, and IE - IF can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. ID.
- FIG. IE illustrates an exploded view of an example of a display unit 1-306 of a HMD.
- the display unit 1-306 can include a front display assembly 1-308, a frame/housing assembly 1-350, and a curtain assembly 1-324.
- the display unit 1-306 can also include a sensor assembly 1-356, logic board assembly 1-358, and cooling assembly 1-360 disposed between the frame assembly 1-350 and the front display assembly 1-308.
- the display unit 1-306 can also include a rear-facing display assembly 1-320 including first and second rear-facing display screens l-322a, 1-322b disposed between the frame 1-350 and the curtain assembly 1-324.
- the display unit 1-306 can also include a motor assembly 1-362 configured as an adjustment mechanism for adjusting the positions of the display screens l-322a-b of the display assembly 1-320 relative to the frame 1-350.
- the display assembly 1-320 is mechanically coupled to the motor assembly 1- 362, with at least one motor for each display screen l-322a-b, such that the motors can translate the display screens l-322a-b to match an interpupillary distance of the user’s eyes.
- the display unit 1-306 can include a dial or button 1- 328 depressible relative to the frame 1-350 and accessible to the user outside the frame 1-350.
- the button 1-328 can be electronically connected to the motor assembly 1-362 via a controller such that the button 1-328 can be manipulated by the user to cause the motors of the motor assembly 1-362 to adjust the positions of the display screens l-322a-b.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IE can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. IB - ID and IF and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. IB - ID and IF can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IE.
- FIG. IF illustrates an exploded view of another example of a display unit 1- 406 of a HMD device similar to other HMD devices described herein.
- the display unit 1-406 can include a front display assembly 1-402, a sensor assembly 1-456, a logic board assembly 1-458, a cooling assembly 1-460, a frame assembly 1-450, a rear-facing display assembly 1- 421, and a curtain assembly 1-424.
- the display unit 1-406 can also include a motor assembly 1-462 for adjusting the positions of first and second display sub-assemblies l-420a, l-420b of the rear-facing display assembly 1-421, including first and second respective display screens for interpupillary adjustments, as described above.
- FIG. IF The various parts, systems, and assemblies shown in the exploded view of FIG. IF are described in greater detail herein with reference to FIGS. IB - IE as well as subsequent figures referenced in the present disclosure.
- the display unit 1-406 shown in FIG. IF can be assembled and integrated with the securement mechanisms shown in FIGS. IB - IE, including the electronic straps, bands, and other components including light seals, connection assemblies, and so forth.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IF can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. IB - IE and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. IB - IE can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IF.
- FIG. 1G illustrates a perspective, exploded view of a front cover assembly 3- 100 of an HMD device described herein, for example the front cover assembly 3-1 of the HMD 3-100 shown in FIG. 1G or any other HMD device shown and described herein.
- the front cover assembly 3-100 shown in FIG. 1G can include a transparent or semi-transparent cover 3-102, shroud 3-104 (or “canopy”), adhesive layers 3-106, display assembly 3-108 including a lenticular lens panel or array 3-110, and a structural trim 3-112.
- the adhesive layer 3-106 can secure the shroud 3-104 and/or transparent cover 3-102 to the display assembly 3-108 and/or the trim 3-112.
- the trim 3-112 can secure the various components of the front cover assembly 3-100 to a frame or chassis of the HMD device.
- the transparent cover 3-102, shroud 3-104, and display assembly 3-108 can be curved to accommodate the curvature of a user’s face.
- the transparent cover 3-102 and the shroud 3-104 can be curved in two or three dimensions, e.g., vertically curved in the Z- direction in and out of the Z-X plane and horizontally curved in the X-direction in and out of the Z-X plane.
- the display assembly 3-108 can include the lenticular lens array 3-110 as well as a display panel having pixels configured to project light through the shroud 3-104 and the transparent cover 3-102.
- the display assembly 3-108 can be curved in at least one direction, for example the horizontal direction, to accommodate the curvature of a user’s face from one side (e.g., left side) of the face to the other (e.g., right side).
- each layer or component of the display assembly 3-108 which will be shown in subsequent figures and described in more detail, but which can include the lenticular lens array 3-110 and a display layer, can be similarly or concentrically curved in the horizontal direction to accommodate the curvature of the user’s face.
- the shroud 3-104 can include a transparent or semitransparent material through which the display assembly 3-108 projects light.
- the shroud 3-104 can include one or more opaque portions, for example opaque ink-printed portions or other opaque film portions on the rear surface of the shroud 3-104.
- the rear surface can be the surface of the shroud 3-104 facing the user’s eyes when the HMD device is donned.
- opaque portions can be on the front surface of the shroud 3- 104 opposite the rear surface.
- the opaque portion or portions of the shroud 3-104 can include perimeter portions visually hiding any components around an outside perimeter of the display screen of the display assembly 3-108. In this way, the opaque portions of the shroud hide any other components, including electronic components, structural components, and so forth, of the HMD device that would otherwise be visible through the transparent or semi-transparent cover 3-102 and/or shroud 3-104.
- the shroud 3-104 can define one or more apertures transparent portions 3-120 through which sensors can send and receive signals.
- the portions 3-120 are apertures through which the sensors can extend or send and receive signals.
- the portions 3-120 are transparent portions, or portions more transparent than surrounding semi-transparent or opaque portions of the shroud, through which sensors can send and receive signals through the shroud and through the transparent cover 3-102.
- the sensors can include cameras, IR sensors, LUX sensors, or any other visual or non-visual environmental sensors of the HMD device.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. 1G can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described herein can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. 1G.
- FIG. 1H illustrates an exploded view of an example of an HMD device 6-100.
- the HMD device 6-100 can include a sensor array or system 6-102 including one or more sensors, cameras, projectors, and so forth mounted to one or more components of the HMD 6-100.
- the sensor system 6-102 can include a bracket 1-338 on which one or more sensors of the sensor system 6-102 can be fixed/secured.
- FIG. II illustrates a portion of an HMD device 6-100 including a front transparent cover 6-104 and a sensor system 6-102.
- the sensor system 6-102 can include a number of different sensors, emitters, receivers, including cameras, IR sensors, projectors, and so forth.
- the transparent cover 6-104 is illustrated in front of the sensor system 6-102 to illustrate relative positions of the various sensors and emitters as well as the orientation of each sensor/emitter of the system 6-102.
- “sideways,” “side,” “lateral,” “horizontal,” and other similar terms refer to orientations or directions as indicated by the X- axis shown in FIG. 1 J.
- the transparent cover 6-104 can define a front, external surface of the HMD device 6-100 and the sensor system 6-102, including the various sensors and components thereof, can be disposed behind the cover 6-104 in the Y- axis/direction.
- the cover 6-104 can be transparent or semi-transparent to allow light to pass through the cover 6-104, both light detected by the sensor system 6-102 and light emitted thereby.
- the HMD device 6-100 can include one or more controllers including processors for electrically coupling the various sensors and emitters of the sensor system 6-102 with one or more mother boards, processing units, and other electronic devices such as display screens and the like.
- the various sensors, emitters, and other components of the sensor system 6-102 can be coupled to various structural frame members, brackets, and so forth of the HMD device 6-100 not shown in FIG. II.
- FIG. II shows the components of the sensor system 6-102 unattached and un-coupled electrically from other components for the sake of illustrative clarity.
- the device can include one or more controllers having processors configured to execute instructions stored on memory components electrically coupled to the processors.
- the instructions can include, or cause the processor to execute, one or more algorithms for self-correcting angles and positions of the various cameras described herein overtime with use as the initial positions, angles, or orientations of the cameras get bumped or deformed due to unintended drop events or other events.
- the sensor system 6-102 can include one or more scene cameras 6-106.
- the system 6-102 can include two scene cameras 6-102 disposed on either side of the nasal bridge or arch of the HMD device 6-100 such that each of the two cameras 6-106 correspond generally in position with left and right eyes of the user behind the cover 6-103.
- the scene cameras 6-106 are oriented generally forward in the Y-direction to capture images in front of the user during use of the HMD 6-100.
- the scene cameras are color cameras and provide images and content for MR video pass through to the display screens facing the user’s eyes when using the HMD device 6-100.
- the scene cameras 6-106 can also be used for environment and object reconstruction.
- the sensor system 6-102 can include a first depth sensor 6-108 pointed generally forward in the Y-direction.
- the first depth sensor 6-108 can be used for environment and object reconstruction as well as user hand and body tracking.
- the sensor system 6-102 can include a second depth sensor 6-110 disposed centrally along the width (e.g., along the X-axis) of the HMD device 6-100.
- the second depth sensor 6-110 can be disposed above the central nasal bridge or accommodating features over the nose of the user when donning the HMD 6-100.
- the second depth sensor 6-110 can be used for environment and object reconstruction as well as hand and body tracking.
- the second depth sensor can include a LIDAR sensor.
- the sensor system 6-102 can include a depth projector 6-112 facing generally forward to project electromagnetic waves, for example in the form of a predetermined pattern of light dots, out into and within a field of view of the user and/or the scene cameras 6-106 or a field of view including and beyond the field of view of the user and/or scene cameras 6-106.
- the depth projector can project electromagnetic waves of light in the form of a dotted light pattern to be reflected off objects and back into the depth sensors noted above, including the depth sensors 6-108, 6-110.
- the depth projector 6-112 can be used for environment and object reconstruction as well as hand and body tracking.
- the sensor system 6-102 can include downward facing cameras 6-114 with a field of view pointed generally downward relative to the HDM device 6-100 in the Z-axis.
- the downward cameras 6-114 can be disposed on left and right sides of the HMD device 6-100 as shown and used for hand and body tracking, headset tracking, and facial avatar detection and creation for display a user avatar on the forward facing display screen of the HMD device 6-100 described elsewhere herein.
- the downward cameras 6-114 can be used to capture facial expressions and movements for the face of the user below the HMD device 6-100, including the cheeks, mouth, and chin.
- the sensor system 6-102 can include jaw cameras 6- 116.
- the jaw cameras 6-116 can be disposed on left and right sides of the HMD device 6-100 as shown and used for hand and body tracking, headset tracking, and facial avatar detection and creation for display a user avatar on the forward facing display screen of the HMD device 6-100 described elsewhere herein.
- the jaw cameras 6-116 can be used to capture facial expressions and movements for the face of the user below the HMD device 6-100, including the user’s jaw, cheeks, mouth, and chin.
- headset tracking, and facial avatar can be used to capture facial expressions and movements for the face of the user below the HMD device 6-100, including the user’s jaw, cheeks, mouth, and chin.
- the sensor system 6-102 can include side cameras 6- 118.
- the side cameras 6-118 can be oriented to capture side views left and right in the X-axis or direction relative to the HMD device 6-100.
- the side cameras 6- 118 can be used for hand and body tracking, headset tracking, and facial avatar detection and re-creation.
- the sensor system 6-102 can include a plurality of eye tracking and gaze tracking sensors for determining an identity, status, and gaze direction of a user’s eyes during and/or before use.
- the eye/gaze tracking sensors can include nasal eye cameras 6-120 disposed on either side of the user’s nose and adjacent the user’s nose when donning the HMD device 6-100.
- the eye/gaze sensors can also include bottom eye cameras 6-122 disposed below respective user eyes for capturing images of the eyes for facial avatar detection and creation, gaze tracking, and iris identification functions.
- the sensor system 6-102 can include infrared illuminators 6-124 pointed outward from the HMD device 6-100 to illuminate the external environment and any object therein with IR light for IR detection with one or more IR sensors of the sensor system 6-102.
- the sensor system 6-102 can include a flicker sensor 6-126 and an ambient light sensor 6-128.
- the flicker sensor 6-126 can detect overhead light refresh rates to avoid display flicker.
- the infrared illuminators 6-124 can include light emitting diodes and can be used especially for low light environments for illuminating user hands and other objects in low light for detection by infrared sensors of the sensor system 6-102.
- multiple sensors including the scene cameras 6-106, the downward cameras 6-114, the jaw cameras 6-116, the side cameras 6-118, the depth projector 6-112, and the depth sensors 6-108, 6-110 can be used in combination with an electrically coupled controller to combine depth data with camera data for hand tracking and for size determination for better hand tracking and object recognition and tracking functions of the HMD device 6-100.
- the downward cameras 6-114, jaw cameras 6-116, and side cameras 6-118 described above and shown in FIG. II can be wide angle cameras operable in the visible and infrared spectrums.
- these cameras 6-114, 6-116, 6-118 can operate only in black and white light detection to simplify image processing and gain sensitivity.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. II can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. 1 J - IL and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. 1 J - IL can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. II.
- FIG. 1 J illustrates a lower perspective view of an example of an HMD 6-200 including a cover or shroud 6-204 secured to a frame 6-230.
- the sensors 6-203 of the sensor system 6-202 can be disposed around a perimeter of the HDM 6- 200 such that the sensors 6-203 are outwardly disposed around a perimeter of a display region or area 6-232 so as not to obstruct a view of the displayed light.
- the sensors can be disposed behind the shroud 6-204 and aligned with transparent portions of the shroud allowing sensors and projectors to allow light back and forth through the shroud 6-204.
- opaque ink or other opaque material or films/layers can be disposed on the shroud 6-204 around the display area 6-232 to hide components of the HMD 6-200 outside the display area 6-232 other than the transparent portions defined by the opaque portions, through which the sensors and projectors send and receive light and electromagnetic signals during operation.
- the shroud 6-204 allows light to pass therethrough from the display (e.g., within the display region 6-232) but not radially outward from the display region around the perimeter of the display and shroud 6- 204.
- the shroud 6-204 includes a transparent portion 6-205 and an opaque portion 6-207, as described above and elsewhere herein.
- the opaque portion 6-207 of the shroud 6-204 can define one or more transparent regions 6- 209 through which the sensors 6-203 of the sensor system 6-202 can send and receive signals.
- the sensors 6-203 of the sensor system 6-202 sending and receiving signals through the shroud 6-204, or more specifically through the transparent regions 6-209 of the (or defined by) the opaque portion 6-207 of the shroud 6-204 can include the same or similar sensors as those shown in the example of FIG.
- depth sensors 6-108 and 6-110 for example depth sensors 6-108 and 6-110, depth projector 6-112, first and second scene cameras 6-106, first and second downward cameras 6-114, first and second side cameras 6-118, and first and second infrared illuminators 6-124.
- depth sensors 6-108 and 6-110 depth projector 6-112
- first and second scene cameras 6-106 first and second downward cameras 6-114
- first and second side cameras 6-118 first and second infrared illuminators 6-124.
- sensors are also shown in the examples of FIGS. IK and IL.
- Other sensors, sensor types, number of sensors, and relative positions thereof can be included in one or more other examples of HMDs.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. 1 J can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. II and IK - IL and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. II and IK - IL can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. 1 J.
- FIG. IK illustrates a front view of a portion of an example of an HMD device 6-300 including a display 6-334, brackets 6-336, 6-338, and frame or housing 6-330.
- the example shown in FIG. IK does not include a front cover or shroud in order to illustrate the brackets 6-336, 6-338.
- the shroud 6-204 shown in FIG. 1J includes the opaque portion 6-207 that would visually cover/block a view of anything outside (e.g., radially/peripherally outside) the di splay/di splay region 6-334, including the sensors 6-303 and bracket 6-338.
- the various sensors of the sensor system 6-302 are coupled to the brackets 6-336, 6-338.
- the scene cameras 6-306 include tight tolerances of angles relative to one another.
- the tolerance of mounting angles between the two scene cameras 6-306 can be 0.5 degrees or less, for example 0.3 degrees or less.
- the scene cameras 6-306 can be mounted to the bracket 6-338 and not the shroud.
- the bracket can include cantilevered arms on which the scene cameras 6-306 and other sensors of the sensor system 6-302 can be mounted to remain un-deformed in position and orientation in the case of a drop event by a user resulting in any deformation of the other bracket 6-226, housing 6-330, and/or shroud.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IK can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. II - 1 J and IL and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. II - 1 J and IL can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IK.
- FIG. IL illustrates a bottom view of an example of an HMD 6-400 including a front display/cover assembly 6-404 and a sensor system 6-402.
- the sensor system 6-402 can be similar to other sensor systems described above and elsewhere herein, including in reference to FIGS. II - IK.
- the jaw cameras 6-416 can be facing downward to capture images of the user’s lower facial features.
- the jaw cameras 6-416 can be coupled directly to the frame or housing 6-430 or one or more internal brackets directly coupled to the frame or housing 6-430 shown.
- the frame or housing 6-430 can include one or more apertures/openings 6-415 through which the jaw cameras 6-416 can send and receive signals.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IL can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. II - IK and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. II - IK can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IL.
- FIG. IM illustrates a rear perspective view of an inter-pupillary distance (IPD) adjustment system 11.1.1-102 including first and second optical modules 11.1. l-104a-b slidably engaging/coupled to respective guide-rods 11.1. l-108a-b and motors 11.1.1-1 lOa-b of left and right adjustment subsystems 11.1. l-106a-b.
- the IPD adjustment system 11.1.1- 102 can be coupled to a bracket 11.1.1-112 and include a button 11.1.1-114 in electrical communication with the motors 11.1.1-1 lOa-b.
- the button 11.1.1-114 can electrically communicate with the first and second motors 11.1.1-1 lOa-b via a processor or other circuitry components to cause the first and second motors 11.1.1-1 lOa-b to activate and cause the first and second optical modules 11.1.1-104a-b, respectively, to change position relative to one another.
- the first and second optical modules 11.1. l-104a-b can include respective display screens configured to project light toward the user’s eyes when donning the HMD 11.1.1-100.
- the user can manipulate (e.g., depress and/or rotate) the button 11.1.1-114 to activate a positional adjustment of the optical modules 11.1.1 - 104a-b to match the inter-pupillary distance of the user’s eyes.
- 11.1.1-104a-b can also include one or more cameras or other sensors/sensor systems for imaging and measuring the IPD of the user such that the optical modules 11.1. l-104a-b can be adjusted to match the IPD.
- the user can manipulate the button 11.1.1-114 to cause an automatic positional adjustment of the first and second optical modules 11.1. l-104a-b.
- the user can manipulate the button 11.1.1-114 to cause a manual adjustment such that the optical modules 11.1. l-104a-b move further or closer away, for example when the user rotates the button 11.1.1-114 one way or the other, until the user visually matches her/his own IPD.
- the manual adjustment is electronically communicated via one or more circuits and power for the movements of the optical modules 11.1. l-104a-b via the motors 11.1.1-1 lOa-b is provided by an electrical power source.
- the adjustment and movement of the optical modules 11.1. l-104a-b via a manipulation of the button 11.1.1-114 is mechanically actuated via the movement of the button 11.1.1-114.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IM can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in any other figures shown and described herein.
- FIG. IN illustrates a front perspective view of a portion of an HMD 11.1.2- 100, including an outer structural frame 11.1.2-102 and an inner or intermediate structural frame 11.1.2-104 defining first and second apertures 11.1.2- 106a, 11.1.2- 106b.
- the HMD 11.1.2-100 can include a first mounting bracket 11.1.2-108 coupled to the inner frame 11.1.2-104.
- the mounting bracket 11.1.2-108 is coupled to the inner frame 11.1.2-104 between the first and second apertures 11.1 ,2-106a-b.
- the mounting bracket 11.1.2-108 can include a middle or central portion
- the middle or central portion 11.1.2-109 may not be the geometric middle or center of the bracket 11.1.2-108. Rather, the middle/central portion 11.1.2-109 can be disposed between first and second cantilevered extension arms extending away from the middle portion 11.1.2-109.
- the mounting bracket 108 includes a first cantilever arm 11.1.2-112 and a second cantilever arm 11.1.2-114 extending away from the middle portion 11.1.2-109 of the mount bracket 11.1.2-108 coupled to the inner frame 11.1.2-104.
- the outer frame 11.1.2-102 can define a curved geometry on a lower side thereof to accommodate a user’s nose when the user dons the HMD
- the curved geometry can be referred to as a nose bridge 11.1.2-111 and be centrally located on a lower side of the HMD 11.1.2-100 as shown.
- the mounting bracket 11.1.2-108 can be connected to the inner frame 11.1.2-104 between the apertures 11.1 ,2-106a-b such that the cantilevered arms 11.1.2-112, 11.1.2-114 extend downward and laterally outward away from the middle portion 11.1.2-109 to compliment the nose bridge 11.1.2-111 geometry of the outer frame 11.1.2-102.
- the mounting bracket 11.1.2-108 is configured to accommodate the user’s nose as noted above.
- the nose bridge 11.1.2-111 geometry accommodates the nose in that the nose bridge 11.1.2-111 provides a curvature that curves with, above, over, and around the user’s nose for comfort and fit.
- the first cantilever arm 11.1.2-112 can extend away from the middle portion
- the first and second cantilever arms 11.1.2-112, 11.1.2-114 are referred to as “cantilevered” or “cantilever” arms because each arm 11.1.2-112, 11.1.2-114, includes a distal free end 11.1.2-116, 11.1.2-118, respectively, which are free of affixation from the inner and outer frames 11.1.2-102, 11.1.2- 104. In this way, the arms 11.1.2-112, 11.1.2-114 are cantilevered from the middle portion
- 11.1.2-109 which can be connected to the inner frame 11.1.2-104, with distal ends 11.1.2- 102, 11.1.2-104 unattached.
- the HMD 11.1.2-100 can include one or more components coupled to the mounting bracket 11.1.2-108.
- the components include a plurality of sensors 11.1.2-1 lOa-f.
- Each sensor of the plurality of sensors 11.1.2- 1 lOa-f can include various types of sensors, including cameras, IR sensors, and so forth.
- one or more of the sensors 11.1.2-1 lOa-f can be used for object recognition in three-dimensional space such that it is important to maintain a precise relative position of two or more of the plurality of sensors 11.1.2-1 lOa-f.
- the cantilevered nature of the mounting bracket 11.1.2-108 can protect the sensors 11.1.2-1 lOa-f from damage and altered positioning in the case of accidental drops by the user. Because the sensors 11.1.2-1 lOa-f are cantilevered on the arms 11.1.2-112, 11.1.2-114 of the mounting bracket 11.1.2-108, stresses and deformations of the inner and/or outer frames 11.1.2-104, 11.1.2-102 are not transferred to the cantilevered arms 11.1.2-112, 11.1.2-114 and thus do not affect the relative positioning of the sensors 11.1.2-1 lOa-f coupled/mounted to the mounting bracket 11.1.2-108.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described herein can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IN.
- FIG. 10 illustrates an example of an optical module 11.3.2-100 for use in an electronic device such as an HMD, including HDM devices described herein.
- the optical module 11.3.2-100 can be one of two optical modules within an HMD, with each optical module aligned to project light toward a user’s eye.
- a first optical module can project light via a display screen toward a user’s first eye
- a second optical module of the same device can project light via another display screen toward the user’s second eye.
- the optical module 11.3.2-100 can include an optical frame or housing 11.3.2-102, which can also be referred to as a barrel or optical module barrel.
- the optical module 11.3.2-100 can also include a display 11.3.2-104, including a display screen or multiple display screens, coupled to the housing 11.3.2-102.
- the display 11.3.2-104 can be coupled to the housing 11.3.2-102 such that the display 11.3.2-104 is configured to project light toward the eye of a user when the HMD of which the display module 11.3.2-100 is a part is donned during use.
- the housing 11.3.2- 102 can surround the display 11.3.2-104 and provide connection features for coupling other components of optical modules described herein.
- the optical module 11.3.2-100 can include one or more cameras 11.3.2-106 coupled to the housing 11.3.2-102.
- the camera 11.3.2-106 can be positioned relative to the display 11.3.2-104 and housing 11.3.2-102 such that the camera 11.3.2-106 is configured to capture one or more images of the user’s eye during use.
- the optical module 11.3.2-100 can also include a light strip 11.3.2-108 surrounding the display 11.3.2-104.
- the light strip 11.3.2-108 is disposed between the display 11.3.2-104 and the camera 11.3.2-106.
- the light strip 11.3.2-108 can include a plurality of lights 11.3.2- 110.
- the plurality of lights can include one or more light emitting diodes (LEDs) or other lights configured to project light toward the user’s eye when the HMD is donned.
- the individual lights 11.3.2-110 of the light strip 11.3.2-108 can be spaced about the strip 11.3.2-108 and thus spaced about the display 11.3.2-104 uniformly or non-uniformly at various locations on the strip 11.3.2-108 and around the display 11.3.2-104.
- the housing 11.3.2-102 defines a viewing opening
- the LEDs are configured and arranged to emit light through the viewing opening 11.3.2-101 and onto the user’s eye.
- the camera 11.3.2- 106 is configured to capture one or more images of the user’s eye through the viewing opening 11.3.2-101.
- 11.3.2-100 shown in FIG. 10 can be replicated in another (e.g., second) optical module disposed with the HMD to interact (e.g., project light and capture images) of another eye of the user.
- another optical module disposed with the HMD to interact (e.g., project light and capture images) of another eye of the user.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. 10 can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts shown in FIGS. IP or otherwise described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described with reference to FIGS. IP or otherwise described herein can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. 10.
- FIG. IP illustrates a cross-sectional view of an example of an optical module
- the housing 11.3.2-202 defines a first aperture or channel 11.3.2-212 and a second aperture or channel 11.3.2-214.
- the channels 11.3.2-212, 11.3.2-214 can be configured to slidably engage respective rails or guide rods of an HMD device to allow the optical module
- the housing 11.3.2-202 can slidably engage the guide rods to secure the optical module 11.3.2-200 in place within the HMD.
- the optical module 11.3.2-200 can also include a lens
- the lens 11.3.2-216 can be configured to direct light from the display assembly 11.3.2-204 to the user’s eye.
- the lens 11.3.2-216 can be a part of a lens assembly including a corrective lens removably attached to the optical module 11.3.2-200.
- the lens 11.3.2- 216 is disposed over the light strip 11.3.2-208 and the one or more eye-tracking cameras
- the camera 11.3.2-206 is configured to capture images of the user’s eye through the lens 11.3.2-216 and the light strip 11.3.2-208 includes lights configured to project light through the lens 11.3.2-216 to the users’ eye during use.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown in FIG. IP can be included, either alone or in any combination, in any of the other examples of devices, features, components, and parts and described herein.
- any of the features, components, and/or parts, including the arrangements and configurations thereof shown and described herein can be included, either alone or in any combination, in the example of the devices, features, components, and parts shown in FIG. IP.
- FIG. 2 is a block diagram of an example of the controller 110 in accordance with some embodiments. While certain specific features are illustrated, those skilled in the art will appreciate from the present disclosure that various other features have not been illustrated for the sake of brevity, and so as not to obscure more pertinent aspects of the embodiments disclosed herein.
- the controller 110 includes one or more processing units 202 (e.g., microprocessors, application-specific integrated-circuits (ASICs), field-programmable gate arrays (FPGAs), graphics processing units (GPUs), central processing units (CPUs), processing cores, and/or the like), one or more input/output (I/O) devices 206, one or more communication interfaces 208 (e g., universal serial bus (USB), FIREWIRE, THUNDERBOLT, IEEE 802.3x, IEEE 802.1 lx, IEEE 802.16x, global system for mobile communications (GSM), code division multiple access (CDMA), time division multiple access (TDMA), global positioning system (GPS), infrared (IR), BLUETOOTH, ZIGBEE, and/or the like type interface), one or more programming (e.g., I/O) interfaces 210, a memory 220, and one or more communication buses 204 for interconnecting these and various processing units 202 (e.g., microprocessors, application
- the one or more communication buses 204 include circuitry that interconnects and controls communications between system components.
- the one or more I/O devices 206 include at least one of a keyboard, a mouse, a touchpad, a joystick, one or more microphones, one or more speakers, one or more image sensors, one or more displays, and/or the like.
- the memory 220 includes high-speed random-access memory, such as dynamic random-access memory (DRAM), static random-access memory (SRAM), double- data-rate random-access memory (DDR RAM), or other random-access solid-state memory devices.
- the memory 220 includes non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices.
- the memory 220 optionally includes one or more storage devices remotely located from the one or more processing units 202.
- the memory 220 comprises a non-transitory computer readable storage medium.
- the memory 220 or the non-transitory computer readable storage medium of the memory 220 stores the following programs, modules and data structures, or a subset thereof including an optional operating system 230 and an XR experience module 240.
- the operating system 230 includes instructions for handling various basic system services and for performing hardware dependent tasks.
- the XR experience module 240 is configured to manage and coordinate one or more XR experiences for one or more users (e.g., a single XR experience for one or more users, or multiple XR experiences for respective groups of one or more users).
- the XR experience module 240 includes a data obtaining unit 242, a tracking unit 244, a coordination unit 246, and a data transmitting unit 248.
- the data obtaining unit 242 is configured to obtain data (e.g., presentation data, interaction data, sensor data, location data, etc.) from at least the display generation component 120 of Figure 1 A, and optionally one or more of the input devices 125, output devices 155, sensors 190, and/or peripheral devices 195.
- the data obtaining unit 242 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the tracking unit 244 is configured to map the scene 105 and to track the position/location of at least the display generation component 120 with respect to the scene 105 of Figure 1 A, and optionally, to one or more of the input devices 125, output devices 155, sensors 190, and/or peripheral devices 195.
- the tracking unit 244 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the tracking unit 244 includes hand tracking unit 245 and/or eye tracking unit 243.
- the hand tracking unit 245 is configured to track the position/location of one or more portions of the user’s hands, and/or motions of one or more portions of the user’s hands with respect to the scene 105 of Figure 1 A, relative to the display generation component 120, and/or relative to a coordinate system defined relative to the user’s hand.
- the hand tracking unit 245 is described in greater detail below with respect to Figure 4.
- the eye tracking unit 243 is configured to track the position and movement of the user’s gaze (or more broadly, the user’s eyes, face, or head) with respect to the scene 105 (e.g., with respect to the physical environment and/or to the user (e.g., the user’s hand)) or with respect to the XR content displayed via the display generation component 120.
- the eye tracking unit 243 is described in greater detail below with respect to Figure 5.
- the coordination unit 246 is configured to manage and coordinate the XR experience presented to the user by the display generation component 120, and optionally, by one or more of the output devices 155 and/or peripheral devices 195. To that end, in various embodiments, the coordination unit 246 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the data transmitting unit 248 is configured to transmit data (e.g., presentation data, location data, etc.) to at least the display generation component 120, and optionally, to one or more of the input devices 125, output devices 155, sensors 190, and/or peripheral devices 195.
- data transmitting unit 248 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the data obtaining unit 242, the tracking unit 244 (e.g., including the eye tracking unit 243 and the hand tracking unit 245), the coordination unit 246, and the data transmitting unit 248 are shown as residing on a single device (e.g., the controller 110), it should be understood that in other embodiments, any combination of the data obtaining unit 242, the tracking unit 244 (e.g., including the eye tracking unit 243 and the hand tracking unit 245), the coordination unit 246, and the data transmitting unit 248 may be located in separate computing devices.
- Figure 2 is intended more as functional description of the various features that may be present in a particular implementation as opposed to a structural schematic of the embodiments described herein.
- FIG. 3 is a block diagram of an example of the display generation component 120 in accordance with some embodiments. While certain specific features are illustrated, those skilled in the art will appreciate from the present disclosure that various other features have not been illustrated for the sake of brevity, and so as not to obscure more pertinent aspects of the embodiments disclosed herein.
- the display generation component 120 includes one or more processing units 302 (e.g., microprocessors, ASICs, FPGAs, GPUs, CPUs, processing cores, and/or the like), one or more input/output (I/O) devices and sensors 306, one or more communication interfaces 308 (e g., USB, FIREWIRE, THUNDERBOLT, IEEE 802.3x, IEEE 802.1 lx, IEEE 802.16x, GSM, CDMA, TDMA, GPS, IR, BLUETOOTH, ZIGBEE, and/or the like type interface), one or more programming (e.g., I/O) interfaces 310, one or more XR displays 312, one or more optional interior- and/or exterior-facing image sensors 314, a memory 320, and one or more communication buses 304 for interconnecting these and various other components.
- processing units 302 e.g., microprocessors, ASICs, FPGAs, GPUs, CPUs, processing cores, and/or the like
- the one or more communication buses 304 include circuitry that interconnects and controls communications between system components.
- the one or more I/O devices and sensors 306 include at least one of an inertial measurement unit (IMU), an accelerometer, a gyroscope, a thermometer, one or more physiological sensors (e.g., blood pressure monitor, heart rate monitor, blood oxygen sensor, blood glucose sensor, etc.), one or more microphones, one or more speakers, a haptics engine, one or more depth sensors (e.g., a structured light, a time-of-flight, or the like), and/or the like.
- IMU inertial measurement unit
- the one or more XR displays 312 are configured to provide the XR experience to the user.
- the one or more XR displays 312 correspond to holographic, digital light processing (DLP), liquid-crystal display (LCD), liquid-crystal on silicon (LcoS), organic light-emitting field-effect transistor (OLET), organic light-emitting diode (OLED), surface-conduction electron-emitter display (SED), fieldemission display (FED), quantum-dot light-emitting diode (QD-LED), micro-electro- mechanical system (MEMS), and/or the like display types.
- DLP digital light processing
- LCD liquid-crystal display
- LcoS liquid-crystal on silicon
- OLET organic light-emitting field-effect transistor
- OLED organic light-emitting diode
- SED surface-conduction electron-emitter display
- FED fieldemission display
- QD-LED quantum-dot light-emitting dio
- the one or more XR displays 312 correspond to diffractive, reflective, polarized, holographic, etc. waveguide displays.
- the display generation component 120 e.g., HMD
- the display generation component 120 includes a single XR display.
- the display generation component 120 includes an XR display for each eye of the user.
- the one or more XR displays 312 are capable of presenting MR and VR content.
- the one or more XR displays 312 are capable of presenting MR or VR content.
- the one or more image sensors 314 are configured to obtain image data that corresponds to at least a portion of the face of the user that includes the eyes of the user (and may be referred to as an eye-tracking camera). In some embodiments, the one or more image sensors 314 are configured to obtain image data that corresponds to at least a portion of the user’s hand(s) and optionally arm(s) of the user (and may be referred to as a hand-tracking camera).
- the one or more image sensors 314 are configured to be forward-facing so as to obtain image data that corresponds to the scene as would be viewed by the user if the display generation component 120 (e.g., HMD) was not present (and may be referred to as a scene camera).
- the one or more optional image sensors 314 can include one or more RGB cameras (e.g., with a complimentary metal-oxide- semiconductor (CMOS) image sensor or a charge-coupled device (CCD) image sensor), one or more infrared (IR) cameras, one or more event-based cameras, and/or the like.
- CMOS complimentary metal-oxide- semiconductor
- CCD charge-coupled device
- the memory 320 includes high-speed random-access memory, such as DRAM, SRAM, DDR RAM, or other random-access solid-state memory devices.
- the memory 320 includes non-volatile memory, such as one or more magnetic disk storage devices, optical disk storage devices, flash memory devices, or other non-volatile solid-state storage devices.
- the memory 320 optionally includes one or more storage devices remotely located from the one or more processing units 302.
- the memory 320 comprises a non-transitory computer readable storage medium.
- the memory 320 or the non-transitory computer readable storage medium of the memory 320 stores the following programs, modules and data structures, or a subset thereof including an optional operating system 330 and an XR presentation module 340.
- the operating system 330 includes instructions for handling various basic system services and for performing hardware dependent tasks.
- the XR presentation module 340 is configured to present XR content to the user via the one or more XR displays 312.
- the XR presentation module 340 includes a data obtaining unit 342, an XR presenting unit 344, an XR map generating unit 346, and a data transmitting unit 348.
- the data obtaining unit 342 is configured to obtain data (e.g., presentation data, interaction data, sensor data, location data, etc.) from at least the controller 110 of Figure 1.
- data e.g., presentation data, interaction data, sensor data, location data, etc.
- the data obtaining unit 342 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the XR presenting unit 344 is configured to present XR content via the one or more XR displays 312. To that end, in various embodiments, the XR presenting unit 344 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the XR map generating unit 346 is configured to generate an XR map (e.g., a 3D map of the mixed reality scene or a map of the physical environment into which computer-generated objects can be placed to generate the extended reality) based on media content data.
- the XR map generating unit 346 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- the data transmitting unit 348 is configured to transmit data (e.g., presentation data, location data, etc.) to at least the controller 110, and optionally one or more of the input devices 125, output devices 155, sensors 190, and/or peripheral devices 195.
- the data transmitting unit 348 includes instructions and/or logic therefor, and heuristics and metadata therefor.
- FIG. 342 Although the data obtaining unit 342, the XR presenting unit 344, the XR map generating unit 346, and the data transmitting unit 348 are shown as residing on a single device (e.g., the display generation component 120 of Figure 1), it should be understood that in other embodiments, any combination of the data obtaining unit 342, the XR presenting unit 344, the XR map generating unit 346, and the data transmitting unit 348 may be located in separate computing devices.
- Figure 3 is intended more as a functional description of the various features that could be present in a particular implementation as opposed to a structural schematic of the embodiments described herein. As recognized by those of ordinary skill in the art, items shown separately could be combined and some items could be separated.
- Figure 4 is a schematic, pictorial illustration of an example embodiment of the hand tracking device 140.
- hand tracking device 140 ( Figure 1) is controlled by hand tracking unit 245 ( Figure 2) to track the position/location of one or more portions of the user’s hands, and/or motions of one or more portions of the user’s hands with respect to the scene 105 of Figure 1 A (e.g., with respect to a portion of the physical environment surrounding the user, with respect to the display generation component 120, or with respect to a portion of the user (e.g., the user’s face, eyes, or head), and/or relative to a coordinate system defined relative to the user’s hand.
- the hand tracking device 140 is part of the display generation component 120 (e.g., embedded in or attached to a head-mounted device). In some embodiments, the hand tracking device 140 is separate from the display generation component 120 (e.g., located in separate housings or attached to separate physical support structures).
- the hand tracking device 140 includes image sensors 404 (e.g., one or more IR cameras, 3D cameras, depth cameras, and/or color cameras, etc.) that capture three-dimensional scene information that includes at least a hand 406 of a human user.
- the image sensors 404 capture the hand images with sufficient resolution to enable the fingers and their respective positions to be distinguished.
- the image sensors 404 typically capture images of other parts of the user’s body, as well, or possibly all of the body, and may have either zoom capabilities or a dedicated sensor with enhanced magnification to capture images of the hand with the desired resolution.
- the image sensors 404 also capture 2D color video images of the hand 406 and other elements of the scene.
- the image sensors 404 are used in conjunction with other image sensors to capture the physical environment of the scene 105, or serve as the image sensors that capture the physical environment of the scene 105. In some embodiments, the image sensors 404 are positioned relative to the user or the user’s environment in a way that a field of view of the image sensors or a portion thereof is used to define an interaction space in which hand movement captured by the image sensors are treated as inputs to the controller 110.
- the image sensors 404 output a sequence of frames containing 3D map data (and possibly color image data, as well) to the controller 110, which extracts high-level information from the map data.
- This high-level information is typically provided via an Application Program Interface (API) to an application running on the controller, which drives the display generation component 120 accordingly.
- API Application Program Interface
- the user may interact with software running on the controller 110 by moving their hand 406 and/or changing their hand posture.
- the image sensors 404 project a pattern of spots onto a scene containing the hand 406 and capture an image of the projected pattern.
- the controller 110 computes the 3D coordinates of points in the scene (including points on the surface of the user’s hand) by triangulation, based on transverse shifts of the spots in the pattern. This approach is advantageous in that it does not require the user to hold or wear any sort of beacon, sensor, or other marker. It gives the depth coordinates of points in the scene relative to a predetermined reference plane, at a certain distance from the image sensors 404.
- the image sensors 404 are assumed to define an orthogonal set of x, y, z axes, so that depth coordinates of points in the scene correspond to z components measured by the image sensors.
- the image sensors 404 e.g., a hand tracking device
- the hand tracking device 140 captures and processes a temporal sequence of depth maps containing the user’s hand, while the user moves their hand (e.g., whole hand or one or more fingers).
- Software running on a processor in the image sensors 404 and/or the controller 110 processes the 3D map data to extract patch descriptors of the hand in these depth maps.
- the software matches these descriptors to patch descriptors stored in a database 408, based on a prior learning process, in order to estimate the pose of the hand in each frame.
- the pose typically includes 3D locations of the user’s hand joints and fingertips.
- the software may also analyze the trajectory of the hands and/or fingers over multiple frames in the sequence in order to identify gestures.
- the pose estimation functions described herein may be interleaved with motion tracking functions, so that patch-based pose estimation is performed only once in every two (or more) frames, while tracking is used to find changes in the pose that occur over the remaining frames.
- the pose, motion, and gesture information are provided via the above-mentioned API to an application program running on the controller 110. This program may, for example, move and modify images presented on the display generation component 120, or perform other functions, in response to the pose and/or gesture information.
- a gesture includes an air gesture.
- An air gesture is a gesture that is detected without the user touching (or independently of) an input element that is part of a device (e.g., computer system 101, one or more input device 125, and/or hand tracking device 140) and is based on detected motion of a portion (e.g., the head, one or more arms, one or more hands, one or more fingers, and/or one or more legs) of the user’s body through the air including motion of the user’s body relative to an absolute reference (e.g., an angle of the user’s arm relative to the ground or a distance of the user’s hand relative to the ground), relative to another portion of the user’s body (e.g., movement of a hand of the user relative to a shoulder of the user, movement of one hand of the user relative to another hand of the user, and/or movement of a finger of the user relative to another finger or portion of a hand of the user), and/or absolute motion of a portion of the user
- input gestures used in the various examples and embodiments described herein include air gestures performed by movement of the user’s finger(s) relative to other finger(s) or part(s) of the user’s hand) for interacting with an XR environment (e.g., a virtual or mixed-reality environment), in accordance with some embodiments.
- XR environment e.g., a virtual or mixed-reality environment
- an air gesture is a gesture that is detected without the user touching an input element that is part of the device (or independently of an input element that is a part of the device) and is based on detected motion of a portion of the user’s body through the air including motion of the user’s body relative to an absolute reference (e.g., an angle of the user’s arm relative to the ground or a distance of the user’s hand relative to the ground), relative to another portion of the user’s body (e.g., movement of a hand of the user relative to a shoulder of the user, movement of one hand of the user relative to another hand of the user, and/or movement of a finger of the user relative to another finger or portion of a hand of the user), and/or absolute motion of a portion of the user’s body (e.g., a tap gesture that includes movement of a hand in a predetermined pose by a predetermined amount and/or speed, or a shake gesture that includes a predetermined speed or amount of rotation of a portion of the user
- the input gesture is an air gesture (e.g., in the absence of physical contact with an input device that provides the computer system with information about which user interface element is the target of the user input, such as contact with a user interface element displayed on a touchscreen, or contact with a mouse or trackpad to move a cursor to the user interface element), the gesture takes into account the user’s attention (e.g., gaze) to determine the target of the user input (e.g., for direct inputs, as described below).
- attention e.g., gaze
- the input gesture is, for example, detected attention (e.g., gaze) toward the user interface element in combination (e.g., concurrent) with movement of a user’s finger(s) and/or hands to perform a pinch and/or tap input, as described in more detail below.
- detected attention e.g., gaze
- the user interface element in combination (e.g., concurrent) with movement of a user’s finger(s) and/or hands to perform a pinch and/or tap input, as described in more detail below.
- input gestures that are directed to a user interface object are performed directly or indirectly with reference to a user interface object.
- a user input is performed directly on the user interface object in accordance with performing the input gesture with the user’s hand at a position that corresponds to the position of the user interface object in the three-dimensional environment (e.g., as determined based on a current viewpoint of the user).
- the input gesture is performed indirectly on the user interface object in accordance with the user performing the input gesture while a position of the user’s hand is not at the position that corresponds to the position of the user interface object in the three-dimensional environment while detecting the user’s attention (e.g., gaze) on the user interface object.
- attention e.g., gaze
- the user is enabled to direct the user’s input to the user interface object by initiating the gesture at, or near, a position corresponding to the displayed position of the user interface object (e.g., within 0.5 cm, 1 cm, 5 cm, or a distance between 0-5 cm, as measured from an outer edge of the option or a center portion of the option).
- a position corresponding to the displayed position of the user interface object e.g., within 0.5 cm, 1 cm, 5 cm, or a distance between 0-5 cm, as measured from an outer edge of the option or a center portion of the option.
- input gestures used in the various examples and embodiments described herein include pinch inputs and tap inputs, for interacting with a virtual or mixed-reality environment, in accordance with some embodiments.
- the pinch inputs and tap inputs described below are performed as air gestures.
- a pinch input is part of an air gesture that includes one or more of: a pinch gesture, a long pinch gesture, a pinch and drag gesture, or a double pinch gesture.
- a pinch gesture that is an air gesture includes movement of two or more fingers of a hand to make contact with one another, that is, optionally, followed by an immediate (e.g., within 0-1 seconds) break in contact from each other.
- a long pinch gesture that is an air gesture includes movement of two or more fingers of a hand to make contact with one another for at least a threshold amount of time (e.g., at least 1 second), before detecting a break in contact with one another.
- a long pinch gesture includes the user holding a pinch gesture (e.g., with the two or more fingers making contact), and the long pinch gesture continues until a break in contact between the two or more fingers is detected.
- a double pinch gesture that is an air gesture comprises two (e.g., or more) pinch inputs (e.g., performed by the same hand) detected in immediate (e.g., within a predefined time period) succession of each other.
- the user performs a first pinch input (e.g., a pinch input or a long pinch input), releases the first pinch input (e.g., breaks contact between the two or more fingers), and performs a second pinch input within a predefined time period (e.g., within 1 second or within 2 seconds) after releasing the first pinch input.
- a first pinch input e.g., a pinch input or a long pinch input
- releases the first pinch input e.g., breaks contact between the two or more fingers
- a second pinch input within a predefined time period (e.g., within 1 second or within 2 seconds) after releasing the first pinch input.
- a pinch and drag gesture that is an air gesture includes a pinch gesture (e.g., a pinch gesture or a long pinch gesture) performed in conjunction with (e.g., followed by) a drag input that changes a position of the user’s hand from a first position (e.g., a start position of the drag) to a second position (e.g., an end position of the drag).
- a pinch gesture e.g., a pinch gesture or a long pinch gesture
- the user maintains the pinch gesture while performing the drag input, and releases the pinch gesture (e.g., opens their two or more fingers) to end the drag gesture (e.g., at the second position).
- the pinch input and the drag input are performed by the same hand (e.g., the user pinches two or more fingers to make contact with one another and moves the same hand to the second position in the air with the drag gesture).
- the pinch input is performed by a first hand of the user and the drag input is performed by the second hand of the user (e.g., the user’s second hand moves from the first position to the second position in the air while the user continues the pinch input with the user’s first hand.
- an input gesture that is an air gesture includes inputs (e.g., pinch and/or tap inputs) performed using both of the user’s two hands.
- the input gesture includes two (e.g., or more) pinch inputs performed in conjunction with (e.g., concurrently with, or within a predefined time period of) each other.
- a first pinch gesture is performed using a first hand of the user (e.g., a pinch input, a long pinch input, or a pinch and drag input), and, in conjunction with performing the pinch input using the first hand, a second pinch input is performed using the other hand (e.g., the second hand of the user’s two hands).
- movement between the user’s two hands is performed (e.g., to increase and/or decrease a distance or relative orientation between the user’s two hands).
- a tap input (e.g., directed to a user interface element) performed as an air gesture includes movement of a user’s finger(s) toward the user interface element, movement of the user’s hand toward the user interface element optionally with the user’s finger(s) extended toward the user interface element, a downward motion of a user’s finger (e.g., mimicking a mouse click motion or a tap on a touchscreen), or other predefined movement of the user’s hand.
- a tap input that is performed as an air gesture is detected based on movement characteristics of the finger or hand performing the tap gesture movement of a finger or hand away from the viewpoint of the user and/or toward an object that is the target of the tap input followed by an end of the movement.
- the end of the movement is detected based on a change in movement characteristics of the finger or hand performing the tap gesture (e.g., an end of movement away from the viewpoint of the user and/or toward the object that is the target of the tap input, a reversal of direction of movement of the finger or hand, and/or a reversal of a direction of acceleration of movement of the finger or hand).
- a change in movement characteristics of the finger or hand performing the tap gesture e.g., an end of movement away from the viewpoint of the user and/or toward the object that is the target of the tap input, a reversal of direction of movement of the finger or hand, and/or a reversal of a direction of acceleration of movement of the finger or hand.
- attention of a user is determined to be directed to a portion of the three-dimensional environment based on detection of gaze directed to the portion of the three-dimensional environment (optionally, without requiring other conditions).
- attention of a user is determined to be directed to a portion of the three-dimensional environment based on detection of gaze directed to the portion of the three-dimensional environment with one or more additional conditions such as requiring that gaze is directed to the portion of the three-dimensional environment for at least a threshold duration (e.g., a dwell duration) and/or requiring that the gaze is directed to the portion of the three-dimensional environment while the viewpoint of the user is within a distance threshold from the portion of the three-dimensional environment in order for the device to determine that attention of the user is directed to the portion of the three- dimensional environment, where if one of the additional conditions is not met, the device determines that attention is not directed to the portion of the three-dimensional environment toward which gaze is directed (e.g., until the one or more additional conditions are met).
- a threshold duration e.g.,
- the detection of a ready state configuration of a user or a portion of a user is detected by the computer system.
- Detection of a ready state configuration of a hand is used by a computer system as an indication that the user is likely preparing to interact with the computer system using one or more air gesture inputs performed by the hand (e.g., a pinch, tap, pinch and drag, double pinch, long pinch, or other air gesture described herein).
- the ready state of the hand is determined based on whether the hand has a predetermined hand shape (e.g., a pre-pinch shape with a thumb and one or more fingers extended and spaced apart ready to make a pinch or grab gesture or a pretap with one or more fingers extended and palm facing away from the user), based on whether the hand is in a predetermined position relative to a viewpoint of the user (e.g., below the user’s head and above the user’s waist and extended out from the body by at least 15, 20, 25, 30, or 50 cm), and/or based on whether the hand has moved in a particular manner (e.g., moved toward a region in front of the user above the user’s waist and below the user’s head or moved away from the user’s body or leg).
- the ready state is used to determine whether interactive elements of the user interface respond to attention (e.g., gaze) inputs.
- user inputs can be detected with controls contained in the hardware input device such as one or more touch-sensitive input elements, one or more pressure-sensitive input elements, one or more buttons, one or more knobs, one or more dials, one or more joysticks, one or more hand or finger coverings that can detect a position or change in position of portions of a hand and/or fingers relative to each other, relative to the user’s body, and/or relative to a physical environment of the user, and/or other hardware input device controls, wherein the user inputs with the controls contained in the hardware input device are used in place of hand and/or finger gestures such as air taps or air pinches in the corresponding air gesture(s).
- controls contained in the hardware input device such as one or more touch-sensitive input elements, one or more pressure-sensitive input elements, one or more buttons, one or more knobs, one or more dials, one or more joysticks, one or more hand or finger coverings that can detect a position or change in position of portions of a hand and/or fingers relative to each other, relative
- a selection input that is described as being performed with an air tap or air pinch input could be alternatively detected with a button press, a tap on a touch-sensitive surface, a press on a pressure-sensitive surface, or other hardware input.
- a movement input that is described as being performed with an air pinch and drag e.g., an air drag gesture or an air swipe gesture
- the hardware input control such as a button press and hold, a touch on a touch-sensitive surface, a press on a pressure-sensitive surface, or other hardware input that is followed by movement of the hardware input device (e.g., along with the hand with which the hardware input device is associated) through space.
- a two-handed input that includes movement of the hands relative to each other could be performed with one air gesture and one hardware input device in the hand that is not performing the air gesture, two hardware input devices held in different hands, or two air gestures performed by different hands using various combinations of air gestures and/or the inputs detected by one or more hardware input devices that are described above.
- the software may be downloaded to the controller 110 in electronic form, over a network, for example, or it may alternatively be provided on tangible, non-transitory media, such as optical, magnetic, or electronic memory media.
- the database 408 is likewise stored in a memory associated with the controller 110.
- some or all of the described functions of the computer may be implemented in dedicated hardware, such as a custom or semi-custom integrated circuit or a programmable digital signal processor (DSP).
- DSP programmable digital signal processor
- controller 110 is shown in Figure 4, by way of example, as a separate unit from the image sensors 404, some or all of the processing functions of the controller may be performed by a suitable microprocessor and software or by dedicated circuitry within the housing of the image sensors 404 (e.g., a hand tracking device) or otherwise associated with the image sensors 404. In some embodiments, at least some of these processing functions may be carried out by a suitable processor that is integrated with the display generation component 120 (e.g., in a television set, a handheld device, or head-mounted device, for example) or with any other suitable computerized device, such as a game console or media player.
- the sensing functions of image sensors 404 may likewise be integrated into the computer or other computerized apparatus that is to be controlled by the sensor output.
- Figure 4 further includes a schematic representation of a depth map 410 captured by the image sensors 404, in accordance with some embodiments.
- the depth map as explained above, comprises a matrix of pixels having respective depth values.
- the pixels 412 corresponding to the hand 406 have been segmented out from the background and the wrist in this map.
- the brightness of each pixel within the depth map 410 corresponds inversely to its depth value, i.e., the measured z distance from the image sensors 404, with the shade of gray growing darker with increasing depth.
- the controller 110 processes these depth values in order to identify and segment a component of the image (i.e., a group of neighboring pixels) having characteristics of a human hand. These characteristics, may include, for example, overall size, shape and motion from frame to frame of the sequence of depth maps.
- Figure 4 also schematically illustrates a hand skeleton 414 that controller 110 ultimately extracts from the depth map 410 of the hand 406, in accordance with some embodiments.
- the hand skeleton 414 is superimposed on a hand background 416 that has been segmented from the original depth map.
- key feature points of the hand e.g., points corresponding to knuckles, fingertips, center of the palm, end of the hand connecting to wrist, etc.
- location and movements of these key feature points over multiple image frames are used by the controller 110 to determine the hand gestures performed by the hand or the current state of the hand, in accordance with some embodiments.
- Figure 5 illustrates an example embodiment of the eye tracking device 130 ( Figure 1).
- the eye tracking device 130 is controlled by the eye tracking unit 243 ( Figure 2) to track the position and movement of the user’s gaze with respect to the scene 105 or with respect to the XR content displayed via the display generation component 120.
- the eye tracking device 130 is integrated with the display generation component 120.
- the display generation component 120 is a head-mounted device such as headset, helmet, goggles, or glasses, or a handheld device placed in a wearable frame
- the head-mounted device includes both a component that generates the XR content for viewing by the user and a component for tracking the gaze of the user relative to the XR content.
- the eye tracking device 130 is separate from the display generation component 120.
- the eye tracking device 130 is optionally a separate device from the handheld device or XR chamber.
- the eye tracking device 130 is a head-mounted device or part of a head-mounted device.
- the head-mounted eye-tracking device 130 is optionally used in conjunction with a display generation component that is also headmounted, or a display generation component that is not head-mounted.
- the eye tracking device 130 is not a head-mounted device, and is optionally used in conjunction with a head-mounted display generation component.
- the eye tracking device 130 is not a head-mounted device, and is optionally part of a non-head- mounted display generation component.
- the display generation component 120 uses a display mechanism (e.g., left and right near-eye display panels) for displaying frames including left and right images in front of a user’s eyes to thus provide 3D virtual views to the user.
- a head-mounted display generation component may include left and right optical lenses (referred to herein as eye lenses) located between the display and the user’s eyes.
- the display generation component may include or be coupled to one or more external video cameras that capture video of the user’s environment for display.
- a head-mounted display generation component may have a transparent or semi-transparent display through which a user may view the physical environment directly and display virtual objects on the transparent or semi-transparent display.
- display generation component projects virtual objects into the physical environment.
- the virtual objects may be projected, for example, on a physical surface or as a holograph, so that an individual, using the system, observes the virtual objects superimposed over the physical environment. In such cases, separate display panels and image frames for the left and right eyes may not be necessary.
- eye tracking device 130 e.g., a gaze tracking device
- eye tracking camera e.g., infrared (IR) or near-IR (NIR) cameras
- illumination sources e.g., IR or NIR light sources such as an array or ring of LEDs
- the eye tracking cameras may be pointed towards the user’s eyes to receive reflected IR or NIR light from the light sources directly from the eyes, or alternatively may be pointed towards “hot” mirrors located between the user’s eyes and the display panels that reflect IR or NIR light from the eyes to the eye tracking cameras while allowing visible light to pass.
- the eye tracking device 130 optionally captures images of the user’s eyes (e.g., as a video stream captured at 60-120 frames per second (fps)), analyze the images to generate gaze tracking information, and communicate the gaze tracking information to the controller 110.
- images of the user’s eyes e.g., as a video stream captured at 60-120 frames per second (fps)
- fps frames per second
- two eyes of the user are separately tracked by respective eye tracking cameras and illumination sources.
- only one eye of the user is tracked by a respective eye tracking camera and illumination sources.
- the eye tracking device 130 is calibrated using a device-specific calibration process to determine parameters of the eye tracking device for the specific operating environment 100, for example the 3D geometric relationship and parameters of the LEDs, cameras, hot mirrors (if present), eye lenses, and display screen.
- the device-specific calibration process may be performed at the factory or another facility prior to delivery of the AR/VR equipment to the end user.
- the device-specific calibration process may be an automated calibration process or a manual calibration process.
- a user-specific calibration process may include an estimation of a specific user’s eye parameters, for example the pupil location, fovea location, optical axis, visual axis, eye spacing, etc.
- images captured by the eye tracking cameras can be processed using a glint-assisted method to determine the current visual axis and point of gaze of the user with respect to the display, in accordance with some embodiments.
- the eye tracking device 130 (e.g., 130A or 130B) includes eye lens(es) 520, and a gaze tracking system that includes at least one eye tracking camera 540 (e.g., infrared (IR) or near-IR (NIR) cameras) positioned on a side of the user’s face for which eye tracking is performed, and an illumination source 530 (e.g., IR or NIR light sources such as an array or ring of NIR light-emitting diodes (LEDs)) that emit light (e.g., IR or NIR light) towards the user’s eye(s) 592.
- IR infrared
- NIR near-IR
- an illumination source 530 e.g., IR or NIR light sources such as an array or ring of NIR light-emitting diodes (LEDs)
- the eye tracking cameras 540 may be pointed towards mirrors 550 located between the user’s eye(s) 592 and a display 510 (e.g., a left or right display panel of a head-mounted display, or a display of a handheld device, a projector, etc.) that reflect IR or NIR light from the eye(s) 592 while allowing visible light to pass (e.g., as shown in the top portion of Figure 5), or alternatively may be pointed towards the user’s eye(s) 592 to receive reflected IR or NIR light from the eye(s) 592 (e.g., as shown in the bottom portion of Figure 5).
- a display 510 e.g., a left or right display panel of a head-mounted display, or a display of a handheld device, a projector, etc.
- a display 510 e.g., a left or right display panel of a head-mounted display, or a display of a handheld device, a projector, etc.
- the controller 110 renders AR or VR frames 562 (e.g., left and right frames for left and right display panels) and provides the frames 562 to the display 510.
- the controller 110 uses gaze tracking input 542 from the eye tracking cameras 540 for various purposes, for example in processing the frames 562 for display.
- the controller 110 optionally estimates the user’s point of gaze on the display 510 based on the gaze tracking input 542 obtained from the eye tracking cameras 540 using the glint-assisted methods or other suitable methods.
- the point of gaze estimated from the gaze tracking input 542 is optionally used to determine the direction in which the user is currently looking.
- the controller 110 may render virtual content differently based on the determined direction of the user’s gaze. For example, the controller 110 may generate virtual content at a higher resolution in a foveal region determined from the user’s current gaze direction than in peripheral regions. As another example, the controller may position or move virtual content in the view based at least in part on the user’s current gaze direction. As another example, the controller may display particular virtual content in the view based at least in part on the user’s current gaze direction. As another example use case in AR applications, the controller 110 may direct external cameras for capturing the physical environments of the XR experience to focus in the determined direction.
- the autofocus mechanism of the external cameras may then focus on an object or surface in the environment that the user is currently looking at on the display 510.
- the eye lenses 520 may be focusable lenses, and the gaze tracking information is used by the controller to adjust the focus of the eye lenses 520 so that the virtual object that the user is currently looking at has the proper vergence to match the convergence of the user’s eyes 592.
- the controller 110 may leverage the gaze tracking information to direct the eye lenses 520 to Adjust focus so that close objects that the user is looking at appear at the right distance.
- the eye tracking device is part of a head-mounted device that includes a display (e.g., display 510), two eye lenses (e.g., eye lens(es) 520), eye tracking cameras (e.g., eye tracking camera(s) 540), and light sources (e.g., light sources 530 (e.g., IR or NIR LEDs), mounted in a wearable housing.
- the light sources emit light (e.g., IR or NIR light) towards the user’s eye(s) 592.
- the light sources may be arranged in rings or circles around each of the lenses as shown in Figure 5.
- eight light sources 530 e.g., LEDs
- the display 510 emits light in the visible light range and does not emit light in the IR or NIR range, and thus does not introduce noise in the gaze tracking system.
- the location and angle of eye tracking camera(s) 540 is given by way of example, and is not intended to be limiting.
- a single eye tracking camera 540 is located on each side of the user’s face.
- two or more NIR cameras 540 may be used on each side of the user’s face.
- a camera 540 with a wider field of view (FOV) and a camera 540 with a narrower FOV may be used on each side of the user’s face.
- a camera 540 that operates at one wavelength e.g., 850 nm
- a camera 540 that operates at a different wavelength e.g., 940 nm
- Embodiments of the gaze tracking system as illustrated in Figure 5 may, for example, be used in computer-generated reality, virtual reality, and/or mixed reality applications to provide computer-generated reality, virtual reality, augmented reality, and/or augmented virtuality experiences to the user.
- FIG. 6 illustrates a glint-assisted gaze tracking pipeline, in accordance with some embodiments.
- the gaze tracking pipeline is implemented by a glint-assisted gaze tracking system (e.g., eye tracking device 130 as illustrated in Figures 1A and 5).
- the glint-assisted gaze tracking system may maintain a tracking state. Initially, the tracking state is off or “NO”. When in the tracking state, the glint-assisted gaze tracking system uses prior information from the previous frame when analyzing the current frame to track the pupil contour and glints in the current frame. When not in the tracking state, the glint-assisted gaze tracking system attempts to detect the pupil and glints in the current frame and, if successful, initializes the tracking state to “YES” and continues with the next frame in the tracking state.
- the gaze tracking cameras may capture left and right images of the user’s left and right eyes.
- the captured images are then input to a gaze tracking pipeline for processing beginning at 610.
- the gaze tracking system may continue to capture images of the user’s eyes, for example at a rate of 60 to 120 frames per second.
- each set of captured images may be input to the pipeline for processing. However, in some embodiments or under some conditions, not all captured frames are processed by the pipeline.
- the method proceeds to element 640.
- the tracking state is NO, then as indicated at 620 the images are analyzed to detect the user’s pupils and glints in the images.
- the method proceeds to element 640. Otherwise, the method returns to element 610 to process next images of the user’s eyes.
- the current frames are analyzed to track the pupils and glints based in part on prior information from the previous frames.
- the tracking state is initialized based on the detected pupils and glints in the current frames.
- Results of processing at element 640 are checked to verify that the results of tracking or detection can be trusted. For example, results may be checked to determine if the pupil and a sufficient number of glints to perform gaze estimation are successfully tracked or detected in the current frames.
- the tracking state is set to NO at element 660, and the method returns to element 610 to process next images of the user’s eyes.
- the method proceeds to element 670.
- the tracking state is set to YES (if not already YES), and the pupil and glint information is passed to element 680 to estimate the user’s point of gaze.
- Figure 6 is intended to serve as one example of eye tracking technology that may be used in a particular implementation.
- eye tracking technologies that currently exist or are developed in the future may be used in place of or in combination with the glint-assisted eye tracking technology describe herein in the computer system 101 for providing XR experiences to users, in accordance with various embodiments.
- the captured portions of real-world environment 602 are used to provide a XR experience to the user, for example, a mixed reality environment in which one or more virtual objects are superimposed over representations of real -world environment 602.
- a three-dimensional environment optionally includes a representation of a table that exists in the physical environment, which is captured and displayed in the three-dimensional environment (e.g., actively via cameras and displays of a computer system, or passively via a transparent or translucent display of the computer system).
- the three-dimensional environment is optionally a mixed reality system in which the three-dimensional environment is based on the physical environment that is captured by one or more sensors of the computer system and displayed via a display generation component.
- the computer system is optionally able to selectively display portions and/or objects of the physical environment such that the respective portions and/or objects of the physical environment appear as if they exist in the three-dimensional environment displayed by the computer system.
- the computer system is optionally able to display virtual objects in the three-dimensional environment to appear as if the virtual objects exist in the real world (e.g., physical environment) by placing the virtual objects at respective locations in the three-dimensional environment that have corresponding locations in the real world.
- the computer system optionally displays a vase such that it appears as if a real vase is placed on top of a table in the physical environment.
- a respective location in the three- dimensional environment has a corresponding location in the physical environment.
- the computer system when the computer system is described as displaying a virtual object at a respective location with respect to a physical object (e.g., such as a location at or near the hand of the user, or at or near a physical table), the computer system displays the virtual object at a particular location in the three-dimensional environment such that it appears as if the virtual object is at or near the physical object in the physical world (e.g., the virtual object is displayed at a location in the three-dimensional environment that corresponds to a location in the physical environment at which the virtual object would be displayed if it were a real object at that particular location).
- a physical object e.g., such as a location at or near the hand of the user, or at or near a physical table
- the computer system displays the virtual object at a particular location in the three-dimensional environment such that it appears as if the virtual object is at or near the physical object in the physical world (e.g., the virtual object is displayed at a location in the three-dimensional environment that corresponds to a location in the
- real world objects that exist in the physical environment that are displayed in the three-dimensional environment can interact with virtual objects that exist only in the three-dimensional environment.
- a three-dimensional environment can include a table and a vase placed on top of the table, with the table being a view of (or a representation of) a physical table in the physical environment, and the vase being a virtual object.
- a three-dimensional environment e.g., a real environment, a virtual environment, or an environment that includes a mix of real and virtual objects
- objects are sometimes referred to as having a depth or simulated depth, or objects are referred to as being visible, displayed, or placed at different depths.
- depth refers to a dimension other than height or width.
- depth is defined relative to a fixed set of coordinates (e.g., where a room or an object has a height, depth, and width defined relative to the fixed set of coordinates).
- depth is defined relative to a location or viewpoint of a user, in which case, the depth dimension varies based on the location of the user and/or the location and angle of the viewpoint of the user.
- depth is defined relative to a location of a user that is positioned relative to a surface of an environment (e.g., a floor of an environment, or a surface of the ground)
- objects that are further away from the user along a line that extends parallel to the surface are considered to have a greater depth in the environment, and/or the depth of an object is measured along an axis that extends outward from a location of the user and is parallel to the surface of the environment (e.g., depth is defined in a cylindrical or substantially cylindrical coordinate system with the position of the user at the center of the cylinder that extends from a head of the user toward feet of the user).
- depth is defined relative to viewpoint of a user (e.g., a direction relative to a point in space that determines which portion of an environment that is visible via a head mounted device or other display)
- objects that are further away from the viewpoint of the user along a line that extends parallel to the direction of the viewpoint of the user are considered to have a greater depth in the environment, and/or the depth of an object is measured along an axis that extends outward from a line that extends from the viewpoint of the user and is parallel to the direction of the viewpoint of the user (e.g., depth is defined in a spherical or substantially spherical coordinate system with the origin of the viewpoint at the center of the sphere that extends outwardly from a head of the user).
- depth is defined relative to a user interface container (e.g., a window or application in which application and/or system content is displayed) where the user interface container has a height and/or width, and depth is a dimension that is orthogonal to the height and/or width of the user interface container.
- a user interface container e.g., a window or application in which application and/or system content is displayed
- depth is a dimension that is orthogonal to the height and/or width of the user interface container.
- the height and or width of the container are typically orthogonal or substantially orthogonal to a line that extends from a location based on the user (e.g., a viewpoint of the user or a location of the user) to the user interface container (e.g., the center of the user interface container, or another characteristic point of the user interface container) when the container is placed in the three- dimensional environment or is initially displayed (e.g., so that the depth dimension for the container extends outward away from the user or the viewpoint of the user).
- a location based on the user e.g., a viewpoint of the user or a location of the user
- the user interface container e.g., the center of the user interface container, or another characteristic point of the user interface container
- depth of an object relative to the user interface container refers to a position of the object along the depth dimension for the user interface container.
- multiple different containers can have different depth dimensions (e.g., different depth dimensions that extend away from the user or the viewpoint of the user in different directions and/or from different starting points).
- the direction of the depth dimension remains constant for the user interface container as the location of the user Interface container, the user and/or the viewpoint of the user changes (e.g., or when multiple different viewers are viewing the same container in the three-dimensional environment such as during an in-person collaboration session and/or when multiple participants are in a real-time communication session with shared virtual content including the container).
- the depth dimension optionally extends into a surface of the curved container.
- z-separation e.g., separation of two objects in a depth dimension
- z-height e.g., distance of one object from another in a depth dimension
- z-position e.g., position of one object in a depth dimension
- z-depth e.g., position of one object in a depth dimension
- simulated z dimension e.g., depth used as a dimension of an object, dimension of an environment, a direction in space, and/or a direction in simulated space
- a user is optionally able to interact with virtual objects in the three-dimensional environment using one or more hands as if the virtual objects were real objects in the physical environment.
- one or more sensors of the computer system optionally capture one or more of the hands of the user and display representations of the hands of the user in the three-dimensional environment (e.g., in a manner similar to displaying a real world object in three-dimensional environment described above), or in some embodiments, the hands of the user are visible via the display generation component via the ability to see the physical environment through the user interface due to the transparency/translucency of a portion of the display generation component that is displaying the user interface or due to projection of the user interface onto a transparent/translucent surface or projection of the user interface onto the user’s eye or into a field of view of the user’s eye.
- the hands of the user are displayed at a respective location in the three-dimensional environment and are treated as if they were objects in the three-dimensional environment that are able to interact with the virtual objects in the three-dimensional environment as if they were physical objects in the physical environment.
- the computer system is able to update display of the representations of the user’s hands in the three-dimensional environment in conjunction with the movement of the user’s hands in the physical environment.
- the computer system is optionally able to determine the “effective” distance between physical objects in the physical world and virtual objects in the three-dimensional environment, for example, for the purpose of determining whether a physical object is directly interacting with a virtual object (e.g., whether a hand is touching, grabbing, holding, etc. a virtual object or within a threshold distance of a virtual object).
- a hand directly interacting with a virtual object optionally includes one or more of a finger of a hand pressing a virtual button, a hand of a user grabbing a virtual vase, two fingers of a hand of the user coming together and pinching/holding a user interface of an application, and any of the other types of interactions described here.
- the computer system optionally determines the distance between the hands of the user and virtual objects when determining whether the user is interacting with virtual objects and/or how the user is interacting with virtual objects.
- the computer system determines the distance between the hands of the user and a virtual object by determining the distance between the location of the hands in the three- dimensional environment and the location of the virtual object of interest in the three- dimensional environment.
- the one or more hands of the user are located at a particular position in the physical world, which the computer system optionally captures and displays at a particular corresponding position in the three-dimensional environment (e.g., the position in the three-dimensional environment at which the hands would be displayed if the hands were virtual, rather than physical, hands).
- the position of the hands in the three- dimensional environment is optionally compared with the position of the virtual object of interest in the three-dimensional environment to determine the distance between the one or more hands of the user and the virtual object.
- the computer system optionally determines a distance between a physical object and a virtual object by comparing positions in the physical world (e.g., as opposed to comparing positions in the three- dimensional environment).
- the computer system when determining the distance between one or more hands of the user and a virtual object, the computer system optionally determines the corresponding location in the physical world of the virtual object (e.g., the position at which the virtual object would be located in the physical world if it were a physical object rather than a virtual object), and then determines the distance between the corresponding physical position and the one of more hands of the user. In some embodiments, the same techniques are optionally used to determine the distance between any physical object and any virtual object.
- the computer system when determining whether a physical object is in contact with a virtual object or whether a physical object is within a threshold distance of a virtual object, the computer system optionally performs any of the techniques described above to map the location of the physical object to the three-dimensional environment and/or map the location of the virtual object to the physical environment.
- the same or similar technique is used to determine where and what the gaze of the user is directed to and/or where and at what a physical stylus held by a user is pointed. For example, if the gaze of the user is directed to a particular position in the physical environment, the computer system optionally determines the corresponding position in the three-dimensional environment (e.g., the virtual position of the gaze), and if a virtual object is located at that corresponding virtual position, the computer system optionally determines that the gaze of the user is directed to that virtual object. Similarly, the computer system is optionally able to determine, based on the orientation of a physical stylus, to where in the physical environment the stylus is pointing.
- the computer system is optionally able to determine, based on the orientation of a physical stylus, to where in the physical environment the stylus is pointing.
- the computer system determines the corresponding virtual position in the three-dimensional environment that corresponds to the location in the physical environment to which the stylus is pointing, and optionally determines that the stylus is pointing at the corresponding virtual position in the three- dimensional environment.
- the embodiments described herein may refer to the location of the user (e.g., the user of the computer system) and/or the location of the computer system in the three-dimensional environment.
- the user of the computer system is holding, wearing, or otherwise located at or near the computer system.
- the location of the computer system is used as a proxy for the location of the user.
- the location of the computer system and/or user in the physical environment corresponds to a respective location in the three-dimensional environment.
- the location of the computer system would be the location in the physical environment (and its corresponding location in the three-dimensional environment) from which, if a user were to stand at that location facing a respective portion of the physical environment that is visible via the display generation component, the user would see the objects in the physical environment in the same positions, orientations, and/or sizes as they are displayed by or visible via the display generation component of the computer system in the three-dimensional environment (e.g., in absolute terms and/or relative to each other).
- the location of the computer system and/or user is the position from which the user would see the virtual objects in the physical environment in the same positions, orientations, and/or sizes as they are displayed by the display generation component of the computer system in the three- dimensional environment (e.g., in absolute terms and/or relative to each other and the real world objects).
- various input methods are described with respect to interactions with a computer system.
- each example may be compatible with and optionally utilizes the input device or input method described with respect to another example.
- various output methods are described with respect to interactions with a computer system.
- each example may be compatible with and optionally utilizes the output device or output method described with respect to another example.
- various methods are described with respect to interactions with a virtual environment or a mixed reality environment through a computer system.
- UI user interfaces
- a computer system such as a portable multifunction device or a head-mounted device, in communication with a display generation component, one or more input devices, and optionally one or cameras.
- Figures 7A-7CH and Figures 17A1-17P include illustrations of three- dimensional environments that are visible via a display generation component (e.g., a display generation component 7100, or a display generation component 120) of a computer system (e.g., computer system 101) and interactions that occur in the three-dimensional environments caused by user inputs directed to the three-dimensional environments and/or inputs received from other computer systems and/or sensors.
- a display generation component e.g., a display generation component 7100, or a display generation component 120
- a computer system e.g., computer system 101
- an input is directed to a virtual object within a three-dimensional environment by a user’s gaze detected in the region occupied by the virtual object, or by a hand gesture performed at a location in the physical environment that corresponds to the region of the virtual object.
- an input is directed to a virtual object within a three-dimensional environment by a hand gesture that is performed (e.g., optionally, at a location in the physical environment that is independent of the region of the virtual object in the three-dimensional environment) while the virtual object has input focus (e.g., while the virtual object has been selected by a concurrently and/or previously detected gaze input, selected by a concurrently or previously detected pointer input, and/or selected by a concurrently and/or previously detected gesture input).
- an input is directed to a virtual object within a three-dimensional environment by an input device that has positioned a focus selector object (e.g., a pointer object or selector object) at the position of the virtual object.
- an input is directed to a virtual object within a three-dimensional environment via other means (e.g., voice and/or control button).
- an input is directed to a representation of a physical object or a virtual object that corresponds to a physical object by the user’s hand movement (e.g., whole hand movement, whole hand movement in a respective posture, movement of one portion of the user’s hand relative to another portion of the hand, and/or relative movement between two hands) and/or manipulation with respect to the physical object (e.g., touching, swiping, tapping, opening, moving toward, and/or moving relative to).
- the computer system displays some changes in the three- dimensional environment (e.g., displaying additional virtual content, ceasing to display existing virtual content, and/or transitioning between different levels of immersion with which visual content is being displayed) in accordance with inputs from sensors (e.g., image sensors, temperature sensors, biometric sensors, motion sensors, and/or proximity sensors) and contextual conditions (e.g., location, time, and/or presence of others in the environment).
- sensors e.g., image sensors, temperature sensors, biometric sensors, motion sensors, and/or proximity sensors
- contextual conditions e.g., location, time, and/or presence of others in the environment.
- the computer system displays some changes in the three-dimensional environment (e.g., displaying additional virtual content, ceasing to display existing virtual content, and/or transitioning between different levels of immersion with which visual content is being displayed) in accordance with inputs from other computers used by other users that are sharing the computer-generated environment with the user of the computer system (e.g., in a shared computer-generated experience, in a shared virtual environment, and/or in a shared virtual or augmented reality environment of a communication session).
- some changes in the three-dimensional environment e.g., displaying additional virtual content, ceasing to display existing virtual content, and/or transitioning between different levels of immersion with which visual content is being displayed
- inputs from other computers used by other users that are sharing the computer-generated environment with the user of the computer system e.g., in a shared computer-generated experience, in a shared virtual environment, and/or in a shared virtual or augmented reality environment of a communication session.
- the computer system displays some changes in the three-dimensional environment (e.g., displaying movement, deformation, and/or changes in visual characteristics of a user interface, a virtual surface, a user interface object, and/or virtual scenery) in accordance with inputs from sensors that detect movement of other persons and objects and movement of the user that may not qualify as a recognized gesture input for triggering an associated operation of the computer system.
- some changes in the three-dimensional environment e.g., displaying movement, deformation, and/or changes in visual characteristics of a user interface, a virtual surface, a user interface object, and/or virtual scenery
- a three-dimensional environment that is visible via a display generation component described herein is a virtual three-dimensional environment that includes virtual objects and content at different virtual positions in the three-dimensional environment without a representation of the physical environment.
- the three-dimensional environment is a mixed reality environment that displays virtual objects at different virtual positions in the three-dimensional environment that are constrained by one or more physical aspects of the physical environment (e.g., positions and orientations of walls, floors, surfaces, direction of gravity, time of day, and/or spatial relationships between physical objects).
- the three-dimensional environment is an augmented reality environment that includes a representation of the physical environment.
- the representation of the physical environment includes respective representations of physical objects and surfaces at different positions in the three-dimensional environment, such that the spatial relationships between the different physical objects and surfaces in the physical environment are reflected by the spatial relationships between the representations of the physical objects and surfaces in the three-dimensional environment.
- virtual objects when virtual objects are placed relative to the positions of the representations of physical objects and surfaces in the three-dimensional environment, they appear to have corresponding spatial relationships with the physical objects and surfaces in the physical environment.
- the computer system transitions between displaying the different types of environments (e.g., transitions between presenting a computer-generated environment or experience with different levels of immersion, adjusting the relative prominence of audio/visual sensory inputs from the virtual content and from the representation of the physical environment) based on user inputs and/or contextual conditions.
- the different types of environments e.g., transitions between presenting a computer-generated environment or experience with different levels of immersion, adjusting the relative prominence of audio/visual sensory inputs from the virtual content and from the representation of the physical environment
- the display generation component includes a pass- through portion in which the representation of the physical environment is displayed or visible.
- the pass-through portion of the display generation component is a transparent or semi-transparent (e.g., see-through) portion of the display generation component revealing at least a portion of a physical environment surrounding and within the field of view of a user (sometimes called “optical passthrough”).
- the pass- through portion is a portion of a head-mounted display or heads-up display that is made semitransparent (e.g., less than 50%, 40%, 30%, 20%, 15%, 10%, or 5% of opacity) or transparent, such that the user can see through it to view the real world surrounding the user without removing the head-mounted display or moving away from the heads-up display.
- the pass-through portion gradually transitions from semi-transparent or transparent to fully opaque when displaying a virtual or mixed reality environment.
- the pass-through portion of the display generation component displays a live feed of images or video of at least a portion of physical environment captured by one or more cameras (e.g., rear facing camera(s) of a mobile device or associated with a head-mounted display, or other cameras that feed image data to the computer system) (sometimes called “digital passthrough”).
- the one or more cameras point at a portion of the physical environment that is directly in front of the user’s eyes (e.g., behind the display generation component relative to the user of the display generation component).
- the one or more cameras point at a portion of the physical environment that is not directly in front of the user’s eyes (e.g., in a different physical environment, or to the side of or behind the user).
- At least some of the virtual objects are displayed in place of (e.g., replacing display of) a portion of the live view (e.g., a portion of the physical environment captured in the live view) of the cameras.
- at least some of the virtual objects and content are projected onto physical surfaces or empty space in the physical environment and are visible through the pass-through portion of the display generation component (e.g., viewable as part of the camera view of the physical environment, or through the transparent or semi-transparent portion of the display generation component).
- at least some of the virtual objects and virtual content are displayed to overlay a portion of the display and block the view of at least a portion of the physical environment visible through the transparent or semi-transparent portion of the display generation component.
- the display generation component displays different views of the three-dimensional environment in accordance with user inputs or movements that change the virtual position of the viewpoint of the currently displayed view of the three- dimensional environment relative to the three-dimensional environment.
- the viewpoint moves in accordance with navigation or locomotion requests (e.g., in-air hand gestures, and/or gestures performed by movement of one portion of the hand relative to another portion of the hand) without requiring movement of the user’s head, torso, and/or the display generation component in the physical environment.
- movement of the user’s head and/or torso, and/or the movement of the display generation component or other location sensing elements of the computer system e.g., due to the user holding the display generation component or wearing the HMD
- movement of the viewpoint e.g., with corresponding movement direction, movement distance, movement speed, and/or change in orientation
- movement of the viewpoint e.g., with corresponding movement direction, movement distance, movement speed, and/or change in orientation
- a virtual object when a virtual object has a preset spatial relationship relative to the viewpoint (e.g., is anchored or fixed to the viewpoint), movement of the viewpoint relative to the three-dimensional environment would cause movement of the virtual object relative to the three-dimensional environment while the position of the virtual object in the field of view is maintained (e.g., the virtual object is said to be head locked).
- a virtual object is body- locked to the user, and moves relative to the three-dimensional environment when the user moves as a whole in the physical environment (e.g., carrying or wearing the display generation component and/or other location sensing component of the computer system), but will not move in the three-dimensional environment in response to the user’ s head movement alone (e.g., the display generation component and/or other location sensing component of the computer system rotating around a fixed location of the user in the physical environment).
- a virtual object is, optionally, locked to another portion of the user, such as a user’s hand or a user’s wrist, and moves in the three-dimensional environment in accordance with movement of the portion of the user in the physical environment, to maintain a preset spatial relationship between the position of the virtual object and the virtual position of the portion of the user in the three-dimensional environment.
- a virtual object is locked to a preset portion of a field of view provided by the display generation component, and moves in the three-dimensional environment in accordance with the movement of the field of view, irrespective of movement of the user that does not cause a change of the field of view.
- the views of a three-dimensional environment sometimes do not include representation(s) of a user’s hand(s), arm(s), and/or wrist(s).
- the representation(s) of a user’s hand(s), arm(s), and/or wrist(s) are included in the views of a three-dimensional environment.
- the representation(s) of a user’s hand(s), arm(s), and/or wrist(s) are included in the views of a three-dimensional environment as part of the representation of the physical environment provided via the display generation component.
- the representations are not part of the representation of the physical environment and are separately captured (e.g., by one or more cameras pointing toward the user’s hand(s), arm(s), and wrist(s)) and displayed in the three-dimensional environment independent of the currently displayed view of the three-dimensional environment.
- the representation(s) include camera images as captured by one or more cameras of the computer system(s), or stylized versions of the arm(s), wrist(s) and/or hand(s) based on information captured by various sensors).
- the representation(s) replace display of, are overlaid on, or block the view of, a portion of the representation of the physical environment.
- real-time visual representations e.g., stylized representations or segmented camera images
- the display generation component does not provide a view of a physical environment, and provides a completely virtual environment (e.g., no camera view and no transparent pass-through portion)
- real-time visual representations e.g., stylized representations or segmented camera images
- the position that corresponds to the user’s hand is optionally indicated in the three-dimensional environment, e.g., by the changing appearance of the virtual content (e.g., through a change in translucency and/or simulated reflective index) at positions in the three-dimensional environment that correspond to the location of the user’s hand in the physical environment.
- the representation of the user’s hand or wrist is outside of the currently displayed view of the three-dimensional environment while the virtual position in the three-dimensional environment that corresponds to the location of the user’s hand or wrist is outside of the current field of view provided via the display generation component; and the representation of the user’s hand or wrist is made visible in the view of the three-dimensional environment in response to the virtual position that corresponds to the location of the user’s hand or wrist being moved within the current field of view due to movement of the display generation component, the user’s hand or wrist, the user’s head, and/or the user as a whole.
- Figures 7A-7R illustrate examples of displaying window controls for a virtual object (e.g., an application window and/or three-dimensional object).
- Figure 8 is a flow diagram of an exemplary method 800 for conditionally displaying window controls.
- Figure 9 is a flow diagram of an exemplary method 900 for updating visual properties of window controls in response to user interactions.
- the user interfaces in Figures 7A-7R are used to illustrate the processes described below, including the processes in Figures 8 and 9.
- Figure 7A illustrates a view of a physical environment that includes a user 7002 interacting with a display generation component 7100.
- the user 7002 uses one or both of their two hands, a hand 7020 and a hand 7022 to provide inputs or instructions to a computer system.
- the computer system also uses the position or movement of an arm of the user, such as the user’s left arm 7028, which is connected to the user’s left hand 7020, as part of an input provided by the user to the computer system.
- a physical environment 7000 includes a physical object 7014, and physical walls 7004 and 7006.
- the physical environment 7000 further includes a physical floor 7008.
- display generation component 7100 of computer system 101 is a touchscreen held by user 7002.
- the display generation component of computer system 101 is a head mounted display (e.g., head mounted display 7100a, as shown in Figures 7F2-7F3, 7K1-7K2, 7T2-7T3, 7AD2-7AD3, 7AN2-7AN3, 7AU2-7AU3, 7BA2-7BA3, 7BD2-7BD3, 7BM2- 7BN2 and 17A1-17B1) worn on user 7002’s head (e.g., what is shown in Figures 7A-7CH and 17A2-17P as being visible via display generation component 7100 of computer system 101 corresponds to user 7002’ s field of view when wearing a head mounted display).
- head mounted display 7100a e.g., head mounted display 7100a, as shown in Figures 7F2-7F3, 7K1-7K2, 7T2-7T3, 7AD2-7AD3, 7AN2-7AN3, 7AU2-7AU3, 7BA2-7BA3, 7BD2-7BD3, 7BM2- 7
- the display generation component is a standalone display, a projector, or another type of display.
- the computer system is in communication with one or more input devices, including cameras or other sensors and input devices that detect movement of the user’s hand(s), movement of the user’s body as whole, and/or movement of the user’s head in the physical environment.
- the one or more input devices detect the movement and the current postures, orientations, and positions of the user’s hand(s), face, and/or body as a whole.
- a representation of the user’s hand 7020’ is displayed in the user interface displayed (e.g., as a passthrough representation and/or as a virtual representation of the user’s hand 7020) on the display of HMD 7100a.
- a representation of the user’s hand 7022’ is displayed in the user interface displayed (e.g., as a passthrough representation and/or as a virtual representation of the user’s hand 7022) on the display of HMD 7100a.
- the user’s hand 7020 and/or the user’s hand 7022 are used to perform one or more gestures (e.g., one or more air gestures), optionally in combination with a gaze input.
- the one or more gestures performed with the user’s hand(s) 7020 and/or 7022 include a direct air gesture input that is based on a position of the representation of the user’s hand(s) 7020’ and/or 7022’ displayed within the user interface on the display of HMD 7100a.
- a direct air gesture input is determined as being directed to a user interface object displayed at a position that intersects with the displayed position of the representation of the user’s hand(s) 7020’ and/or 7022’ in the user interface.
- the one or more gestures performed with the user’s hand(s) 7020 and/or 7022 include an indirect air gesture input that is based on a virtual object displayed at a position that corresponds a position at which the user’s attention is currently detected (e.g., and/or is optionally not based on a position of the representation of the user’s hand(s) 7020’ and/or 7022’ displayed within the user interface).
- an indirect air gesture is performed with respect to a user interface object while detecting the user’s attention (e.g., based on gaze or other indication of user attention) on the user interface object, such as a gaze and pinch (e.g., or other gesture performed with the user’s hand).
- user inputs are detected via a touch-sensitive surface or touchscreen.
- the one or more input devices include an eye tracking component that detects location and movement of the user’s gaze.
- the display generation component, and optionally, the one or more input devices and the computer system are parts of a head-mounted device that moves and rotates with the user’s head in the physical environment, and changes the viewpoint of the user in the three- dimensional environment provided via the display generation component.
- the display generation component is a heads-up display that does not move or rotate with the user’s head or the user’s body as a whole, but, optionally, changes the viewpoint of the user in the three-dimensional environment in accordance with the movement of the user’s head or body relative to the display generation component.
- the display generation component e.g., a touchscreen
- the display generation component is optionally moved and rotated by the user’s hand relative to the physical environment or relative to the user’s head, and changes the viewpoint of the user in the three-dimensional environment in accordance with the movement of the display generation component relative to the user’s head or face or relative to the physical environment.
- the display generation component 7100 comprises a head mounted display (HMD) 7100a.
- HMD head mounted display
- the head mounted display 7100a includes one or more displays that displays a representation of a portion of the three-dimensional environment 7000’ that corresponds to the perspective of the user
- an HMD typically includes multiple displays including a display for a right eye and a separate display for a left eye that display slightly different images to generate user interfaces with stereoscopic depth, in the figures a single image is shown that corresponds to the image for a single eye and depth information is indicated with other annotations or description of the figures.
- HMD 7100a includes one or more sensors (e.g., one or more interior- and/or exterior-facing image sensors 314), such as sensor 7101a, sensor 7101b and/or sensor 7101c for detecting a state of the user, including facial and/or eye tracking of the user (e.g., using one or more inward-facing sensors 7101a and/or 7101b) and/or tracking hand, torso, or other movements of the user (e.g., using one or more outward-facing sensors 7101c).
- sensors e.g., one or more interior- and/or exterior-facing image sensors 314
- sensor 7101a e.g., sensor 7101a, sensor 7101b and/or sensor 7101c for detecting a state of the user, including facial and/or eye tracking of the user (e.g., using one or more inward-facing sensors 7101a and/or 7101b) and/or tracking hand, torso, or other movements of the user (e.g., using one or more out
- HMD 7100a includes one or more input devices that are optionally located on a housing of HMD 7100a, such as one or more buttons, trackpads, touchscreens, scroll wheels, digital crowns that are rotatable and depressible or other input devices.
- input elements are mechanical input elements
- input elements are solid state input elements that respond to press inputs based on detected pressure or intensity.
- HMD 7100a includes one or more of button 701a, button 701b and digital crown 703 for providing inputs to HMD 7100a. It will be understood that additional and/or alternative input devices may be included in HMD 7100a.
- Figure 7F3 (e.g., and Figures 7K2, 7T3, 7AD3, 7AN3, 7AU3, 7BA3 and 7BD3) illustrates a top-down view of the user 7002 in the physical environment 7000.
- the user 7002 is wearing HMD 7100a, such that the user’s hand(s) 7020 and/or 7022 (e.g., that are optionally used to provide air gestures or other user inputs) are physically present within the physical environment 7000 behind the display of HMD 7100a.
- Figure 7F2 (e.g., and Figures 7K1, 7T2, 7AD2, 7AN2, 7AU2, 7BA2, 7BD2, and 7BM2-7BN2) illustrates an alternative display generation component of the computer system than the display illustrated in Figures 7A-7E, 7G-7J, 7L-7T1, 7U-7AD1, 7AE-7AN1, 7AO-7U1, 7V-7BA1, 7BC-7BD1, 7BL-7N1 and 7BO-7CH.
- the computer system e.g., display generation component 7100 displays a view of a three-dimensional environment (e.g., an environment 7000’, a virtual three-dimensional environment, an augmented reality environment, a pass- through view of a physical environment, or a camera view of a physical environment).
- a three-dimensional environment e.g., an environment 7000’, a virtual three-dimensional environment, an augmented reality environment, a pass- through view of a physical environment, or a camera view of a physical environment.
- the three-dimensional environment is a virtual three-dimensional environment without a representation of the physical environment 7000.
- the three-dimensional environment is a mixed reality environment that is a virtual environment that is augmented by sensor data corresponding to the physical environment.
- the three-dimensional environment is an augmented reality environment that includes one or more virtual objects (e.g., an application window 702 and/or a virtual object 7028) and a representation of at least a portion of a physical environment (e.g., representations 7004’, 7006’ of walls, a representation 7008’ of a floor, and/or a representation 7014’ of a physical object 7014 in the physical environment 7000) surrounding the display generation component 7100.
- the representation of the physical environment includes a camera view of the physical environment.
- the representation of the physical environment includes a view of the physical environment through a transparent or semitransparent portion of the first display generation component.
- the application window 702 is displayed in a first view of the three-dimensional environment 7000’ at a first position.
- the application window 702 is associated with a first application that is executing on the computer system.
- the application window 702 displays content for the first application.
- the application window 702 is displayed with a first horizontal position, a first vertical position, and a first depth, or a perceived distance from the user, (e.g., a position defined by an x-axis, a y-axis, and a z-axis) within the first view of the three-dimensional environment 7000’.
- the application window 702 is locked (also referred to herein as anchored) to the three-dimensional environment, such that as the field of view of the three-dimensional environment changes, the application window 702 is maintained at its position within the three-dimensional environment.
- the application window 702 would be displayed in a peripheral region of a field of view of the user’s eyes while looking at the three-dimensional environment via the display generation component.
- the user is enabled to move a position of the application window 702 to place it in a different position in the three-dimensional environment 7000’, such that the application window 702 becomes locked to the new position in the three-dimensional environment.
- the grabber 706-1 is a selectable user interface object for the application window 702 that, when selected by the user (e.g., using gaze and/or gestures, such as air gestures), enables the user to reposition the application window 702 within the three-dimensional environment 7000’.
- the grabber 706-1 is displayed along a bottom-center edge of the application window 702.
- the grabber 706-1 is displayed at a different position relative to application window 702.
- a shape and/or size of the grabber bar changes based on a size of the application window 702. For example, a size of grabber 706-1 increases and/or decreases as a size of the application window 702 increases and/or decreases.
- the application window 702 is a two-dimensional object (e.g., the application window 702 appears flat from the viewpoint of the user).
- the grabber 706-1 is automatically, and without user input, displayed with the application window 702 while application window 702 is displayed in the three-dimensional environment. In some embodiments, the grabber 706-1 is only displayed while the user’s attention is directed to the application window 702 and disappears in response to the user’s attention moving away from the application window 702. In some embodiments, the grabber 706-1 is displayed in response to detecting the user’s gaze at a bottom center portion, or other predefined portion, of application window 702.
- Figure 7B further illustrates a virtual object 7028 that optionally does not correspond to a physical object in the physical environment.
- the virtual object 7028 is a three-dimensional object, such as a ball.
- the virtual object 7028 is associated with a second application, distinct from the first application associated with the application window 702.
- the first application associated with the application window 702 and/or the second application associated with the virtual object 7028 is a system application (e.g., associated with an operating system) of the computer system or is an application associated with a third-party that is executed by the computer system.
- Figure 7C illustrates the computer system detecting a user’s attention 710-1 (e.g., a user’s gaze) directed to a top left comer, or an area surrounding the top left corner (e.g., within a predefined area that overlaps the top left comer), of the application window 702.
- the user’s attention 710-1 is detected as a gaze input.
- the display generation component 7100 optionally displays an indication (e.g., a cursor or other user interface object) that corresponds to the detected user’s gaze input, such that the indication moves within the display area of the display generation component 7100 in accordance with movement of the user’s gaze.
- the user’s attention 710-1 is detected as another type of input, such as an air gesture.
- the user’s attention 710-1 represents a position of a cursor (e.g., or other visual indicator) associated with an input device (e.g., controlled by the user’s hand instead of the user’s gaze).
- the application window 702 is displayed in front of (e.g., closer to the user than) the virtual object 7028.
- the computer system optionally visually deemphasizes the virtual object 7028 by dimming the virtual object 7028, pushing the virtual object 7028 back (e.g., further away from the viewpoint of the user 7002) within the three-dimensional environment 7000’, and/or decreasing a size of the virtual object 7028.
- the computer system automatically changes a perceived depth of respective objects in accordance with the detected user’s attention.
- the computer system in response to detecting the user’s attention 710-1 directed to the top left corner of the application window 702, the computer system displays a close affordance 7030 for closing the application window 702, as illustrated in Figures 7D (e.g., Figures 7D1-7D3).
- the computer system 101 displays the close affordance 7030 after detecting the user’s attention 710-1 has been maintained for a threshold amount of time, and/or has satisfied other attention criteria. For example, in response to detecting a gaze input directed to the top left corner that is not maintained for a threshold amount of time (e.g., the user gazes elsewhere in the three-dimensional environment before satisfying the threshold amount of time), the computer system does not display the close affordance 7030.
- the close affordance 7030 is displayed as a distinct user interface element from the application window 702.
- the close affordance 7030 is separated from the application window 702 by a non-zero distance such that the portion of the three-dimensional environment 7000’ between the application window 702 and the close affordance 7030 is visible.
- the close affordance 7030 is displayed above the application window 702, near the corner of the application window 702.
- the computer system optionally maintains display of the grabber 706-1.
- the computer system ceases display of the grabber 706-1.
- the title bar 716a is displayed below application window 702, optionally in response to detecting the user’s attention directed to a bottom portion of application window 702.
- a user interface object 705 is displayed concurrently with the grabber 706-1 and/or the title bar 716a (e.g., which has the same or analogous functionality as title bar 716, as described with reference to Figures 7S).
- the user interface object 705 is displayed proximate to (e.g., to the left and/or to the right of) the title bar 716a and/or grabber 706-1.
- the user interface object 705 corresponds to a minimized version of a close icon and/or a minimized version of one or more other controls (e.g., a control menu, or another type of control object).
- the user interface object 705 is displayed before detecting the user’s attention directed to the bottom portion of application window 702 (e.g., and, optionally, user interface object 705 continues to be displayed after the user’s attention is directed away from the bottom portion of application window 702).
- close affordance 7030-2 in response to detecting that the user’s attention 710- la is directed to the user interface object 705, the user interface object 705 is updated (e.g., from a minimized state, or a reduced state) to display close affordance 7030-2, as illustrated in Figure 7D2.
- close affordance 7030-2 has the same or analogous functionality as close affordance 7030, although close affordance 7030-2 is optionally displayed at a different position than close affordance 7030 relative to application window 702.
- the computer system while displaying the close affordance 7030 (or close affordance 7030-2), the computer system optionally detects a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a pinch input, a tap input, or another selection input) directed to the close affordance 7030 (or close affordance 7030-2, as illustrated in Figure 7D2).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a pinch input, a tap input, or another selection input
- the computer system in response to detecting the user input directed to the close affordance 7030 or 7030-2, the computer system ceases display of
- Figure 7D3 illustrates the computer system 101 detecting the user’s attention 710-2 directed to a bottom right corner of the application window 702.
- the computer system displays a resize affordance 708-1 (e.g., the resize affordance 708-1 shown in Figures 7F (e.g., Figures 7F1 and 7F2)) corresponding to the bottom right corner.
- the computer system before the resize affordance 708-1 is displayed, the computer system enables the user to access functionality and/or perform operations associated with the resize affordance 708-1. For example, without displaying resize affordance 708-1, the user is enabled to perform a gesture or other user input while directing the user’s attention to the bottom right corner of the application window 702 to resize the application window 702 (e.g., and in response to detecting the gesture or other user input, the application window 702 is resized in accordance with the gesture or other user input). In some embodiments, resize affordance 708-1 is displayed during and/or after the user performing the gesture.
- the resize affordance 708-1 is displayed in response to detecting the user’s attention 710-2 satisfies attention criteria. For example, the user maintains the user’s gaze at the bottom right comer for a threshold amount of time (e.g., 1 second, 2 seconds, 5 seconds, or another amount of time).
- the resize affordance 708-1 is displayed with a size and/or shape that is based on the application window 702. For example, in some embodiments, a size of resize affordance 708-1 is based on a size of the application window 702.
- the resize affordance 708-1 is displayed with an L-shape around the comer of application window 702 (e.g., to extend along a portion of the bottom edge and a portion of the right edge of application window 702), wherein a contour of the L-shape follows a contour of the comer of application window 702.
- the user’s attention 710-2 satisfies attention criteria that includes a criterion that is met when the user’s attention 710-2 is directed to a first area having a first size that corresponds to a respective portion of the application window 702.
- attention criteria that includes a criterion that is met when the user’s attention 710-2 is directed to a first area having a first size that corresponds to a respective portion of the application window 702.
- the user’s attention 710-2 is directed to an area having the first size that is centered about the bottom right corner of the application window 702.
- the computer system in response to detecting the user’s attention 710-2 directed to the bottom right corner of the application window 702, displays an animated transition from displaying the grabber 706-1 and the resize affordance 708-1.
- the animated transition includes displaying the grabber 706-1 gradually shifting to the right, as illustrated in Figure 7E, until it is replaced with the resize affordance 708-1 displayed at the bottom right comer of the application window 702.
- the grabber 706-1 is displayed as a single bar that morphs into an L-shape of the resize affordance 708-1.
- the animated transition includes reducing a size of the grabber 706-1 and/or fading the grabber 706-1, optionally without shifting a position of the grabber 706-1, until the grabber 706-1 is no longer displayed, and, optionally concurrently or after reducing the size of grabber 706-1, increasing a size of the resize affordance 708-1 and/or fading in the resize affordance 708-1.
- the animated transition removes the grabber 706-1 and initiates display of the resize affordance 708-1 at the corner in which the gaze of the user is detected.
- the grabber 706-1 is not displayed while the resize affordance 708-1 is displayed.
- the resize affordance 708-1 is displayed in accordance with a determination that the user’s attention 710-2 is directed to the bottom right comer of the application window for at least a threshold amount of time. For example, display of resize affordance 708-1 is delayed until the threshold amount of time has been met (e.g., and the user is optionally enabled to resize application window 702 before resize affordance 708-1 is displayed).
- the animated transition between displaying grabber 706-1 and resize affordance 708-1 is one example of an animation displayed for displaying object management controls, including the resize affordance, the grabber, the close affordance, the title bar, and/or other affordances that are dynamically displayed in response to detecting the user’s attention is directed to a respective portion of application window 702 (e.g., or other virtual object).
- the affordances described herein are responsive to detecting the user’s attention, such as the user’s gaze, such that the affordances are displayed in accordance with a determination that the user’s attention is directed to a portion of the displayed area that corresponds to the affordance (e.g., indicating that the user intends to interact with the affordance).
- the animation to display a respective affordance for the respective portion of application window 702 is initiated, and the user is enabled perform an operation associated with the affordance whether the animation is complete or incomplete.
- the user is enabled to select or otherwise perform a respective operation associated with the respective affordance (e.g., even before the respective affordance is displayed), so long as the user’s attention has dwelled at the respective position corresponding to the affordance for a threshold amount of time.
- the animation to display the respective affordance is initiated after the user’s attention has dwelled and the threshold amount of time has passed, but the user is enabled to interact with the affordance (e.g., by directing the user’s attention to a position that corresponds to the position of the affordance) before the affordance is displayed.
- application window 702 (e.g., or other virtual object, such as a three-dimensional virtual object) is divided into a plurality of regions, such as a left edge region, a left comer region, bottom region, a right corner region, and a right edge region.
- each of the plurality of regions optionally includes an area outside of and/or proximate to application window 702.
- the left comer region includes an area that extends beyond the outside of the left corner of application window 702.
- a respective affordance is enabled to appear at any one of the regions (e.g., the same affordance appears at any one of the regions, or a different affordance appears at different regions based on the region).
- resize affordance 708-1 appears at a bottom left comer region of application window 702, at a bottom right corner region of application window 702, at a top left corner region of application window 702 and/or at a top right corner region of application window 702 based on which of the corner regions at which the user’s attention is currently directed.
- the system determines a current state of each of the plurality of regions, and optionally performs an operation (e.g., and/or enables an operation to be performed in response to a user input) in accordance with the current state (e.g., and/or a change to the current state of a respective region).
- an operation e.g., and/or enables an operation to be performed in response to a user input
- the current state e.g., and/or a change to the current state of a respective region.
- the possible states include: the affordance is not visible and does not allow for user interaction, the affordance is not visible but does allow for interaction to perform an operation associated with the affordance, the affordance is displayed but the user’s attention is not dwelled in the region (e.g., the user’s attention is detected as being directed to the region for less than a threshold amount of time), the affordance is displayed and the user’s attention is dwelled in the region (e.g., the user’s attention is detected as directed to the region for at least the threshold amount of time), and the affordance is displayed and selected (e.g., pressed or otherwise interacted with).
- the title bar and/or other affordances are visible in only one of the plurality of regions (e.g., without being visible or available in the other regions of the plurality of regions).
- a user in a first region of the plurality of regions, a user is enabled to perform an operation associated with a first respective affordance (e.g., even if the first respective affordance is inactive or not displayed), and/or the user is enabled to direct the user’s attention to and/or select the first respective affordance within the first region, while in a second region of the plurality of regions, the first respective affordance is hidden and/or disabled (e.g., such that the user is not enabled to select or interact with the first respective affordance).
- the computer system in response to detecting a user input directed to (e.g., the user’s attention directed to) the first region that includes the first respective affordance (e.g., the user’s gaze is directed to a respective region for at least a threshold amount of time), the computer system provides visual feedback (e.g., a change in opacity, blur, and/or other visual feedback) in the region where the affordance is currently displayed.
- visual feedback e.g., a change in opacity, blur, and/or other visual feedback
- the computer system in response to detecting that the user’s attention is directed to a third region of the plurality of regions (e.g., a region other than the first region), where the first respective affordance is not displayed in the third region, the computer system displays an animation to display a second respective affordance (e.g., the same or a different affordance than the first respective affordance) for the third region to which the user is currently directing the user’s attention.
- a second respective affordance e.g., the same or a different affordance than the first respective affordance
- the grabber bar 706-1 is animated into a resize affordance 708-1 in response to detecting that the user’s attention is directed to the bottom right corner region.
- the resize affordance 708-1 continues to be displayed while detecting the user’s attention directed to the bottom right corner of application window 702. In some embodiments, in response to detecting that the user’s attention is no longer directed to the bottom right corner of the application window 702, the resize affordance 708- 1 is no longer displayed, and the grabber 706-1 is optionally redisplayed.
- detecting that the user’s attention is no longer directed to the bottom right corner of the application window 702 includes determining that the user’s attention is directed outside of a second area having a second size (e.g., distinct from the first area having the first size that corresponds to a respective portion of the application window 702 used to determine that the user’s attention satisfies attention criteria).
- the second area having the second size is a larger region than the first area having the first size.
- the second area completely encompasses the first area.
- detecting that the user’s attention is no longer directed to the bottom right comer is based on whether the user’s attention has moved outside of a larger sized region than the first area in which the user’s attention is directed to determine that the user’s attention satisfies attention criteria (e.g., and displaying the resize affordance in accordance with the determination that the user’s attention satisfies the attention criteria).
- a resize affordance (e.g., analogous to the resize affordance 708-1) is displayed proximate to the bottom left corner in response to detecting the user’s attention is directed to the bottom left corner.
- a mirror image of the resize affordance 708-1 is displayed at the bottom left corner. Accordingly, based on detecting which of the bottom comers the user’s attention is directed to, the computer system displays a corresponding resize affordance at the respective corner (e.g., the bottom right and/or bottom left comer).
- the computer system in response to detecting the user’s attention being directed to another area of the application window 702, other than the bottom right and/or bottom left comers, displays a respective affordance and/or forgoes displaying respective affordances that do not correspond to the current location to which the user’s attention is directed. For example, detecting the user’s attention directed to the top left corner of the application window 702 causes the computer system to cease display of resize affordance 708-1 and display the close affordance 7030, and, optionally, display (e.g., or redisplays) the grabber 706-1.
- different affordances are associated with respective portions of the application window 702 (e.g., and/or the virtual object 7028), such that the user invokes display of a respective affordance by directing the user’s attention to the respective portion of the application window 702 associated with the respective affordance.
- the examples described herein associate the bottom comers of the application window 702 with resize affordances, and a top left comer of application window 702 with a close affordance, it will be understood that the corners may be assigned to a different type of affordance based on the application window (e.g., different applications may associate different affordances with the corners).
- some application windows and/or virtual objects cannot be resized, and a resize affordance is not displayed in response to the user gazing at a corner of the application window and/or virtual object.
- one or more application windows and/or virtual objects cannot be repositioned within the three-dimensional object, and a grabber is not displayed for the one or more application windows and/or virtual objects.
- Figures 7F illustrates the computer system 101 detecting the user’s attention 710-4 directed to the resize affordance 708-1.
- the computer system in response to detecting that the user’s attention is directed to the resize affordance 708-1 (e.g., detecting that the user is gazing at resize affordance 708-1), the computer system updates one or more visual properties of the resize affordance 708-1 to indicate that the computer system detects the user’s attention directed to the resize affordance 708-1.
- Figure 7G illustrates a resize affordance 708-2 (e.g., an updated version of the resize affordance 708-1) displayed with a color distinct from the color of the resize affordance 708-1 in Figures 7F.
- updating the one or more visual properties of the resize affordance 708-1 includes displaying the (updated) resize affordance 708-2 with a different size, color, and/or transparency while the user’s attention 710-4 is detected as being directed to the resize affordance 708-2.
- resize affordance 708-1 after updating the one or more visual properties of resize affordance 708-1, and optionally before detecting additional user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of selection input) selecting (updated) resize affordance 708- 2, the computer system detects that the user’s attention is directed to another portion of the three-dimensional environment that does not correspond to the resize affordance 708-2.
- additional user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input
- the user’s attention 710-5 shifts to the left of resize affordance 708-3 (e.g., which is analogous to the resize affordance 708-2, but with different properties as described below).
- the computer system in response to detecting that the user’s attention is no longer directed to the resize affordance 708-2 (e.g., and/or detecting that the user’s attention is directed to another portion of the three-dimensional environment), the computer system reverses the updates to the one or more visual properties of the resize affordance, as represented by the resize affordance 708-3 in Figure 7H.
- the updated visual properties of resize affordance 708-2 are not maintained, and instead the resize affordance 708-3 is displayed with the same visual properties as the resize affordance 708-1 (e.g., is displayed with the same appearance as the resize affordance 708-1 in Figures 7F).
- the computer system in response to detecting that the user’s attention is not directed to the resize affordance 708-3, the computer system optionally maintains the update to the one or more visual properties of the resize affordance 708-2 (e.g., the resize affordance 708-3 has the same appearance as the resize affordance 708-2), depending on where the user’s attention is directed (e.g., whether the user’s attention is directed outside of a proximity of the resize affordance 708-3).
- the resize affordance 708-3 ceases to be displayed, and if the user looks at another portion of the application window 702 (e.g., but not at the resize affordance 708-3), the appearance of the resize affordance 708-3 is optionally maintained (e.g., with the different size and/or color of the resized affordance 708-2).
- the resize affordance 708-3 continues to be displayed, but with a different visual appearance (e.g., the color and/or size is changed, and is different from the color and/or size of the resized affordance 708-2 in Figure 7G). As such, depending on where the user’s gaze is detected, resize affordance 708-3 ceases to be displayed or is maintained, optionally with different visual properties.
- the computer system 101 detects that the user’s attention is redirected to (e.g., after being directed away from) the resize affordance 708-3 before a threshold amount of time has passed. For example, the user had quickly looked away from resize affordance 708-3 before looking back at resize affordance 708-3, and the computer system maintains display of the resize affordance 708-3 (e.g., for up to the threshold amount of time).
- the computer system 101 detects that the user’s attention is directed to another portion of the three-dimensional environment that does not correspond to the resize affordance 708-3 for the threshold amount of time, and after the threshold amount of time has passed, the computer system ceases display of the resize affordance 708-3, and optionally redisplays the grabber 706-1, as illustrated in Figure 7J.
- the computer system ceases display of resize affordance 708-3 by displaying an animated transition, such as decreasing a size of the resize affordance 708-3 and/or fading the resize affordance 708-3 until it is no longer displayed.
- the animated transition includes shifting the resize affordance down and/or around the respective corner of application window 702, optionally in a direction towards the grabber 706-1.
- the resize affordance 708-4 shifts down and to the left as if traveling and/or morphing into the grabber 706-1 in Figure 7J.
- the grabber 706-1 is optionally automatically redisplayed, without additional user input (e.g., without the user gazing at a bottom center edge of application window 702), in response to ceasing display of resize affordance 708-3.
- the grabber 706-1 is redisplayed in response to detecting the user’s attention 710-6 is directed to a bottom center edge of application window 702 (e.g., including an area outside of the application window 702 that is along the bottom edge of the application window 702, as illustrated in Figure 71).
- the computer system 101 displays an animated transition of moving resize affordance 708-4 until it is displayed as grabber 706-1.
- Figure 7J illustrates detecting the user’s attention 710-7 directed to the bottom right corner of the application window 702.
- the computer system 101 displays (e.g., or redisplays) the resize affordance 708-1, and in response to detecting that the user’s attention is directed to the resize affordance 708-1, updates the resize affordance (e.g., to the resize affordance 708-3) in a first manner, such as by changing the color of the resize affordance (e.g., as described above with reference to Figures 7F-7G).
- the computer system 101 detects a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to the resize affordance 708-3 (e.g., a user input performed by the user’s hand 7020 in Figures 7K (e.g., Figures 7K1, 7K2 and 7K3)).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- the user input directed to the resize affordance 708-3 is an air gesture, such as a pinch gesture or a tap input, that is detected while the user’s gaze is detected as being directed to the resize affordance 708-3.
- the computer system in response to the user input directed to the resize affordance 708-3, the computer system updates display of the resize affordance 708-3 in a second manner (e.g., that is different from the first manner).
- the computer system changes a size of the resize affordance 708-3 to a (e.g., smaller) size of a resize affordance 708-5, optionally while maintaining the updated color of the resize affordance, as illustrated in Figures 7K (e.g., Figures 7K1, 7K2 and 7K3).
- the computer system provides two levels of visual feedback to the user as the user interacts with the resize affordance.
- the computer system first changes a color of the resize affordance to indicate the computer system detects the user’s attention is on the resize affordance, and upon detecting a further user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) to interact with the resize affordance, the computer system changes a size of the resize affordance to indicate that the resize affordance has been selected by the user input.
- a further user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a
- changing a size of the resize affordance 708-3 to the size of resize affordance 708-5 includes changing a width, or thickness of the resize affordance without changing a length.
- the resize affordance 708-3 and the resize affordance 708-5 differ with respect to a first dimension (e.g., width), but are the same with respect to other dimensions (e.g., thickness).
- the computer system detects that the user’s attention is directed to another portion of the three-dimensional environment that does not correspond to the resize affordance 708-5 after detecting the user input (e.g., via the user’s hand 7020, such as an air gesture) in Figures 7K (e.g., Figures 7K1, 7K2 and 7K3).
- Figures 7K e.g., Figures 7K1, 7K2 and 7K3
- the updates to the visual properties are optionally maintained (e.g., resize affordance 708-5 continues to be displayed with the updated size and the updated color) even while the user’s attention is directed to a portion of the three-dimensional environment that does not correspond to resize affordance 708-5.
- the computer system continues the user’s interaction with resize affordance 708-5 (e.g., even while the user’s attention 710-5 is not directed to resize affordance 708-5).
- the user input detected via the user’s hand 7020 is enabled to continue to interact with the resize affordance 708-5 (e.g., the user is enabled to continue resizing the window by dragging the resize affordance 708-5 in one or more directions).
- the user input detected via the user’s hand 7020 indicates movement of the user’s hand 7020 that causes movement of the resize affordance 708-5.
- the user input continues by the user moving the user’s hand 7020 in a respective direction and/or by a respective amount (e.g., the user performs a drag gesture or an air drag gesture), optionally while maintaining the pinch gesture.
- the user performs a pinch and drag gesture while gazing at the resize affordance 708-5.
- the computer system 101 moves the resize affordance by an amount and/or with a direction of movement corresponding to the user’s drag gesture, and resizes the application window 702, as illustrated in Figure 7L.
- the application window 702 is resized in accordance with a direction and/or amount of movement of the user input detected via the user’s hand 7020.
- the user’s hand 7020 moves up and to the left, and in response, the corner of the application window 702 moves up and to the left, thereby decreasing a size of the application window 702 by a corresponding amount.
- movement of the user’s hand 7020 in the opposite direction causes the application window 702 to increase in size by an amount corresponding to the amount of movement of the user’s hand 7020.
- resizing the application window 702 includes maintaining a position of one or more edges of the application window 702 within the three- dimensional environment. For example, in Figure 7L, the top edge and the left edge of the application window 702 remain in a same position within the three-dimensional environment before and after resizing the application window 702.
- the edge(s) that maintain their respective positions are the edge(s) opposite (e.g., that are not a part of) the corner of the resize affordance.
- Figures 7K e.g., Figures 7K1 and 7K3
- the resize affordance 708-5 is displayed in a bottom right comer, such that the top edge and left edge of the application window 702 are maintained while resizing the application window 702, while the bottom right comer moves inward toward the top edge and left edge of the application window 702 to decrease the size of the application window 702.
- resizing the application window 702 includes maintaining a center of the application window 702 at a same position before, after and/or during the resizing of the application window 702. For example, as the application window 702 decreases in size, a plurality of (or, optionally, all of) the edges of application window 702 move inward (e.g., uniformly, and by the same distance), toward the center of application window 702 to decrease the size of the application window 702. Similarly, as the application window 702 increases in size, a plurality of (or, optionally, all of) the edges of application window 702 move outward, away from the center of the application window 702 while maintaining the center of the application window 702 at a same position.
- the computer system increases or decreases a size of the resize affordance 708-5 to indicate that the resize affordance 708-5 is currently selected by the user.
- the resize affordance 708-5 has a first size when currently selected by the user input, and the resize affordance 708-5 has a different size (e.g., a second size different from the first size) while the application window 702 is being resized.
- the resize affordance 708-6 is displayed with a size that is based on (e.g., proportional to) a size of application window 702.
- the application window 702 in Figure 7L is decreased in size
- the resize affordance 708-3 in Figure 7H the resize affordance 708-6 is decreased in size (optionally by an amount proportional to an amount of the decrease in size of application window 702).
- the resize affordance 708-5 changes in size by an amount based on (i) being currently selected by a user input and/or (ii) a current size of application window 702.
- the resize affordance 708-5 changes in size as the user resizes application window 702.
- the resize affordance 708-6 is displayed in a first updated state (e.g., with a second color), but without a user input currently selecting resize affordance 708-6, the resize affordance 708-6 is not displayed in its second updated state (e.g., with a different size).
- the computer system displays resize affordance 708-6 with a grey color to provide visual feedback that the computer system detects the user’s gaze is directed to the resize affordance 708-6 (e.g., whereas the resize affordance 708-6 is displayed with a white color if the user’s gaze is not directed to the resize affordance, as described above).
- Figure 7M illustrates detecting the user’s attention 710-10 directed to the virtual object 7028.
- the application window 702 continues to be displayed in the three-dimensional environment, even while the user’s attention is not directed to the application window 702.
- the grabber 706-2 is optionally displayed for the application window 702 even while the user’s attention is not directed to application window 702.
- the grabber 706-2 optionally ceases to be displayed while the user’s attention is not directed to the application window 702, and in response to detecting the user’s attention is directed to the application window 702 (e.g., or directed to a respective portion, such as a bottom center portion, of the application window 702), the grabber 706-2 is displayed.
- a size of the grabber 706-2 is decreased relative to a size of the grabber 706-1 after the application window 702 decreases in size.
- the grabber 706-1 is displayed with a size proportional to the application window 702 and updates in size as a size of the application window 702 changes.
- a size of application window 702 is dependent on a perceived distance away from the user (e.g., and/or the viewpoint of the user).
- the application window 702 decreases in size, and if the application window 702 (e.g., and its associated controls, such as the grabber 706-1) is moved in position towards the user, the application window 702 (e.g., and its associated controls, such as grabber 706-1) increases in size in accordance with the closer position to the user.
- the application window 702 e.g., and its associated controls, such as grabber 706-1
- a platter 7029 in response to detecting that the user’s attention 710-10 is directed to the virtual object 7028, a platter 7029 is displayed below the virtual object 7028.
- the platter 7029 comprises a flat surface that appears substantially parallel to the floor 7008’.
- the platter 7029 is displayed as a surface, optionally a floating surface, on which the virtual object 7028 is situated in the three- dimensional environment.
- the platter 7029 is displayed for three- dimensional virtual objects, such as the virtual object 7028, while two-dimensional objects, such as the application window 702, are displayed without a platter.
- a size of the platter 7029 is based on a size of the virtual object 7028.
- the platter 7029 is displayed while the user’s attention 710-10 continues to be directed to the virtual object 7028 and/or one or more controls (e.g., the grabber 712-1, the resize affordance 714-1, and/or the close affordance 717) for the virtual object 7028, and optionally is no longer displayed in response to detecting the user’s attention has moved away from the virtual object 7028 and/or the one or more controls for the virtual object 7028.
- one or more controls e.g., the grabber 712-1, the resize affordance 714-1, and/or the close affordance 71
- Figure 7N illustrates a grabber 712-1, for moving a position of the virtual object 7028.
- the grabber 712-1 is displayed as a distinct user interface object from the platter 7029, and is displayed along an edge of the platter 7029 (e.g., the closest edge of the platter 7029).
- the grabber 712-1 is automatically displayed, without additional user input, concurrently with display of the platter 7029 (e.g., in response to detecting the user’s attention is directed to the virtual object 7028).
- the grabber 712-1 is displayed in response to detecting that the user’s attention is directed to a bottom portion of the virtual object 7028, such as to the edge of the platter 7029 in an area proximate to where the grabber 712-1 is displayed in Figure 7N. For example, the user gazes at the closest edge of the platter 7029, and the grabber 712-1 is displayed.
- the user selects the grabber 712-1 using a selection input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a pinch input, or other selection input) and moves the virtual object 7028 within the three- dimensional environment (e.g., via a drag gesture, an air drag gesture, or other movement of the user input).
- a selection input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a pinch input or other selection input
- Figure 7N further illustrates the computer system 101 detecting the user’s attention 710-11 directed to a comer of the platter 7029.
- the computer system 101 in response to detecting that the user’s attention 710-11 is directed to the bottom left comer of the platter 7029, the computer system 101 displays a resize affordance 714-1 at the bottom left corner, as illustrated in Figure 70.
- the resize affordance 714-1 is displayed as an L-shape that extends along a portion of the edges of the platter 7029 with a contour that matches a contour of the corner of platter 7029.
- the user gazes at the bottom left comer or the bottom right corner of platter 7029, and in response, a respective resize affordance is displayed at the respective corner of platter 7029.
- an animated transition is displayed along the edge of platter 7029, and the animated transition includes ceasing display of grabber 712-1 and initiating display of resize affordance 714-1 (e.g., including any of the animated transitions between the grabber 706-1 and the resize affordance 708-1 described above with reference to Figures 7E-7F).
- Figure 70 illustrates the computer system 101 detecting the user’s attention 710-12 directed to the resize affordance 714-1, and in response to detecting the user’s attention 710-12, the computer system 101 changes a color of resize affordance 708-1 and/or updates one or more other visual properties of the resize affordance 708-1.
- the computer system in response to detecting a user input, such as an air pinch gesture, selecting the resize affordance 708-1, the computer system changes a size of the resize affordance 708-1 (e.g., to indicate it has been selected by the user) and/or updates one or more other visual properties of the resize affordance 708-1.
- Figure 7P illustrates the computer system 101 detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) that indicates a direction and an amount of movement based on movement of the user’s hand 7020, and changing a size of the virtual object 7028 in accordance with the user input (e.g., in accordance with the movement and/or direction of movement of the user’s hand 7020).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- the resize affordance 714-2 changes in size to indicate that the computer system detects the user interaction (e.g., gaze inputs and/or air gestures).
- the resize affordance 714-2 optionally increases in size while the user is interacting with the resize affordance 714-2 (e.g., by moving the user’s hand 7020 while the resize affordance 714-2 is selected), such as performing an air drag gesture on the resize affordance while the user’s attention 710-13 is directed to the resize affordance 714-2.
- the resize affordance 714-2 increases in size while the user is interacting with the resize affordance 714-2, because the user is decreasing a size of the virtual object 7028, the overall size of the resize affordance 714-2 appears to decrease (e.g., in accordance with the decrease in size of virtual object 7028).
- resize affordance 714-2 is displayed with a size proportional to the virtual object 7028 such that, as the virtual object 7028 decreases in size, the resize affordance 714-2 also decreases in size.
- the resize affordance 714-2 decreases in size by an amount that is less than the amount that the virtual object decreases in size, because the resize affordance 714-2 increases in size while the user is interacting with the resize affordance 714-2.
- Figure 7Q illustrates that, after the user input directed to the resize affordance 714-2 is no longer detected, the resize affordance 714-2 ceases to be displayed, and optionally, the grabber 712-2 is redisplayed under the platter 7029 of the virtual object 7028.
- the grabber 712-2 is displayed with a size based on the resized virtual object 7028 (e.g., the grabber 712-2 decreases in size in accordance with the decrease in size of virtual object 7028).
- Figure 7Q further illustrates the computer system 101 detecting that the user’s attention 710-14 is directed to a top left corner of the virtual object 7028.
- Figure 7R illustrates that, in response to detecting the user’s attention 710-14 directed to a top left corner of virtual object 7028, the computer system 101 displays the close affordance 717 for the virtual object 7028.
- the close icon 717 is displayed at a different portion of the virtual object 7028 (e.g., in accordance with the virtual object 7028 being a three-dimensional object).
- the close affordance 717 is displayed below the platter 7029 of virtual object 7028.
- the computer system detects a user input (e.g., via the user’s hand 7020), such as a tap gesture (e.g., a touch gesture, air gesture, or other selection input) and/or a pinch gesture, such as an air pinch gesture, directed to the close affordance 717 while the user is gazing at the close affordance 717, indicating that the user’s attention 710-15 is directed to the close affordance 717 (e.g., while the close affordance 717 is in a ready state, as described above).
- a user input e.g., via the user’s hand 7020
- a tap gesture e.g., a touch gesture, air gesture, or other selection input
- a pinch gesture such as an air pinch gesture
- the computer system In response to the user input selecting the close affordance 717, the computer system ceases display of the virtual object 7028 in the three-dimensional environment, as illustrated in Figure 7S. For example, the computer system closes the virtual object 7028, including optionally closing an application associated with the virtual object 7028.
- Figures 7S-7AD (e.g., Figures 7AD1, 7AD2 and 7AD3) illustrate examples of displaying a title bar that expands in response to detecting the user’s attention directed to the title bar.
- Figure 10 is a flow diagram of an exemplary method 1000 for displaying a title bar near an application window that provides additional control options for a user.
- the user interfaces in Figures 7S-7AD are used to illustrate the processes described below, including the processes in Figure 10.
- Figure 7S illustrates detecting the user’s attention 710-16, such as a gaze input, directed to the application window 702.
- the application window 702 is displayed concurrently with the title bar 716.
- the title bar 716 displays a name or other indication, such as an application icon, of an application associated with the application window 702 (e.g., “App 1”).
- the title bar 716 indicates respective content that is currently displayed in the application window 702.
- the title bar 716 includes a name of a document displayed in the application window 702 and/or a name of a website displayed in the application window 702.
- multiple tabs are displayed in a tab bar (optionally that includes the title bar 716), wherein the content associated with each tab is available for display in the application window 702.
- a user is enabled to switch, or otherwise navigate, between tabs to display different content (e.g., different documents, different webpages, or other content) for a same application and/or other applications within the application window 702.
- indications for other available tabs are displayed in a tab bar next to title bar 716, such that selecting another tab in the tab bar switches the content displayed in application window 702.
- the currently selected tab is displayed as the current title bar for the content currently displayed in the application window 702.
- the currently selected tab is displayed with a visual appearance that is different than the other tabs in the tab bar.
- the currently selected tab bar is displayed with an application icon, a different level of translucency, a different size and/or a different color to indicate it is the currently active tab.
- the title bar 716 is displayed as a distinct user interface object with a nonzero distance between the application window 702 and the title bar 716.
- the title bar 716 is displayed while the computer system detects that the user’s attention 710-16 is directed to the application window 702.
- the title bar 716 is displayed even if the user’s attention is not detected as being directed to the application window 702 (e.g., the title bar 716 is optionally displayed in Figures 7M-7R while the user is gazing at the virtual object 7028).
- the title bar 716 (and, optionally, the application window 702 and/or other controls of the application window 702) are visually deemphasized (e.g., dimmed, decreased in size, or otherwise deemphasized) if (e.g., and/or while) the user’s attention is not detected as being directed to the application window 702.
- Figures 7T-7AH illustrate examples of conditionally displaying privacy indicators.
- Figure 11 is a flow diagram of an exemplary method 1100 for maintaining a privacy indicator with an application window as the application window moves in the display area.
- the user interfaces in Figures 7T-7AH are used to illustrate the processes described below, including the processes in Figure 11.
- Figures 7T illustrates detecting that the user’s attention 710-17 continues to be directed to the application window 702.
- the application associated with the application window 702 begins accessing, using, and/or collecting sensor data from one or more sensors of the computer system.
- the application associated with the application window 702 accesses one or more of a microphone, a camera, a location, or other sensors of the computer system, that provides the application with access to sensitive, or private, data of the user.
- the computer system in response to detecting that the application is accessing one or more of the sensors that provides access to sensitive data of the user, the computer system displays a privacy indicator 718-1 above the application window 702, while optionally also maintaining display of the title bar 716.
- the privacy indicator 718-1 is displayed with a first set of properties that indicates which of the one or more sensors are being accessed by the application associated with application window 702.
- the privacy indicator 718- 1 is displayed with a respective color that corresponds to the type of sensor (e.g., a red indicator represents a camera is being accessed, an orange indicator represents a microphone is being accessed, and/or a blue indicator represents location data is being accessed). It will be understood that different visual properties and/or colors may be assigned to particular sensors to indicate which of the sensors are currently being accessed by the application associated with the application window 702.
- the privacy indicator 718-1 is displayed even if the user is not currently directing the user’s attention to the application window 702.
- the computer system indicates to the user when an application is accessing sensitive data using one or more sensors of the computer system whether or not the user is currently interacting with, or paying attention to, the application window 702.
- Figure 7U illustrates the computer system 101 detecting that the user’s attention 710-18 is directed to the privacy indicator 718-1.
- the computer system in response to detecting that the user’s attention 710-18 is directed to the privacy indicator 718-1, displays additional information about the sensor(s) being accessed by the application associated with the application window 702, as illustrated in Figure 7V.
- the privacy indicator 718-1 is expanded into an expanded privacy indicator 718-2 which includes a textual indication (e.g., and/or an icon representing the sensor) that the microphone is being accessed by the application associated with application window 702.
- the expanded privacy indicator 718-2 optionally provides selectable options for the user to disable access to the sensor, such that the application is no longer able to use the sensor to collect sensitive data. In some embodiments, in accordance with a determination that two or more sensors are currently being accessed by the application, the expanded privacy indicator 718-2 lists, or otherwise indicates, each of the sensors that are being accessed by the application(s) associated with the application window 702.
- Figure 7V illustrates detecting that the user’s attention 710-19 is directed to the grabber 706-2 for the application window 702.
- the computer system in response to detecting that the user’s attention 710-19 is directed to the grabber 706-2 (e.g., in Figure 7V), the computer system displays a grabber 706-3 (e.g., updates the display of the grabber 706-2 to the grabber 706-3, and/or replaces display of the grabber 706- 2 with the grabber 706-3) with a color distinct from the color of grabber 706-2 to indicate that the computer system detects that the user’s attention was directed to grabber 706-2 (e.g., and is directed to the currently displayed grabber 706-3), similar to the change in visual properties of the resize affordance 708-2 described with reference to Figures 7F-7G .
- Figure 7X illustrates detecting a user input via the user’s hand 7020 selecting a grabber 706-4 (e.g., which is an updated display of the grabber 706-3 and/or replaces the grabber 706-3).
- the user input is an air tap gesture, an air pinch gesture, or another selection gesture (e.g., air gesture) detected while the user is gazing at grabber 706-4.
- the user input in response to detecting the user input selecting the grabber 706-4, the user input continues by moving the user’s hand 7020 to drag (e.g., via an air drag gesture) the grabber 706-4 from its respective position to another position within the three-dimensional environment.
- the user is enabled to move the grabber 706-4 and associated application window 702 in three dimensions, including changing the position of the application window 702 in a horizontal direction (e.g., left and/or right), in a vertical direction (e.g., up and/or down), and/or in a depth (e.g., forwards and/or backwards) in the three-dimensional environment.
- a horizontal direction e.g., left and/or right
- a vertical direction e.g., up and/or down
- a depth e.g., forwards and/or backwards
- Figure 7Y illustrates that, in response to the user’s movement (e.g., from left to right and/or optionally away from the user’s body) of the user’s hand 7020 while a grabber 706-5 (e.g., which is analogous to the grabber 706-4, but represents the grabber as the user continues to interact with the grabber) is selected (e.g., and user’s attention 710-22 is detected as directed to the grabber 706-5), the application window 702 moves in the three-dimensional environment in accordance with a position and/or amount of the movement of the user’s hand 7020. For example, if the user moves the user’s hand to the left, the application window 702 moves to the left in accordance with the user’s hand movement.
- the title bar 716 and/or the privacy indicator 718-1 continue to be displayed at a same respective position relative to the application window 702 as the user moves the application window 702 in the three-dimensional environment.
- Figure 7Z illustrates that, in response to detecting an end of a user input, for example the user releases a pinch gesture, or air pinch gesture (e.g., the user’s movement of the user’s hand 7020) directed to the grabber 706-5, stops moving the user’s hand 7020, or liftoff (e.g., of a user input directed to the grabber 706-5), the grabber 706-6 is displayed without the updates to the one or more visual properties (e.g., the color of the grabber 706-6 is the same as the color of grabber 706-2 in Figure 7V, before detecting user input and/or attention directed to grabber 706-2).
- a pinch gesture e.g., the user’s movement of the user’s hand 7020
- liftoff e.g., of a user input directed to the grabber 706-5
- the grabber 706-6 is the same as the grabber 706-2 (e.g., the computer system 101 redisplays the grabber 706-2).
- Figure 7Z also illustrates that the application window 702 has been repositioned (e.g., using the grabber 706-5) to a different position in the three-dimensional environment.
- Figure 7Z further illustrates that the title bar 716 and the privacy indicator 718-1 continue to be displayed at their same respective positions relative to the application window 702, even while the application window 702 was being moved (e.g., and after the application window 702 has been moved). For example, the title bar 716 continues to be displayed above the application window 702 at the center of the application window 702, and the privacy indicator 718-1 is displayed to the right of the title bar 716 and above the application window 702.
- a size of the title bar 716 and/or privacy indicator 718-1 is based on a size of the application window 702, and the respective sizes of the title bar 716 and/or the privacy indicator 718-1 (e.g., and/or the privacy indicator 718-2) change as the application window 702 is resized and/or is positioned as farther away or closer to the user within the three- dimensional environment (e.g., where the application window 702 appears larger in size as it is positioned closer to the viewpoint of the user and smaller in size as it is positioned farther from the viewpoint of the user).
- the privacy indicator 718-1 in accordance with a determination that the application associated with the application window 702 is no longer accessing one or more sensors of the computer system that correspond to sensitive data, the privacy indicator 718-1 ceases to be displayed. As such, the privacy indicator 718-1 indicates to the user, in real-time, if an application is accessing a one or more sensors of the computer system.
- one or more additional indications are displayed above (or along another edge of) the application window 702.
- a sharing indication 719 is optionally displayed above the application window 702, optionally to the right and/or left, or otherwise proximate to, the title bar 716.
- the sharing indication 719 includes an indication of whether the content displayed in the application window 702 is shared by the user of computer system 101 with one or more other users.
- the sharing indication 719 includes names, avatars, initials, and/or another indication of one or more other users that are able to view the content displayed in the application window 702 (e.g., at a respective computer system used by the other user).
- the sharing indication 719 includes an indication of another user that is sharing content in the application window 702, if the content displayed in the application window 702 is being shared by another computer system used by the other user (e.g., the user of computer system 101 can view and/or interact with the application window 702, but optionally does not execute the application associated with application window 702). In some embodiments, if the content displayed in application window 702 is not shared with any other users and/or other devices, sharing indication 719 indicates that application window 702 is private by displaying text and/or an icon in sharing indication 719 and/or sharing indication 719 optionally ceases to be displayed.
- Figure 7Z illustrates an additional control option 721.
- one or more additional controls are displayed for the application window 702.
- the additional control option 721 is a control for opening another application window and/or tab for the application window 702.
- the other control options are displayed proximate to the title bar 716 and/or are displayed along other edges of the application window 702 (e.g., to the right and/or left and/or under the application window 702).
- the other control options include an option to enable and/or disable a gaze cursor. For example, while the gaze cursor is enabled, a cursor is displayed, via display generation component 7100, indicating the location corresponding to the user’s attention (e.g., where the user is gazing).
- the other control options include an option to display the application window 702 with a landscape orientation and/or a portrait orientation.
- Figure 7AA illustrates that, in response to detecting that the user’s attention 710-23 directed to the title bar 716 (e.g., as shown in Figure 7Z), an expanded title bar 716-2 is displayed (e.g., by increasing a size of the title bar 716).
- the expanded title bar 716-2 includes one or more control options for interacting with the application window 702.
- the one or more control options displayed in the expanded title bar 716-2 depend on the current content displayed in the application window 702.
- control options are based on the respective application and/or content in the application window 702.
- the expanded title bar 716-2 expands downward, such that it at least partially overlaps the application window 702. In some embodiments, the title bar 716-2 expands to the side (e.g., to the left and/or right) of the application window 702 and/or expands in another manner such that it does not obscure or overlay a portion of the application window 702.
- the user selects a respective control of the one or more control options displayed in expanded the title bar 716-2.
- the user selects a respective control by directing the user’s attention 710-24 to the respective control and performing a user input to select the control (e.g., a gaze and air pinch gesture, a gaze and air tap gesture, or another selection input optionally coupled with a gaze input).
- a user input e.g., a gaze and air pinch gesture, a gaze and air tap gesture, or another selection input optionally coupled with a gaze input.
- the user’s hand 7020 performs a selection input (e.g., an air pinch gesture or an air tap) while the user’s attention 710-25 is directed to the control “Open App 2” (e.g., as shown in Figure 7AB).
- an application window 720 is displayed in the three-dimensional environment, as illustrated in Figure 7AB.
- the application window 720 is associated with App 2.
- the application window 720 and the application window 702 are associated with the same application (e.g., two windows for the same application are displayed).
- the application window 720 and the application window 702 are associated with distinct applications.
- the computer system temporarily displays one or more of the available controls (e.g., close affordance, resize affordances, sharing indication, and/or other controls) at their respective positions to show the user that the controls are available (e.g., concurrently with displaying the application window 720).
- the one or more available controls are displayed for a threshold amount of time (e.g., 2 seconds, 3 seconds, or another amount of time), and then cease to be displayed (e.g., unless the user’s attention is directed to one of the available controls and/or the user’s attention is redirected back to one of the available controls within the threshold amount of time).
- ceasing to display the one or more available controls includes the available controls fading out and/or gradually disappearing, and/or displaying another animation as the one or more available controls cease to be displayed.
- the one or more available controls gradually move toward the application window 720 (e.g., decrease a distance between the available controls and the application window 720), until the one or more available controls overlap with (e.g., or disappear behind) the application window 720, optionally while gradually fading the controls.
- the application window 702 in response to displaying the application window 720 (e.g., or while the application window 720 is displayed), the application window 702 is visually deemphasized in the three-dimensional environment.
- the application window 702 is optionally pushed backwards (e.g., to appear father away from the user) in the three-dimensional environment, displayed with a different level of translucency, and/or dimmed relative to the three-dimensional environment and/or relative to the application window 720.
- the application window 702 is displayed with a visual deemphasis relative to the application window 720.
- the title bar 716 and/or the other controls are also visually deemphasized (e.g., in a same manner or a different manner as application window 702).
- the privacy indicator 718-1 is not visually deemphasized while the application window 702 is visually deemphasized.
- a size and/or translucency of the privacy indicator 718-1 is not changed in response to displaying the application window 720.
- the grabber 706-6 is optionally displayed under the application window 702, even while the application window 702 is visually deemphasized.
- the application window 720 is displayed with a title bar 722 (similar to the title bar 716 described above for the application window 702) that indicates the content displayed in the application window 720.
- the application window 720 is displayed with a grabber 726, which includes the same functionality described above with respect grabber 706-1.
- the application associated with the application window 720 is accessing one or more sensors that are associated with sensitive data.
- the computer system 101 displays a privacy indicator 724-1 above the application window 720.
- the privacy indicator 724-1 is displayed at a same position relative to the application window 720 as the position of the privacy indicator 718-1 relative to the application window 702.
- the privacy indicator 724-1 is displayed with a respective color based on the type of sensor that is being accessed by the application associated with the application window 720.
- the privacy indicator 724-1 is displayed with a first color associated with a first type of sensor (e.g., a camera) while the application associated with the application window 720 is accessing the first type of sensor, and the privacy indicator 718-1 is displayed with a second color associated with a second type of sensor (e.g., a microphone) while the application associated with application window 702 is accessing the second type of sensor.
- a first type of sensor e.g., a camera
- a second type of sensor e.g., a microphone
- Figure 7AB further illustrates detecting the user’s attention 710-25 directed to the title bar 722 of the application window 720.
- the computer system in response to detecting the user’s attention is directed to the title bar 722, and optionally in accordance with a determination that the user’s attention 710-25 remains directed to the title bar 722 for a threshold amount of time (e.g., or satisfies other attention-based criteria with respect to the title bar 722), the computer system displays an expanded title bar 722-2 for the application window 720, as illustrated in Figure 7AC.
- the expanded title bar 722-2 includes one or more control options.
- the user selects, using the user’s hand 7020 in combination with a gaze input (e.g., user’s attention 710-26), “Control C” from the expanded title bar 722-2.
- the computer system 101 in response to detecting user selection of Control C, performs an operation associated with Control C, for example by displaying a user interface object C 728 in the application window 720, as illustrated in Figures 7AD.
- Figures 7AD illustrates that while the user’s attention 710-27 is directed to the application window 720, in response to detecting that the application associated with the application window 720 is no longer accessing a sensor of the one or more sensors of the computer system 101, the privacy indicator 724-1 ceases to be displayed for the application window 720.
- the privacy indicator 718-1 continues to be displayed.
- the application windows 702 and 720 display distinct privacy indicators associated with the respective application window, such that the user is provided with information about which application is using a sensor at any given time. Further, because the respective privacy indicator moves with its respective application window within the three-dimensional environment, the user can easily identify the particular application that is accessing a sensor or other sensitive data.
- Figure 7AE further illustrates detecting the user’s attention 710-28 directed to a top-center portion of the three-dimensional environment.
- the computer system in response to detecting that the user’s attention 710-28 is directed to the top-center portion of the three- dimensional environment, the computer system displays a dot indicator 730 that represents a launch control for launching a system control platter.
- the computer system in response to detecting the user’s attention 710-29 directed to the dot indicator 730 for a threshold amount of time (e.g., the user’s gaze lingers on dot indicator 730), the computer system displays a system control platter 734, as illustrated in Figure 7AG.
- the system control platter 734 is displayed proximate to the dot indicator 730.
- the dot indicator 730 is optionally not displayed while the system control platter 734 is displayed.
- the system control platter 734 includes one or more selectable controls for controlling and/or performing system functions of the computer system 101.
- the system control platter 734 optionally includes a selectable control for changing a system volume, a selectable control for performing a search function on the computer system 101, a selectable control for viewing notifications, a selectable control for viewing and/or modifying one or more settings of the computer system, and/or a selectable control for invoking a virtual assistant of the computer system.
- system control platter 734 is concurrently displayed with system information, such as an indication of one or more sensors 732 that are being accessed by an application executing on the computer system 101, a connectivity indication (e.g., for WiFi and/or cellular connectivity), and/or a battery indication showing a battery level of the computer system.
- system information such as an indication of one or more sensors 732 that are being accessed by an application executing on the computer system 101, a connectivity indication (e.g., for WiFi and/or cellular connectivity), and/or a battery indication showing a battery level of the computer system.
- the indication of the one or more sensors 732 includes an icon displaying the type of sensor that is being accessed by one or more applications of the computer system.
- a sensor indication 732 includes a microphone icon (e.g., or a location icon, or a camera icon, and/or another sensor icon) to indicate the microphone is currently being accessed by one or more applications.
- an icon is displayed for each of the sensors that are currently being accessed by one or more applications.
- the computer system 101 in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) selecting the sensor icon 732, as illustrated in Figure 7AG, the computer system 101 displays a user interface element that indicates the names of applications that are currently accessing the sensor corresponding to the selected sensor icon 732. For example, in Figure 7AH, the computer system displays, in an expanded indicator 736, that the microphone is being accessed by App 1. In some embodiments, the expanded indicator 736 includes a listing of all of the applications that are currently accessing the respective sensor, such as the microphone.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such
- Figure 7AI illustrates an exemplary view of a three-dimensional environment, as described above with respect to Figure 7A, which includes a representation of a physical environment, in addition to multiple user interface elements.
- the view includes a representation of a physical environment, including a representation of a first wall 7006’, a representation of a floor 7008’, and a representation of a physical object 7014’.
- the view also includes a first title region 744 (also referred to herein as a title bar, such as title bar 716), a first content region 746 (e.g., optionally displayed in an application window, such as application window 702), a first user interface element 748, a second user interface element 750, and a first grabber 752.
- the view also includes a second title region 754, a second content region 756, a third user interface element 760, and a second grabber 756.
- a second title region 754 reference the first content region 746 and the first title region 744, but one skilled in the art would understand that the descriptions specific to the first content region 746 and/or the first title region 744 are readily applicable to any suitable user interface element (e.g., any of the user interface elements of Figure 7AI described above).
- the user’s attention 710-32 is directed to first content region 746.
- different user interface elements that are visible in the view of the three dimensional environment are displayed with different respective parameter values (e.g., respective brightness values, color values, or luminance values).
- the first content region 746 includes first content, and the first content is constrained to having an appearance in which the respective parameter has a value that is within a first range of values (e.g., the respective parameter has a value that is below T2, the respective parameter value is between Ti and T2).
- the first title region 744, that is concurrently displayed with the first content region 746 e.g., and optionally displayed with the grabber 752) has an appearance in which the respective parameter has a value that is outside the first range of values (e.g., above T2).
- the display generation component 7100 is a HMD, which displays a view of a three-dimensional environment that includes representations of the physical environment around the user.
- One or more, or all, of the representations of objects in the physical environment e.g., the representations 7004’ and 7006’ of walls , the representation 7008’ of the floor, the representation 7014’ of a physical object, and/or other representations of other objects in the physical environment
- the respective parameter value e.g., a brightness, a color, and/or a luminance
- the first range of values e.g., an appearance of the physical environment in the view of the three-dimensional environment is constrained to a first range of values for the respective parameter.
- one or more virtual objects e.g., objects that are not present in the physical environment
- the respective parameter has a value that is outside the first range of values (e.g., with a higher brightness value, a different color, and/or a higher luminance value, compared to the representations of the physical environment).
- This increases the visibility of the one or more virtual objects in the view of the three-dimensional environment (e.g., when the one or more virtual objects are displayed over the representations of the physical environment), regardless of a viewpoint of the user (e.g., which changes frequently during the use of a HMD).
- a user of the HMD interacts with the view of the three-dimensional environment through gaze gestures (e.g., and/or gaze gestures in combination with an air gesture or physical user input), and a virtual object that corresponds to the user’s attention (e.g., a location at which the user is gazing), has an appearance in which the respective parameter has a value that is outside the first range of values (e.g., as discussed in greater detail below, with reference to a visual indicator 764 of Figure 7 AM), which assists the user in identifying the location the computer system detects the user’s attention (e.g., gaze) is directed to.
- gaze gestures e.g., and/or gaze gestures in combination with an air gesture or physical user input
- a virtual object that corresponds to the user’s attention e.g., a location at which the user is gazing
- the respective parameter has a value that is outside the first range of values (e.g., as discussed in greater detail below, with reference to a visual indicator 764 of Figure 7
- the first title region 744 has an appearance in which the respective parameter has a value that is outside the first range of values (e.g., above T2) only if the first content is displayed within a sub-range of selectable values for the first content.
- the range of selectable values for the first content is any value below T2. If the respective parameter for the first content has a value that is below Ti, the first title region 744
- I l l has an appearance in which the respective parameter has a value that is not outside the first range of values (e.g., is below T2). If the respective parameter for the first content has a value that is above Ti but below T2 (e.g., the sub-range Ti to T2 is a sub-range of a range of values that is below T2), the first title region 744 has an appearance in which the respective parameter has a value that is outside the first range of values (e.g., is above T2).
- the respective parameter is a brightness
- the patterned background of the first content region 746 reflects a lower brightness value (e.g., corresponding to a “normal white,” which represents the brightest possible white value for content of the first content region 746) relative to the brightness value for the first title region 744, which has a pure white background (e.g., a brighter white than the brightest “normal white” for the first content region 746).
- the black bars displayed below the display generation component 7100 denote the range of selectable values for respective parameter for the first content region 746 and the first title region 763.
- the value for the first content region 746 which has selectable values for the respective parameter that are between Ti and T2.
- the selectable values for the respective parameter for the first title region 744 are above T2.
- the respective parameter is a color value
- the patterned background of the first content region 746 reflects a first color (or range of color values) that can be selected for the first content region 746.
- the pure white background of the first title region 744 reflects a color that is outside the possible colors that can be selected for the first content region 746 (e.g., some color values are reserved for use with the first title region 744, and reserved color values are not available for selection for the first content region 746).
- the first content region 746 is an application user interface (e.g., also referred to herein as an application window) for a first application, and the first application determines the first range of values that can be selected for the first content region 746. If the respective parameter is a brightness value, then the first application determines a first range of brightness values that can be selected for the first content region 746. For example, the first application determines that lower brightness values can be selected for the first content region 746. Higher brightness values are reserved (e.g., by the operating system, or by the first application) for the first title region 744. If the respective parameter is a color value, then the first application determines a first range of color values that can be selected for the first content region 746.
- the respective parameter is a color value
- the first application determines that blue color values can be selected for the first content region 746, while green color values are reserved (e.g., by the operating system, or by the first application) for the first title region 744 (e.g., to ensure sufficient color contrast between the first content region 746 and the first title region 744).
- the first title region 744 includes a first portion (e.g., a top half) and a second portion (e.g., a bottom half), and the first portion of the first title region 744 is displayed with a first value for the respective parameter that is outside the first range of values, and the second portion of the first title region 744 is displayed with a second value for the respective parameter that is within the first range of values (e.g., to provide a smoother visual transition between the first content region 746 and the first portion of the first title region 744).
- a first portion e.g., a top half
- a second portion e.g., a bottom half
- the first portion of the first title region 744 is displayed with the first value for the respective parameter that is outside the first range of values, which includes changing an appearance of the first portion of the first title region 744 due to a simulated lighting effect (e.g., a specular highlighting effect, a glow effect, or a halo effect) that is applied to the first title region 744 (e.g., and localized to the first portion of the first title region).
- a simulated lighting effect e.g., a specular highlighting effect, a glow effect, or a halo effect
- the first portion of the first title region 744 is displayed with the first value for the respective parameter that is outside the first range of values, which includes changing an appearance of the first portion of the first title region 744 due to presence of a visual indication (e.g., a visual indicator 764 described below with reference to Figures 7AM-7AN (e.g., Figures 7AN1, 7AN2 and 7AN3)) of a location of a gaze of the user.
- a visual indication e.g., a visual indicator 764 described below with reference to Figures 7AM-7AN (e.g., Figures 7AN1, 7AN2 and 7AN3)
- a first portion of a respective user interface element is displayed with the first value for the respective parameter that is outside the first range of values (and a second portion of the respective user interface element is displayed with a second value for the respective parameter that is within the first range of values), for respective user interface elements that are continuously displayed (e.g., regardless of user inputs and/or where the user’s attention is currently directed), such as the first title region 744, the second title region 754, the first grabber 752, and/or the second grabber 758.
- the first title region 744 includes more than two portions, and at least one portion of the first title region of the first title region 744 is displayed with a respective value for the respective parameter that is outside the first range of values, and at least one portion the first title region 744 is displayed with a respective value for the respective parameter that is within the first range of values.
- multiple user interface elements have an appearance in which the respective parameter has a value that is outside the first range of values, and the descriptions and examples above are applicable to each of the multiple user interface elements.
- the dot indicator 730, the first grabber 752, the second title region 754, and/or the second grabber 758 have respective values for the respective parameter that are outside the first range of values (e.g., all at or above T2), and have appearances analogous to that of the first title region 744 as described above.
- one or more, and optionally all of, user interface elements of a particular type have an appearance in which the respective parameter has a value that is outside the first range of values.
- detecting the user’s attention 710-33 directed to the bottom right corner of the first content region 746 includes detecting a gaze input, an air gesture, or a combination of gaze input and air gesture.
- the user’s attention 710-33 represents a position of a cursor associated with an input device, or represents a gaze input of the user.
- the animated transition includes displaying the first grabber 752 gradually shifting to the right, until it is replaced with the resize affordance 753 displayed at the bottom right corner of the first content region 746 (e.g., in a similar manner as shown in, and described above with reference to, Figure 7E).
- the resize affordance 753 is displayed with an appearance in which the respective parameter has a value that is outside the first range of values (e.g., in addition to, and concurrently with, the first title region 744, which also has an appearance in which the respective parameter has a value that is outside the first range of values).
- the value for the respective parameter for the resize affordance 753 is the same as the value for the first title region 744 (e.g., as described above with reference to Figure 7AI).
- the range of selectable values for the respective parameter for the resize affordance 753 is the same as the selectable range of values for the respective parameter for the first title region 744 (e.g., as shown by the identically sized black bars for the value of 744 and the value of 753, in Figures 7AI and 7AJ, respectively).
- the range of selectable values for the respective parameter for the resize affordance 753 is different from the range of selectable values for the respective parameter for the first title region 744 (e.g., optionally the range of selectable values can overlap, such that, for example, both ranges of selectable values are above T2).
- the first grabber 752 is not displayed with an appearance in which the respective parameter has a value that is outside the first range of values, but the resize affordance 753 is displayed with an appearance in which the respective parameter has a value that is outside the first range of values (e.g., to provide visual feedback that the first grabber 752 has been replaced with the resize affordance 753).
- the first grabber 752 is displayed with an appearance an appearance in which the respective parameter has a value that is outside the first range of values (e.g., prior to being replaced by the resize affordance 753), and the resize affordance 753 is displayed with an appearance in which the respective parameter has a value that is outside the first range of values (e.g., after replacing the first grabber 752).
- the resize affordance 753 ceases to be displayed (e.g., and the first grabber 752 is redisplayed) when the computer system detects that the user’s attention is no longer directed to the lower right corner of the first content area 746 (e.g., as shown in Figure 7AK, where the user’s attention 710-34 is directed to the upper right corner of the second content region 756, the first grabber 752 is redisplayed, and the resize affordance 753 is no longer displayed).
- the computer system detects the user’s attention 710-34 directed to an upper right corner of the second content region 756.
- detecting the user’s attention 710-34 directed to the upper right corner of the second content region 756 includes detecting a gaze input, an air gesture, or a combination of gaze input and air gesture.
- user’s attention 710-34 represents a position of a cursor associated with an input device.
- the computer system In response to detecting the user’s attention 710-34 directed to the upper right corner of the first content region 746, the computer system displays a close affordance 762 (e.g., a close affordance analogous to the close affordance 7030 described above with reference to Figure 7C) for ceasing display of the first content region 746.
- the close affordance 762 is displayed with an appearance in which the respective parameter has a value that is outside the first range of values (e.g., in addition to, and concurrently with, the first title region 744, which also has an appearance in which the respective parameter has a value that is outside the first range of values).
- the range of selectable values for the respective parameter of the close affordance 762 being above T2.
- the close affordance 762 ceases to be displayed when the computer system detects that the user’s attention is no longer directed to the upper right corner of the second content area 760.
- the content in the second content region 756 changes (e.g., the appearance of the content in the second content region 756 is different in Figure 7AL, as compared to the content in the second content region 756 in Figure 7AK).
- the range of selectable values for the respective parameter for the second content region 756 is changed such that the range is between T3 and T4 (e.g., instead of ranging between Ti and T2, as in Figure 7AK, where T3 and T4 are distinct from Ti and T2).
- the content is dynamic content (e.g., a media item such as a video), which changes over time, including a change in the brightness, color, and/or luminance of the content over time.
- the content in the second content region 756 changes in response to a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) (e.g., in response to the user interacting with the content).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input e.g., in response to the user interacting with the content.
- the content in the second content region 756 changes in response to detecting occurrence of an event (e.g., a night mode or dark mode may be activated in response to detecting that it is a certain time (e.g., 7:00 PM)).
- the content in the second content region 756 changes in response to detecting a change in the physical environment (e.g., the computer system detects that the user has moved to a physical environment with an increased amount of and/or brighter lighting, such that a brightness of the content is automatically, without user input, increased to maintain visibility).
- the close affordance 762 in response to the content in the second content region 756 changing over time, is displayed with an updated value for the respective parameter that is outside the first range of values based on a current state of the content in the second content region.
- the updated value for the respective parameter of close affordance 762 is selected in accordance with the change in the respective parameter of the content in the second content region 756. For example, if the respective parameter is brightness, and the content in the second content region 756 becomes brighter (e.g., as shown by the lower end of the range of selectable values, T3.
- the updated value for the brightness parameter of the close affordance 762 is higher than the original brightness parameter value for the close affordance 762 (e.g., the range of selectable values for the close affordance 762 is shifted upwards by the same (or proportional) amount as the difference between T3 and Ti).
- the close affordance 762 is displayed with the updated value for the respective parameter that is outside the first range of values, in accordance with a determination that the content in the second content region is displayed with an appearance in which the respective parameter is within a first sub-range of values (e.g., the sub-range T3 to T4, which is a sub-range of the range Ti to T2).
- a first sub-range of values e.g., the sub-range T3 to T4, which is a sub-range of the range Ti to T2.
- Figure 7AM shows the visual indicator 764, which reflects a location in the view of the three-dimensional environment that corresponds to the user’s attention (e.g., a detected gaze of the user).
- the visual indicator 764 changes location as the user’s attention moves between different locations (e.g., the visual indicator 764 moves in accordance with the user’s gaze).
- the visual indicator 764 is displayed as overlaid with other visible user interface elements (e.g., the visual indicator 764 appears “on top” of any other user interface element and/or passthrough of the three-dimensional environment at the location of the visual indicator 764).
- the visual indicator 764 is displayed with an appearance in which the respective parameter has a value that is outside the first range of values (e.g., to improve the visibility of the visual indicator 764, such that the user can easily identify where the computer system detects the user’s attention as being focused).
- the range of selectable values for the respective parameter of the visual indicator 764 is above T2.
- the visual indicator 764 is a circle (e.g., as shown in Figure 7AM), or another shape.
- the visual indicator 764 is another suitable shape (e.g., a square, a diamond, or a cross shape).
- the visual indicator 764 includes a virtual lighting effect (e.g., a specular highlighting effect, a glow effect, or a halo effect).
- a feathering effect is applied to the visual indicator 764 (e.g., a smoothing and/or blurring effect is applied near the edges of, or to the entirety of, the visual indicator 764).
- the visual indicator 764 is at least partially transparent (e.g., such that content in the first content region 746, over which the visual indicator 764 is displayed, is at least partially visible).
- Figures 7AN e.g., Figures 7AN1, 7AN2 and 7AN3
- the user’s attention is directed to the second user interface element 750.
- the visual indicator 764 has a default shape (e.g., a circle shape, as in Figure 7AM), and in response to the user’s attention being directed to second user interface element 750, the shape of the visual indicator 764 is at least partially altered to conform to the outline of the second user interface element 750 (e.g., the left and bottom edges of the second user interface element 750).
- the shape of the visual indicator 764 may be fully altered to conform to the outline of the second user interface element 750 (e.g., the default shape of the of the visual indicator 764 is a circle that is large enough to encompass the entirety of the second user interface element 750, and the altered shape of the visual indicator 764 is a rectangle that matches the boundaries of the second user interface element 750).
- the visual indicator 764 is displayed as only appearing as overlapping second user interface element 750, without displaying portions of the visual indicator 764 that are positioned outside of the boundaries of second user interface element 750.
- the second user interface element 750 (e.g., and/or the first user interface element 748) is an affordance or control element for performing functions with the first content region 746.
- the first content region 746 is an application user interface
- the second user interface element 750 (e.g., and/or the first user interface element 748) is a button, slider, or other control for performing functions (e.g., navigating within the application user interface, navigating to a different user interface, opening a new tab or window for the corresponding application, controlling media playback, starting a communication session, and/or adjusting settings) in the application user interface.
- the first user interface element 748 and/or the second user interface element 750 are displayed as overlaying the first content region 746, and the first user interface element 748 and the second user interface element 750 are displayed as a virtual augmentation (e.g., as augmenting portions of the content of the first content region 746, over which the first user interface element 748 and/or the second user interface element 750 are displayed).
- the first content region 746 is a representation of a virtual three-dimensional environment (e.g., a virtual beach that includes sand and waves), the first user interface element 748 is displayed as a virtual augmentation that makes sand of the virtual beach appear to sparkle, and the second user interface element 750 is displayed as a virtual augmentation that makes waves of the virtual beach appear to sparkle (e.g., optionally in a different manner than the sand).
- the first content region 746 includes a representation of a physical environment (e.g., and the first user interface element 748 and/or the second user interface element 750 augment representations of physical objects/features of the physical environment).
- the first content region 746 includes a photo and/or video, and the first user interface element 748 and/or the second user interface element 750 augment one or more portions of the photo and/or video.
- Figure 7AO shows an alternative view of the three-dimensional environment, which includes a text object 766 displayed as overlaying a portion of the representation of the physical environment.
- the text object 766 includes a title of a user interface element, such as an application name corresponding to an application icon, a contact name corresponding to an avatar for a user, a document name corresponding to a file, a title or description corresponding to a physical or virtual object that is visible in the view of the three-dimensional environment, and/or a textual indication of a virtual environment/experience.
- the user interface object 766 includes text of media content, such as lyrics to a song that is playing, or closed captioning or subtitles for audio/video media that is playing.
- At least a portion of the representation of the physical environment is constrained to have an appearance in which the respective parameter has a value that is within a first range of values, as illustrated by the patterned fill for the representation of the wall 7006’ and by the value of 7006’ being between Ti and T2.
- the text object 766 has an appearance in which the respective parameter has a value that is outside the first range of values (e.g., to ensure the text object 766 is visible when displayed over the representation of the wall 7006’), as shown by the pure white fill of the letters in the text object 766, and the value of 766 being above T2.
- a user interface 769 (e.g., a third content region) overlays (e.g., occludes, from the current viewpoint of the user) a region 767 of the text object 766, and the region 767 of the text object 766 has an appearance in which the respective parameter has a different (e.g., lower) value (e.g., as compared to the other portions of the text object 766).
- the very bright appearance can cause visibility issues for the user interface 769 that is displayed over a portion of the text object 766, so the overlaid portion of the text object 766 is displayed with a different appearance (e.g., dimmed) to avoid visibility issues.
- the different value is also outside the first range of values (e.g., but less so than the value for the respective parameter for the other portions of the text object 766 that are not overlaid or occluded). In some embodiments, the different value is within the first range of values.
- the appearance of the text object 766 is updated to reflect the relative movement. For example, portions of the text object 766 that were not previously overlaid or occluded, that become overlaid or occluded as a result of the movement of the user interface 769, are displayed with the appearance in which the respective parameter has the different value. Portions of the text object 766 that were previously overlaid or occluded, that are no longer overlaid or occluded as a result of the movement of the user interface 769, are displayed with the appearance in which the respective parameter has the value that is outside the first range of values.
- Figure 7AP is analogous to Figure 7AO, but shows that in some embodiments, the text object 766 has a value for the respective parameter that is below (and outside) the first range of values (e.g., the value of 766 is below Ti). For example, if the respective parameter is a brightness value, then in Figure 7AO, the text object 766 appears brighter (e.g., more white) than the representation of the physical environment (e.g., the representation of the wall 7006’, over which the text object 766 is displayed). In Figure 7AP, the text object appears darker (e.g., more black) than the representation of the physical environment (e.g., the representation of the physical object 7014’, over which the text object 766 is displayed).
- the respective parameter is a brightness value
- the text object 766 appears brighter (e.g., more white) than the representation of the physical environment (e.g., the representation of the wall 7006’, over which the text object 766 is displayed).
- the text object appears darker (e.g.,
- the user interface 769 overlays a different region 771 of the text object 766.
- the region 771 has an appearance in which the respective parameter has a different (e.g., higher) value (e.g., from the rest of the text object 766), and has analogous properties and behavior as the region 767 described with reference to Figure 7AO (e.g., except that the region 771 is shown in Figure 7AP as being less dark than the rest of the dark text object 766, in contrast to the region 767 which is shown in Figure 7AO as being less bright than the rest of the bright text object 766).
- Figure 7AQ illustrates a three-dimensional object 770a (e.g., a virtual chair or another virtual object).
- three-dimensional object 770a is displayed in the three-dimensional environment in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) requesting to view three-dimensional object 770a (e.g., in response to a request to open an application associated with three-dimensional object 770a and/or selection of three-dimensional object 770a from an application).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed
- three-dimensional object 770a is displayed with a platter 772.
- three-dimensional object 770a is positioned atop platter 772.
- platter 772 is a two-dimensional object (e.g., or a three-dimensional object with a small thickness (e.g., 1 mm, 5 mm, or another thickness)) that extends backwards (e.g., in the z-dimension and/or the depth dimension) in the three-dimensional environment.
- platter 772 extends such that it appears as a flat surface under three-dimensional object 770a.
- platter 772 is displayed as parallel, or substantially parallel, to the horizon (e.g., the floor or the latitudinal plane relative to the display of computer system 7100).
- a simulated shadow and/or reflection of three- dimensional object 770a is displayed on platter 772.
- the simulated shadow and/or reflection of three-dimensional object 770a on platter 772 automatically, without additional user input, moves in accordance with the movement of the three-dimensional object 770a and, optionally, a direction of a simulated or real light source.
- three-dimensional object 770a is associated with one or more object management controls for the object 770a.
- object 770a is displayed concurrently with grabber 7104a for moving a position of object 770a within the three- dimensional environment (e.g., grabber 7104a for object 770a having the same, analogous, or different, functionality as described with reference to grabber 706-1 for application window 702), and/or with affordance 774-1 (e.g., a minimized close affordance and/or a minimized controls menu).
- affordance 774-1 e.g., a minimized close affordance and/or a minimized controls menu
- additional and/or alternative object management controls are optionally displayed with object 770a, such as a resize affordance.
- object management controls for object 770a are available (e.g., for the user to interact with and/or perform operations associated with the respective object management controls) but optionally not displayed (e.g., until the user’s attention is directed to a respective region for the respective object management control), as described above with reference to Figures 7D1-7D2.
- the one or more object management controls include platter 772.
- platter 772 is displayed under object 770a, regardless of whether the user’s attention is directed to object 770a (e.g., and/or a region around object 770a).
- platter 772 is displayed in response to detecting that the user’s attention is directed to object 770a and optionally ceases to be displayed in accordance with a determination that the user’s attention is no longer directed to object 770a.
- the one or more object management controls are displayed concurrently with object 770a without detecting a user input.
- the one or more object management controls are displayed in response to detecting a user input directed to object 770a, for example, in response to detecting a gaze input (e.g., directed to a bottom portion, or other respective portion or region, of the object 770a), and/or another gesture (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to object 770a.
- a gaze input e.g., directed to a bottom portion, or other respective portion or region, of the object 770a
- another gesture e.g., a direct air gesture such as an air tap or air pinch at a location with which the user
- the one or more object management controls are optionally not displayed concurrently with the object 770a until a user input is detected as being directed to the object 770a (e.g., a gaze input directed to a bottom of object 770a or another portion of object 770a).
- the one or more object management controls are displayed in accordance with a determination that the gaze input is maintained for a threshold amount of time (e.g., the gaze input satisfies a gaze criterion).
- the one or more object management controls are displayed with an orientation that faces the user’s current viewpoint (e.g., such that the user can more easily read and/or interact with the one or more object management controls without requiring the user to change position and/or change the user’s viewpoint).
- the one or more object management controls are displayed with an orientation selected based at least in part on the user’s current viewpoint, and are displayed with a first spatial relationship relative to the object 770a (e.g., at a first size and at a first angle/orientation relative to the object 770a).
- the one or more object management controls are optionally updated to be displayed with the same orientation relative to the user’s current viewpoint, even while an orientation of the object 770a does not change relative to the user’s viewpoint.
- the one or more object management controls are displayed with a different spatial relationship (e.g., different orientation, angle and/or size), distinct from the first spatial relationship, relative to the object 770a.
- affordance 774-1 before detecting the user’s attention directed to affordance 774-1, affordance 774-1 is displayed as a minimized version of one or more object management controls, optionally displayed as a dot that is expanded in response to detecting the user’s attention (e.g., for at least a threshold amount of time) directed to the dot.
- a dot that is expanded in response to detecting the user’s attention (e.g., for at least a threshold amount of time) directed to the dot.
- affordance 774-2 is displayed. For example, if the user’s attention is detected for less than the threshold amount time, the dot does not expand to close affordance 774-2.
- affordance 774- 1 expands to display close affordance 774-2 and/or to display title bar 7029, as illustrated in Figure 7AR.
- close affordance 774-2 is an expanded version of affordance 774-1.
- close affordance 774-2 includes an icon and/or text (e.g., an “x” icon, or a “close” button).
- close affordance 774-2 in response to user selection of close affordance 774-2, object 770a ceases to be displayed (e.g., is closed and/or removed from the user interface). In some embodiments, close affordance 774-2 continues to be displayed while the user’s attention 710-37 continues to be detected as being directed to close affordance 774- 2.
- one or more additional object management controls are optionally displayed, such as title bar 7106 and/or other object management controls that were not displayed prior to detecting the user’s attention as directed to affordance 774-1.
- the one or more object management controls include a resize affordance (e.g., resize affordance 776), and/or an object menu that includes a plurality of selectable options for performing operations with respect to object 770a.
- one or more of the object management controls for object 770a are visually deemphasized (e.g., shrink in size, blur, decrease in luminance and/or cease to be displayed).
- grabber 7104a decreases in size to grabber 7104b ( Figure 7AR) while the close affordance 774-2 is displayed.
- grabber 7104b shrinks in size (e.g., to a decreased length) to create additional space for close affordance 774-2 (e.g., which is an expanded version of affordance 774-1), such that close affordance 774-2 does not overlap with grabber 7104b.
- title bar 7106 is optionally also displayed with a size and/or at a position such that it does not overlap with close affordance 774-2 while close affordance 774-2 is displayed.
- some of the object management controls, but not others, are visually deemphasized while the close affordance 774-2 (e.g., or another object management control) is visually emphasized (e.g., as an expanded version).
- the visual deemphasis of the one or more object management controls is undone. For example, grabber 7104b increases to the size of grabber 7104a while affordance 774-1 is displayed in its minimized version.
- close affordance 774-2 includes a plurality of options for closing one or more objects and/or application windows, as illustrated in Figure 7AS1.
- close affordance 774-2 expands to include menu 774-3.
- the plurality of options displayed in menu 774-3 include the options described with reference to Figures 7BC-7BL.
- the menu 774-3 is displayed in accordance with a determination that the user’s attention 774-2 is directed to affordance 774-1 (e.g., from the display in Figure 7AQ to Figure 7AS1 directly) and/or to close affordance 774-2 (e.g., from the display in Figure 7AQ to Figure 7AR to Figure 7AS1) for at least a threshold amount of time, or satisfies one or more other attention criteria.
- affordance 774-1 e.g., from the display in Figure 7AQ to Figure 7AS1 directly
- close affordance 774-2 e.g., from the display in Figure 7AQ to Figure 7AR to Figure 7AS1
- menu 774-3 is displayed in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to close affordance 774-2.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input e.g., a gaze input, a drag input, and/or another type of user input
- Figure 7AS2 illustrates detecting the user’s attention 710-38 directed to object 770a.
- a user input e.g., an air gesture or other input
- the object 770a is rotated.
- the user’s hand 7020 moves (e.g., in a drag gesture or other gesture) from left to right to rotate object 770a.
- the user provides an input for turning the object 770a such that the orientation of object 770a (e.g., relative to the current viewpoint of the user) changes.
- object 770a changes from a front view to a side view as object 770b, illustrated in Figure 7AT.
- close affordance 774-2, title bar 7106, grabber 7104b and/or one or more other object management controls continue to be displayed while the user interacts with object 770a.
- at least a subset of the object management controls e.g., a resize affordance and/or other controls
- Figure 7AT illustrates that, in some embodiments, the one or more object management controls (e.g., at least one of the object management controls and/or all of the object management controls) do not rotate and/or change in position or viewing angle relative to the user’s viewpoint.
- the one or more object management controls do not rotate with the object 770b.
- the one or more object management controls are displayed at a different orientation relative to the object 770b (as opposed to the orientation of the one or more object management controls displayed relative to the object 770a before rotating the object 770a).
- the one or more object management controls are displayed to maintain the same orientation relative to the user’s current viewpoint, without regard to the updated orientation of object 770b.
- the one or more object management controls are displayed with a second spatial relationship relative to the object 770b (e.g., at a second size and/or at a second angle/orientation relative to the object 770a).
- Figure 7AT illustrates the user’s attention 710-39 is directed to grabber 7104b.
- the computer system detects an input (e.g., an air gesture and/or other selection input) for moving object 770b using the grabber 7104b.
- object 770b is moved, within the three- dimensional environment 7000’, from a first position illustrated in Figure 7AT to a second position illustrated in Figures 7AU (e.g., Figures 7AU1, 7AU2 and 7AU3).
- the computer system continues to detect the input (e.g., a drag input, a swipe input, and/or another type of input) for moving object 770b and updates the display to gradually show object 770b moving in accordance with a speed, direction and/or amount of movement indicated by the input (e.g., the displayed position of the object 770b changes over a period of time). For example, in Figure 7AV, the object 770c appears closer to the viewpoint of the user and moved to the left relative to the position of object 770b in Figure 7AU.
- the input e.g., a drag input, a swipe input, and/or another type of input
- the close affordance 774-2 in accordance with a determination that the user’s attention is not directed to the close affordance 774-2 for at least a threshold amount of time (e.g., and/or that the user has not interacted with the close affordance 774-2 for at least a threshold amount of time), the close affordance 774-2 is visually deemphasized, for example, is redisplayed as its minimized version as affordance 774-1 (e.g., as illustrated in Figures 7AU (e.g., Figures 7AU1, 7AU2 and 7AU3)) and/or ceases to be displayed.
- Figures 7AU e.g., Figures 7AU1, 7AU2 and 7AU3
- the object 770c is displayed with a larger size (e.g., relative to object 770b) as the object 770c is moved closer to the viewpoint of the user.
- the size of object 770c is updated based on a perceived distance between object 770c and the viewpoint of the user.
- the one or more object management controls optionally update in size based on the perceived distance between the object management controls and the viewpoint of the user (e.g., optionally by a lesser amount than the amount that size of object 770c is updated).
- the displayed sizes of the object management controls are optionally not changed or change to a smaller extent than the change to the displayed size of the object 770c, such that the content of the one or more object management controls remains clear to the viewer (e.g., when the first virtual object gets farther away from the current viewpoint and becomes very small, the one or more object management controls also get farther away but increase in intrinsic size (e.g., the size of the one or more object management controls relative to the three-dimensional environment) so that the one or more object management controls are displayed at a size that is not too small for the user to see) and/or is not unnecessarily obtrusive (e.g., when the object 770c gets closer to the current viewpoint of the user and becomes very large, the one or more object management controls also get closer but shrink in intrinsic size so that their displayed size(s) are not overly large as perceived from the user’s current viewpoint).
- the displayed size of the object 770c is changed by a different amount than the displayed size(s
- the sizes of (e.g., and/or one or more other visual properties of) the one or more object management controls are updated gradually, over a period of time, as the object 770b is moved closer to the user, displayed as object 770c.
- the one or more object management controls change in size and/or position based on the movement of object 770b.
- the object management controls are displayed as gradually decreasing in size and/or the object management controls gradually increase in size relative to the three-dimensional environment (e.g., to a size that is selected based at least in part on their perceived size as they are moved away from and/or towards the user).
- the visual appearance e.g., amount of visual deemphasis updates gradually (e.g., changes over a period of time) as the object 770b is moved in the three-dimensional environment.
- Figure 7AV further illustrates that, in some embodiments, the object management controls are optionally visually deemphasized while the user is interacting with object 770c.
- the one or more object management controls such as affordance 774- 4, grabber 7104b and/or title bar 7106 are optionally blurred, faded, decreased in size and/or ceased to be displayed in Figure 7AV while object 770c is being moved within the three- dimensional environment.
- a subset, less than all, of the one or more object management controls are visually deemphasized while the user is interacting with object 770c.
- platter 772 and/or grabber 7104b are not visually deemphasized as the user moves the object 770c, while affordance 774-4 and/or title bar 7106 (e.g., and/or one or more other object management controls) are visually deemphasized.
- one or more of the object management controls for object 770a gradually, over a time period, change in appearance (e.g., affordance 774-4 is visually deemphasized and/or changes its angle relative object 770a) while the object 770a is moved in the three-dimensional environment.
- the one or more object management controls are no longer visually deemphasized and are returned to the same visual appearance as displayed before the object 770b was moved. For example, the one or more object management controls cease to be displayed as object 770b is being moved, and reappear after object 770b is displayed at its new position and is no longer being moved.
- Figure 7AW illustrates detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) to reposition the object 770d within the three-dimensional environment 7000’.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input e.g., a tap input, a gaze input, a drag input, and/or another type of user input
- Figure 7AX illustrates that object 770d (now labeled as 770e) is moved, within the three-dimensional environment, to a position that is closer to the viewpoint of the user.
- the user’s attention 710-42 is detected as being directed to a portion of platter 772.
- resize affordance 776 is displayed (e.g., at the respective comer), as shown in Figure 7AY.
- resize affordance 776 is displayed at a position that corresponds to the detected user’s attention 710-42.
- resize affordance 776 is displayed as shifting backwards in z-space. For example, in response to detecting that the user’s attention is directed to a back corner of platter 772 (e.g., that is displayed farther away, in z-space, from the user’s current viewpoint), the resize affordance appears at the respective back corner, and is displayed as being farther away, in z-space, from the user’s current viewpoint.
- Figure 7AY illustrates that, while displaying resize affordance 776, in response to detecting the user’s attention 710-43 directed to resize affordance 776 and/or in response to detecting a user input (e.g., an air gesture and/or other type of user input) from the user’s hand 7020, resize affordance 776 is selected such that the user is enabled to change the size of object 770f.
- resize affordance 776 optionally has at least some of the functionality of resize affordance 708-1, as described above with reference to Figures 7D1-7D3.
- grabber 7104a (e.g., and/or one or more other object management controls) is optionally not displayed concurrently with resize affordance 776 and/or an animated transition is displayed whereby grabber 7104a gradually shifts to the right and/or morphs into resize affordance 776.
- object 770f in response to the user input illustrated in Figure 7AY, object 770f is resized to a smaller size, illustrated as object 770g in Figure 7AZ.
- object 770g in response to detecting that the user’s attention is directed to the resize affordance 708-1 (e.g., detecting that the user is gazing at resize affordance 708-1), the computer system updates one or more visual properties of the resize affordance 708-1 to indicate that the computer system detects the user’s attention directed to the resize affordance 708-1 (e.g., and/or while the user performs the resizing operation, the appearance of the resize affordance is updated to indicate that the resize affordance is currently selected).
- Figure 7AZ further illustrates an arrow indicating user movement (e.g., and/or movement of the user’s viewpoint) in the physical environment.
- the three-dimensional environment is updated in Figures 7BA (e.g., Figures 7BA1, 7BA2 and 7BA3) to display the current view of the user.
- Figures 7BA e.g., Figures 7BA1, 7BA2 and 7BA3
- the user moves to the right in the physical environment and/or turns the user’s head to the right, such that the current view of the user is shifted to the right relative to the view of the user in Figure 7AZ.
- object 770h continues to be displayed at its position relative to the three-dimensional environment (e.g., object 770h is world-locked). For example, object 770h does not shift to the right in response to the user movement.
- the one or more object management controls are optionally maintained at a same position relative to object 770h in the current view.
- the angle (e.g., orientation) at which the one or more object management controls are displayed relative to the object 770h is not changed (e.g., and thus, the angle at which the one or more object management controls are viewed from the current viewpoint of the user is distinct in the current view relative to the previous view illustrated in Figure 7AZ).
- the angles of the one or more object management controls update in response to detecting that one or more criteria are met. For example, in response to detecting the user’s attention directed and object 770h and/or at the one or more object management controls, the one or more object management controls optionally shift to an angle that faces the user as perceived at the current viewpoint of the user.
- the one or more object management controls are also not displayed in the three-dimensional environment 7000’, but are optionally maintained such that if the user’s viewpoint changes to include (e.g., reinclude) object 770h, the one or more object management controls continue to be displayed with object 770h.
- Figure 7BB illustrates a second change in the current viewpoint of the user to a position behind object 770i (e.g., such that the back of the chair is in the current view).
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and
- the one or more object management controls are automatically updated to be displayed with preferred angle(s) and/or size(s) (e.g., that is determined based on a distance and/or angle between object 770i and the user’s current viewpoint) in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) for moving the object 770i and/or in response to detecting the user’s attention is directed to the object 770i.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is
- Figures 7BC-7BL illustrate examples of closing different applications or groups of applications in a three-dimensional environment in response to different inputs.
- Figure 14 is a flow diagram of an exemplary method 1400 for closing different applications or groups of applications in a three-dimensional environment in response to different inputs.
- the user interfaces in Figures 7BC-7BL are used to illustrate the processes described below, including the processes in Figure 14.
- Figures 7BC-7BE illustrate a close affordance displayed in proximity to a user interface of a first application (e.g., overlapping the user interface of the first application, or displayed with an offset from the user interface of the first application) and associated with the user interface of the first application.
- the close affordance includes or is activatable to cause display of an affordance to cease displaying other applications within a viewport of the three-dimensional environment.
- the close affordance is activatable for dismissing (e.g., closing or hiding) one or more user interfaces of the first application (e.g., including the main region and/or one or more, or all, auxiliary regions of the application user interface) and/or of other applications.
- Figure 7BC illustrates a viewport of a three-dimensional environment in which respective user interfaces of four applications are displayed.
- one or more additional user interfaces of additional applications are also displayed in the three- dimensional environment outside the viewport shown in Figure 7BC.
- the respective user interfaces of respective applications within the viewport include user interface 7202 of a first application, App A; user interface 7206 of a second application, App B; user interface 7204 of a third application, App C; and user interface 7208 application of a fourth application, App D.
- attention of user e.g., by a gaze input that meets a gaze dwell time threshold, a gaze input that has an associated movement that is below a movement threshold, and/or a gaze input that has a gaze velocity over a portion of the user interface of the application that is below a velocity threshold
- user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting a hand gesture, tap inputs, press inputs, pinch inputs, body movements, gaze input and drag input, and/or voice commands
- a user interface of an application e.g., user interface 7208 or a different user interface
- attention of the user is directed is visually emphasized, relative to one or more other user interfaces (e.g., user interfaces 7202, 7204, and 7206, or one or more of user interfaces 7202, 7204, and 7206) in the viewport to which attention of the user is not directed to (e.g., in that attention or gaze input toward the one or more other user interfaces is not sustained and does not persist beyond a gaze dwell time threshold).
- App A, App B, App C and/or App D are the same application (e.g., one or more of user interfaces 7202, 7204, 7206 and 7208 are different user interfaces of the same application).
- App A, App B, App C and App D are distinct applications.
- the user interfaces of respective applications in the viewport have similar visual appearance.
- Figure 7BC illustrates close affordance 7212-1 associated with user interface 7208 of App D.
- close affordance 7212-1 is displayed in proximity to, or within a threshold proximity of, user interface 7208 of App D as shown in Figure 7BC.
- user interface 7208 (e.g., the main user interface element, or a different portion of the user interface) has a dimension di along a first dimension or first edge (e.g., a long edge) and a dimension d2 along a second dimension or second edge (e.g., a short edge).
- Close affordance 7212-1 is displayed at a distance g2 away from an edge (e.g., a lower edge or a top edge) of user interface 7208, in a direction of or parallel to dimension d2, and at a distance gi away from an edge (e.g., a left edge or a right edge) of user interface 7208, in a direction of or parallel to dimension di.
- close affordance 7212-1 is displayed in proximity to user interface 7208 such that distance gi is less than 30%, 20%, 15%, 10%, 5%, 1%, or other threshold percentage of dimension di and/or distance g2is less than 30%, 20%, 15%, 10%, 5%, 1%, or other threshold percentage of dimension d2.
- gi extends left of the left edge of user interface 7208 instead of extending to the right of the left edge as shown in Figure 7BC.
- g2 extends above the bottom edge of user interface 7208 instead of extending below the bottom edge as shown in Figure 7BC, such that close affordance 7212-1 overlaps with a portion of user interface 7208 (optionally being offset from user interface 7208 along a simulated depth dimension, or a long a z- direction, such as in front of user interface 7208).
- close affordance 7212-1 is displayed next to a second auxiliary user interface element of user interface 7208.
- the second auxiliary user interface element is distinct from (and optionally separate from) the main user interface element (e.g., the main user interface element having the dimensions di and d2 as shown in Figure 7BC, or a different user interface element having different dimensions).
- the second auxiliary user interface element is also displayed within a threshold proximity to the main user interface element of user interface 7208, similar to close affordance 7212-1.
- Figure 7BC also illustrates user input 7210 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to close affordance 7212-1.
- user input 7210 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to close affordance 7212-1.
- computer system 101 In response to detecting user input 7210 directed to close affordance in 7212-1 ( Figure 7BC), and in accordance with a determination that user input 7210 is a first type of input, computer system 101 displays expanded close affordance 7212-2 and an option list 7216-1 corresponding to different options for closing one or more user interfaces of respective applications, as shown in Figures 7BD (e.g., Figures 7BD1, 7BD2 and 7BD3).
- Figures 7BD e.g., Figures 7BD1, 7BD2 and 7BD3
- Examples of the first type of user input include a gaze input that exceeds a dwell time threshold, a gaze input that has a gaze velocity that is below a velocity threshold, air gestures such as a long press or long pinch, a long tap, a tap sequence (e.g., one or more taps), a movement gesture that meets selection criteria (e.g., having a movement velocity below a threshold, and/or includes movement greater than distance threshold), a head motion (e.g., a head motion coupled with a gaze input and/or an air gesture), a hover input, or a passive input based on a body posture of the user that causes the close affordance to expand.
- air gestures such as a long press or long pinch, a long tap, a tap sequence (e.g., one or more taps), a movement gesture that meets selection criteria (e.g., having a movement velocity below a threshold, and/or includes movement greater than distance threshold), a head motion (e.g., a head motion coupled with a gaze input
- close affordance 7212-1 is replaced by expanded close affordance 7212-2 in response to: a gaze input directed to close affordance 7212-1 (e.g., directly on close affordance 7212-1, or in proximity (e.g., within a threshold distance of close affordance 7212-1, optionally within the distance gi or distance g2) for longer than a gaze dwell time threshold; a gaze input with a gaze velocity below a threshold velocity; and/or a hand gesture that includes a pinch gesture, with or without a movement input (e.g., a pinch or pinch and drag input and/or other air gesture).
- a gaze input directed to close affordance 7212-1 e.g., directly on close affordance 7212-1, or in proximity (e.g., within a threshold distance of close affordance 7212-1, optionally within the distance gi or distance g2) for longer than a gaze dwell time threshold
- a gaze input with a gaze velocity below a threshold velocity e.g., a pinch or pinch and drag input and/
- close affordance 7212- 2 is displayed in response to a first portion of user input 7210 directed to close affordance 7212-1 (e.g., including a gaze input directed to close affordance 7212-1) to indicate that the close affordance has focus for subsequent interaction.
- display of close affordance 7212-1 is replaced with display of close affordance 7212-2 prior to detection of a second portion of user input 7210 that includes a hand gesture (e.g., in response to which options list 7216-1 is displayed).
- a centroid of close affordance 7212- 2 is coincident with a centroid of close affordance 7212-1.
- computer system 101 displays an animated transition that transforms close affordance 7212-1 into close affordance 7212-2 by expanding close affordance 7212-1 to a size of close affordance 7212-2 and/or dynamically updating a graphical representation (e.g., a glyph or icon) displayed on or within close affordance 7212-2.
- a graphical representation e.g., a glyph or icon
- close affordances 7212-1 and 7212-2 are each displayed close to and/or collinear with another auxiliary user interface element of user interface 7208, such as grabber 7214.
- close affordance 7212-1 is displayed at a location outside a main user interface element (e.g., platter) of user interface 7208 (e.g., where content of App D is presented).
- option list 7216-1 displayed in response to the detection of user input 7210 and in accordance with the determination that user input 7210 is the first type of input, is a context menu and/or a pop-up user interface element.
- Options list 7216-1 includes one or more options for closing one or more user interfaces of respective applications.
- option 7216-2 e.g., labeled “Hide This App Window” is an option for hiding user interface 7208 associated with close affordance 7212-2 (e.g., the user interface associated with the close affordance to which user input 7210 was directed). Hiding a respective user interface is sometimes referred to hereinafter as hiding an application window.
- hiding a user interface of an application includes ceasing to display the user interface of the application and optionally switching the application from running in a foreground state to running in a background state. In some embodiments, while running the application in a background state, information about a state and/or content of the user interface of the application is retained. In some embodiments, hiding a user interface of an application includes closing the application. In some embodiments, closing the application includes terminating all processes associated with the application that are running on computer system 101 and/or not retaining information about a state and/or content of the user interface of the application.
- hiding a respective user interface of an application includes terminating processes associated with displaying content on the respective user interface, but allowing other processes associated with the application to continue (e.g., in a foreground state, or in a background state).
- hiding a user interface of an application includes continuing to run the application in a foreground state but without displaying the user interface for that application in a viewport of the three-dimensional environment. For example, in such an approach, the hidden user interface may be quickly brought back into view (e.g., in response to a request to unhide and redisplay the user interface).
- option list 7216-1 includes option 7216-3 (e.g., labeled “Hide All Windows of this App”) to hide all application user interfaces of the application associated with close affordance 7212-2.
- option 7216-3 is displayed in accordance with a determination that two or more user interfaces of the application associated with close affordance 7212-2 are open in the three-dimensional environment, optionally including scenarios in which one or more of such user interfaces are outside the viewport.
- option 7216-3 includes closing only user interfaces of the application associated with close affordance 7212-2 that are within the viewport.
- Figures 7BI-7BL further illustrate embodiments in which an option for closing two or more user interfaces of a respective application is provided (e.g., where the option for closing two or more user interfaces of the respective application is provided in response to detecting a second type of input that different from the first type of input).
- option 7216-3 is not displayed in option list 7216-1.
- option list 7216-1 includes option 7216-4 (e.g., optionally displayed as a second element in option list 7216-1 in scenarios in which option 7216-3 is not displayed and/or the order of the options displayed is changed).
- Option 7216-4 e.g., labeled “Hide Other Apps”
- hiding all other applications includes hiding all user interfaces in the three-dimensional environment that are within the viewport and that are user interfaces of applications other than the application associated with close affordance 7212-2.
- hiding all other applications includes hiding all user interfaces of applications in the three-dimensional environment other than user interfaces of the application associated with close affordance 7212-2, not just those within the viewport.
- Figures 7BD illustrates detecting user input 7218 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7216-4 to hide all other applications.
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to option 7216-4 to hide all other applications.
- Figure 7BD2 illustrates a direct air gesture is performed using a gesture performed by the user’s hand 7020 that is based on the position of the representation of the user’s hand 7020’ that is directed to a position, in the user interface, that intersects with the position of option 7216-4, and/or an indirect air gesture is performed (e.g., as an alternative to the direct air gesture and/or in addition to the direct air gesture) using a gesture performed by the user’s hand 7022 while the user’s gaze input 7218a is directed to option 7216-4, but is not based on the position of the representation of the user’s hand 7022’.
- user input 7218 is a same type of input as user input 7210.
- user input 7218 includes a different type of input from the first type of input.
- user input 7218 includes a quick pinch input, or a short tap or press input directed to option 7216-4.
- Providing an option to use a single input to hide all other applications allows a user to more quickly focus on an application of interest (e.g., hiding all other application user interfaces to watch a movie, or hiding all other application user interfaces to initiate an immersive communication session) without having to individually close each of the non-relevant applications.
- Figure 7BE illustrates that, in response to detecting user input 7218 directed to option 7216-4 ( Figures 7BD) for hiding user interfaces of other applications, different from the application associated with the close affordance 7212-2, computer system 101 ceases to display user interface 7202 of App A, user interface 7204 of App C, and user interface 7206 of App B.
- only user interface 7208 of App D remains in the viewport of the three-dimensional environment.
- no other application user interfaces except those associated with App D remain in the three-dimensional environment.
- other user interfaces of App D e.g., either in the viewport or outside the viewport
- other user interfaces of applications other than App D remain opened in the three-dimensional environment outside of the viewport, after detecting user input 7218 directed to option 7216- 4.
- Figures 7BE-7BH illustrate different example scenarios resulting from a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to a close affordance displayed in proximity to a user interface of an application and associated with the user interface of the application, based on whether any other application user interfaces remain in a viewport of the three-dimensional environment after the user interface of the application is closed in response to the user input.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- Figure 7BE illustrates option list 7216-1 displayed in response to activation of close affordance 7212-2 associated with user interface 7208 of application App D.
- Figure 7BE illustrates a transition from Figures 7BD upon detecting user input 7218 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to third option 7216-4 (e.g., the viewport illustrated in Figure 7BE follows from the detection of user input 7218 of Figures 7BD).
- user input 7218 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- option list 7216-1 is no longer displayed once an operation is performed (e.g., other applications are hidden) in response to detecting input 7218.
- display of option list 7216-1 is maintained for a first time period after user interfaces of other applications are hidden, and after the first time period, close affordance 7212-2 and option list 7216-1 cease to be displayed, and close affordance 7212-1 is redisplayed.
- a time interval between computer system 101 detecting user input 7218 and detecting user input 7220 is longer than the first time period.
- an additional user input e.g., an input of the first type, including a gaze input, an air gesture, a head gesture, and/or other types of input
- close affordance 7212 e.g., displayed as close affordance 7212-1, or as close affordance 7212-2 following an initial gaze input directed to close affordance 7212-1
- option list 7216-1 is displayed in response, as shown in Figure 7BE.
- display of option list 7216-1 is maintained from the scenario of Figures 7BD or redisplayed via a separate invocation, a user is enabled to continue interacting with option list 7216-1.
- Figure 7BE illustrates user input 7220 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7216-2 for hiding user interface 7208 associated with close affordance 7212-2.
- computer system 101 hides user interface 7208 of application App D, as shown in Figure 7BF.
- computer system 101 automatically displays home menu user interface 7222 without any additional user input.
- computer system 101 automatically displays home menu user interface 7222 without any additional user input in accordance with a determination that no user interface of any application is in the viewport (e.g., even if at least one user interface of an application is present in the three-dimensional environment outside the viewport shown in Figure 7BF).
- computer system 101 automatically displays home menu user interface 7222 without any additional user input when no user interface of any application is anywhere within the three-dimensional environment, including within and outside the viewport (e.g., computer system 101 would not automatically display home menu user interface 7222 if there is at least one application user interface outside the viewport, even if there is no user interface of any application in the viewport).
- Figure 7BF shows home menu user interface 7222 displaying a collection of representations 7224, 7226, 7228, 7230, 7232, 7234, 7236, and 7238 (also collectively referred to herein as representations 7224 - 7238).
- representations 7224 - 7238 correspond to various software applications that can be executed on computer system 101 (e.g., an email application, a web browser, a messaging application, a maps application, a video player, or an audio player, or other software application), representations of one or more other people with whom a user is able to initiate or maintain (e.g., continue or resume) communication or otherwise interact with, or who are capable of interacting with the user, and/or a collection of computer generated (e.g., augmented reality, virtual reality, mixed reality, or other extended reality) three-dimensional environments that can be displayed as (or in) the three-dimensional environment.
- computer generated e.g., augmented reality, virtual reality, mixed reality, or other extended reality
- user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a pinch input, a tap input, a gaze input, and/or other input
- a representation e.g., one of representations 7224 - 7238
- home menu user interface 7222 launches a software application associated with the representation, initiates or joins a communication session, or loads a computer-generated environment into the three-dimensional environment.
- Figure 7BG illustrates a different scenario from that depicted in Figure 7BE in which one or more additional user interfaces are displayed in the viewport (e.g., Figure 7BG is similar to Figures 7BD except that user input 7240 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) in Figure 7BG is directed to option 7216-2, instead of to option 7216-4 as user input 7218 is in Figures 7BD).
- user input 7240 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- computer system 101 In response to detecting user input 7240 directed to option 7216-2 to hide the user interface associated with close affordance 7212-2 (Figure 7BG), computer system 101 hides user interface 7208 of application App D, as shown in Figure 7BH, without displaying home menu user interface 7222, as is the case in Figure 7BF, due to the continued presence of user interface 7202 associated with application App A, user interface 7206 associated with application App B, and user interface 7204 associated with application App C in the viewport illustrated in Figure 7BH. While Figure 7BH shows more than one remaining user interface of one or more respective applications in the viewport, in some embodiments, home menu user interface 7222 is not displayed as long as there is at least one remaining user interface of an application in the viewport.
- computer system 101 if no user interface of any application remains in the viewport, but at least one user interface is open in a portion of the three- dimensional environment outside the viewport, computer system 101 forgoes automatically displaying home menu user interface 7222. In some embodiments, however, if no user interface of any application remains in the viewport, computer system 101 automatically displays home menu user interface 7222 without regard to whether (e.g., even if) at least one user interface is open in a portion of the three-dimensional environment outside the viewport.
- Figures 7BI-7BL illustrate example interactions with a close affordance, in response to a second type of input that is different from the first type of input described with reference to user input 7210 of Figure 7BC.
- Figures 7BI and 7BK illustrate that, in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input)directed to close affordance 7212 (e.g., whether displayed as close affordance 7212-1, or as close affordance 7212-2 in response to a gaze input directed to close affordance 7212-1), and in accordance with a determination that the user input is a second type of input, computer system 101 displays an option list 7246-1 corresponding to different options for closing one or more user interfaces of the application associated with close affordance 7212.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as
- Examples of the second type of input include a quick pinch shorter than a pinch threshold time, a pinch gesture that includes a movement velocity above a velocity threshold, a tap sequence (e.g., one or more taps), or a combination input that includes a pinch gesture and a movement input (e.g., an air pinch gesture and movement, and/or another type of air gesture).
- a quick pinch shorter than a pinch threshold time e.g., a pinch gesture that includes a movement velocity above a velocity threshold
- a tap sequence e.g., one or more taps
- a combination input that includes a pinch gesture and a movement input
- the close affordances 7212 of Figures 7BC, 7BI, and 7BK are all analogous.
- close affordance 7212 of Figure 7BC would also, if activated by a user input of the second type instead of the first type, result in display of option list 7246-1 instead of option list 7216-1 in the scenario of Figures 7BD (e.g., if user input 7210 of Figure 7BC were an instance of the second type of input instead of an instance of the first type of input).
- the user interface associated with the close affordance in response to detecting a user input of the second type directed to a close affordance, is closed.
- computer system 101 closes just user interface 7242-1, instead of displaying option list 7264-1 (e.g., and without closing any other displayed user interfaces).
- option list 7246-1 includes: option 7246-2 (e.g., labeled “Hide This Window”) to hide user interface 7242-1 of the application associated with close affordance 7212-2; and option 7246-3 (e.g., labeled “Hide All of This App”) to hide all user interfaces of the application associated with close affordance 7212-2.
- option 7246-2 e.g., labeled “Hide This Window”
- option 7246-3 e.g., labeled “Hide All of This App”
- Figures 7BI-7BJ illustrate an example outcome in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to an option in option list 7246-1.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to an option in option list 7246-1.
- user input 7244 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input
- option 7246-3 to hide all user interfaces of the application associated with close affordance 7212-2.
- computer system 101 hides user interface 7242-1 and user interface 7242-2 of application App D, as shown in Figure 7BJ, thereby hiding all user interfaces of application App D.
- computer system 101 does not display home menu user interface 7222 in conjunction with or after hiding user interfaces 7242-1 and 7242-2, due to the continued presence of user interface 7202 associated with application App A, user interface 7206 associated with application App B, and user interface 7204 associated with application App C in the viewport as illustrated in Figure 7BJ. While Figure 7BJ shows no remaining user interface of application App D in the viewport, in some embodiments, one or more open user interfaces of application App D are permitted to remain in a portion of the three-dimensional environment outside the viewport.
- option 7246-3 in response to detecting user input 7244 selecting option 7246-3 ( Figure 7BI), computer system 101 closes all user interfaces of application App D both within and outside of the viewport, such that no user interface of application App D remains in any portion of the three-dimensional environment. Accordingly, in some embodiments, option 7246-3 is analogous to option 7216-3 described herein with reference to Figures 7BD (e.g., to hide all application user interfaces of the application associated with close affordance 7212).
- Figures 7BK-7BL illustrate another example outcome in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to an option in option list 7246-1, different from the outcome illustrated in Figures 7BI-7BJ.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input e.g., a tap input, a gaze input, a drag input, and/or another type of user input
- Figure 7BK is analogous to Figure 7BI, except that Figure 7BK illustrates user input 7248 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7246-2 to hide the user interface associated with close affordance 7212-2 instead of user input 7244 directed to option 7246-3 to hide all user interfaces of the application associated with close affordance 7212-2, as in Figure 7BI.
- user input 7248 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- option 7246-2 to hide the user interface associated with close afford
- computer system 101 hides user interface 7242-1 of application App D, in accordance with user interface 7242-1 being associated with the close affordance 7212-2 via which option list 7246-1 was invoked, while maintaining display of user interface 7242-2 of application App D, as shown in Figure 7BL.
- option 7246-2 is analogous to option 7216-2 described herein with reference to Figures 7BE and 7BG (e.g., to close the current user interface associated with the activated close affordance 7212).
- other user interfaces of the application App D remain open, within and/or outside of the viewport.
- Figures 7BM1-7CH illustrate examples of displaying a home menu user interface after closing one or more application user interfaces.
- Figure 16 is a flow diagram of an exemplary method 1600 for displaying a home menu user interface after closing one or more application user interfaces.
- the user interfaces in Figures 7BM1-7CH are used to illustrate the processes described below, including the processes in Figure 16.
- Figures 7BM1-7BP illustrate different example scenarios resulting from the closing of a user interface of an application (e.g., in response to a user input directed to a close affordance), depending on whether any other application user interfaces remain in the viewport of the three-dimensional environment after the user interface of the application is closed.
- Figures 7BM illustrates option 7252 that is associated with close affordance 7212 for user interface 7208 of application App D, and that in some embodiments or circumstances is displayed in response to an input directed to close affordance 7212 (optionally a particular type of input as described herein with reference to Figures 7BC-7BL).
- Figures 7BM also illustrates user input 7250 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252, labeled “Close,” for closing a user interface associated with close affordance 7212-2 (e.g., user interface 7208 of application App D).
- user input 7250 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- option 7252 labeled “Close”
- close affordance 7212-2 e.g., user interface 7208 of application App D.
- Figure 7BM2 illustrates a direct air gesture is performed using a gesture performed by the user’s hand 7020 that is based on the position of the representation of the user’s hand 7020’ that is directed to a position, in the user interface, that intersects with the position of option 7252, and/or an indirect air gesture is performed (e.g., as an alternative to the direct air gesture and/or in addition to the direct air gesture) using a gesture performed by the user’s hand 7022 while the user’s gaze input 7250a is directed to option 7252, but is not based on the position of the representation of the user’s hand 7022’.
- Figures 7BM is similar to Figure 7BE, except that, for simplicity of illustration, Figures 7BM includes just the single option 7252.
- option 7252 is in some embodiments analogous to options 7216-2, 7216-3, 7246-2, or 7246-3 described herein with reference to Figures 7BC-7BL.
- Figures 7BN illustrates that, in response to detecting user input 7250 corresponding to selection of option 7252 (e.g., in Figures 7BM), computer system 101 closes user interface 7208 of application App D (and optionally closes all user interfaces associated with application App D).
- computer system 101 automatically displays home menu user interface 7222 without any additional user input.
- the location at which home menu user interface 7222 is displayed is based on the location of user interface 7208 of application App D just prior to user interface 7208 being closed.
- the location of home menu user interface 7222 in Figures 7BN is based on user interface 7208 being closed from the location of user interface 7208 as shown in Figures 7BM. If user interface 7208 were closed while displayed at different location, then home menu user interface 7222 would be displayed at a location that is based on the different location from which user interface 7208 was closed.
- Figure 7BO in contrast to Figures 7BM, one or more additional user interfaces are displayed in the viewport.
- Figure 7BO additionally shows user interface 7202 of application App A displayed in the viewport, in addition to user interface 7208 of application App D.
- Figure 7BO is thus similar to Figure 7BG in that user interfaces of multiple applications are displayed in the viewport.
- Figure 7BO also shows user input 7254 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252.
- user input 7254 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to option 7252.
- computer system 101 closes user interface 7208 of application App D, as shown in Figure 7BP, without displaying home menu user interface 7222 (e.g., as shown in Figures 7BN) due to the continued presence of user interface 7202 associated with application App A in the viewport illustrated in Figure 7BP.
- computer system 101 would also forgo automatically displaying home menu user interface 7222 if a user interface is open in a portion of the three-dimensional environment outside the viewport, even if no user interface of any application remains in the viewport.
- Figures 7BQ-7BR illustrate placement of the home menu user interface in the three-dimensional environment, where the home menu user interface is displayed in accordance with closing a respective user interface, based on respective locations of the respective user interface and one or more other user interfaces (e.g., taking into account the location(s) of user interface(s) closed within a first time threshold from the closing of the respective user interface and not taking into account the location(s) of user interface(s) closed outside of the first time threshold).
- Figure 7BQ is similar to Figures 7BM, except that Figure 7BQ additionally includes frame 7258-1 and frame 7258-2 (e.g., shown with dashed lines).
- Frame 7258-1 indicates the position at which user interface 7202 of application App A was closed
- frame 7258-2 indicates the position at which user interface 7204 of App C was closed, at respective times (e.g., shown in timing diagram 7259) prior to computer system 101 detecting user input 7256 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252 to close user interface 7208 of application App D.
- user input 7256 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an
- Timing diagram 7259 shows an example temporal sequence of the closing of user interface 7202, user interface 7204, and user interface 7208 along a time axis or timeline extending to the right, in which later or subsequent events are placed along the timeline to the right of earlier events.
- Computer system 101 determines whether to position home menu user interface 7222 based on a location of user interface of the application that was just closed, or whether to take into account respective locations of additional user interfaces that were closed within a first time threshold 7260 (e.g., indicated with a left right arrow in timing diagram 7259) from a time an input corresponding to a request to close the last closed user interface was detected or a time the last closed user interface (e.g., user interface 7208 in timing diagram 7259) was closed.
- a first time threshold 7260 e.g., indicated with a left right arrow in timing diagram 7259
- Timing diagram 7259 illustrates that user interface 7204 was closed prior to closing user interface 7208 and within the first time threshold 7260 of the closing of user interface 7208.
- User interface 7202 although also closed prior to closing user interface 7208, was closed at an earlier time that was beyond first time threshold 7260.
- computer system 101 positions home menu user interface 7222 in a lower right portion of the viewport, as shown in Figure 7BR, based on the position of frame 7258-2 in addition to the position of user interface 7208 of application App D, and not based on the position of frame 7258-1.
- the position of home menu user interface 7222 in Figure 7BR is thus lower and shifted to the right compared to the position of home menu user interface 7222 in Figures 7BN (e.g., where home menu user interface 7222 is placed based on the location of user interface 7208 alone).
- Figures 7BS-7BT illustrate placement of the home menu user interface in the three-dimensional environment, where the home menu user interface is displayed in accordance with closing a respective user interface (e.g., in response to a user input directed to a close affordance associated with the respective user interface), based on respective locations of the respective user interface and one or more other user interfaces, all of which were closed within the first time threshold of the closing of the respective user interface.
- Figure 7BS shows three dashed-line frames of three closed user interfaces at their respective positions prior to being closed: frame 7258-5 representing the position of user interface 7205 of App F, frame 7258-4 representing the position of user interface 7203 of App E, and frame 7258-3 representing the position of user interface 7208 of App D.
- frame 7258-5 representing the position of user interface 7205 of App F
- frame 7258-4 representing the position of user interface 7203 of App E
- frame 7258-3 representing the position of user interface 7208 of App D.
- timing diagram 7207 user interface 7208 of application App D is the last closed user interface, and both user interface 7205 and user interface 7203 are closed within first time threshold 7260 of the closing of user interface 7208.
- computer system 101 in conjunction with closing the last user interface (e.g., user interface 7208 of App D) of a cluster of user interfaces closed within the first time threshold 7260 of the closing of user interface 7208 (e.g., including user interface 7208 of App D, user interface 7203 of App E, and user interface 7205 of App F), places home menu user interface 7222 at a position corresponding to a characteristic portion 7263 of the cluster.
- characteristic portion 7263 of the cluster is a centroid of the cluster, a region of maximum overlap between the dashed line frames, or a collective location of the cluster in one or more directions or dimensions (e.g., in a lateral dimension or x-direction or azimuthal direction; in a depth dimension or z-direction or distance from the viewpoint of the user; and/or in a height dimension or y-direction or altitudinal angle from the viewpoint of the user).
- home menu user interface 7222 is automatically placed at position corresponding to characteristic portion 7263 of the cluster.
- home menu user interface 7222 is displayed at a position that coincides with characteristic portion 7263 in one or more directions (e.g., in a lateral dimension or x-direction or azimuthal direction). In some embodiments, home menu user interface 7222 is displayed with an offset from characteristic portion 7263 in one or more directions (e.g., in a depth dimension or z- direction or distance from the viewpoint of the user; and/or in a height dimension or y- direction or altitudinal angle from the viewpoint of the user).
- home menu user interface 7222 in Figure 7BS is laterally centered on characteristic portion 7263 of the cluster (e.g., along a lateral dimension or x- direction, and/or at a respective azimuthal angle with respect to a viewpoint of the user).
- home menu user interface 7222 is displayed offset from the positions of the respective dashed-line frames of the cluster, such as in front of the positions of the cluster (e.g., along the depth dimension, closer to the viewpoint of user, and/or at a lesser radial distance from the viewpoint of the user) (e.g., a respective distance in front of characteristic portion 7263, where the position of characteristic portion 7263 in the z- direction or radial direction is the position of the frame closest to the viewpoint of the user), and/or above characteristic portion 7263 of the cluster (e.g., a respective amount higher along a y-direction, further from representation 7008’ of the floor, such as raised by between 1-10° relative to characteristic portion 7263 as determined from a viewpoint of the user, and/or a respective amount higher in altitudinal angle from the viewpoint of the user).
- a central portion of home menu user interface 7222 is aligned with characteristic portion 7263, or a top portion of the home menu user interface 7222
- Top view 7275 of Figure 7BT illustrates a location of home menu user interface 7222, and locations of user interface 7203, user interface 7205, and user interface 7208 just prior to being closed, as seen from above (e.g., from the top, rather than from the front as in the viewport illustrated in Figure 7BT as visible via computer system 101).
- Coordinate diagram 7225 in the coordinate system of the three-dimensional environment for top view 7275 shows the z-axis (e.g., depth dimension or radial direction) extending upwards and the x-axis extending to the right.
- the depth dimension e.g., z-direction or radial direction
- Top view 7275 of Figure 7BT shows that home menu user interface 7222 is placed in front of the locations of the closed user interfaces 7203, 7205, and 7208.
- Figures 7BU-7BX illustrate different example placements of the home menu user interface, where the home menu user interface is displayed in accordance with closing a respective user interface, based on the amount of time between the closing of the respective user interface and the closing of another user interface (e.g., most recently closed) prior to detecting the user input corresponding to the request to close the respective user interface.
- Figure 7BU is similar to Figure 7BS in that there are two other user interfaces that were closed (e.g., indicated by frame 7258-1 representing closed user interface 7202 of App A and frame 7258-2 representing closed user interface 7204 of App C) within first time threshold 7260 prior to the closing of the last user interface (e.g., user interface 7208).
- frame 7258-1 representing closed user interface 7202 of App A
- frame 7258-2 representing closed user interface 7204 of App C
- Timing diagram 7209 shows that user interface 7202 of App A is closed first, followed by user interface 7204 of App C, and finally user interface 7208 of App D is closed in response to detecting user input 7262 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) (Figure 7BU) directed to option 7252 for closing user interface 7208 associated with close affordance 7212. Both user interface 7202 and user interface 7204 are closed within first time threshold 7260 of the closing of user interface 7208.
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- Timing diagram 7209 also shows a time delay 7213 (e.g., within first time threshold 7260) between the closing of the last user interface 7208 and the closing of the penultimately closed user interface 7204.
- a time delay 7213 e.g., within first time threshold 7260
- the user pauses (e.g., by directing attention to a different user interface, or to a different portion of the viewport, and/or interacting with one or more other user interface objects) for a length of time (e.g., time delay 7213) before closing the last user interface 7208 of App D.
- Figure 7BV illustrates that, in response to detecting user input 7262 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) corresponding to selection of option 7252 associated with close affordance 7212-2 (Figure 7BU), computer system 101 ceases to display user interface 7208, and automatically displays home menu user interface 7222 in three-dimensional environment at the location depicted in Figure 7BV.
- user input 7262 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- time delay 7213 between the closing of user interface 7208 and the penultimately closed user interface is greater than a delay time threshold (e.g., between 25-75% of first time threshold 7260, or defined independently of first time threshold 7260, such as a threshold of Is, 2s, 5s, 8s, 10s, 15s, or other amount of time)
- computer system 101 disregards the positions of closed application user interfaces other than user interface 7208 (e.g., even other application user interfaces closed within first threshold time 7260) in determining where to display home menu user interface 7222.
- Figure 7BV thus illustrates the impact of closing a last closed application user interface more than a delay time threshold after closing other application user interfaces.
- top view 7211 of Figure 7BV illustrates that the respective locations of user interface 7202, user interface 7204, and user interface 7208, prior to being closed, as seen from above, are within spatial threshold 7215 from one another and thus satisfy a spatial proximity requirement for forming a cluster (e.g., in accordance with a determination that a respective application user interface at a respective location is within spatial threshold 7215 from another (e.g., one or more, or all) application user interfaces of the cluster).
- pairwise distances between pairs (e.g., all pairs, or some of the pairs) of the closed user interfaces are determined and user interfaces having pairwise distances above a threshold are excluded from the cluster.
- an average location (e.g., centroid) of a group of user interfaces is determined, and user interfaces that are outside a threshold radius from the average location are excluded from the cluster (e.g., a user interface is excluded from the cluster if a characteristic portion, such as a centroid or respective edge or respective corner, is outside the threshold radius, or if or any portion of the user interface is outside the threshold radius; or the user interface is included in the cluster if the characteristic portion is within the threshold radius or if any portion of the user interface is within the threshold radius).
- a characteristic portion such as a centroid or respective edge or respective corner
- the placement of a home menu user interface such as home menu user interface 7222 is determined based on respective locations of a cluster of closed application user interface.
- the cluster includes application user interfaces that meet proximity criteria (also called herein cluster criteria).
- the proximity criteria include a spatial proximity requirement that respective closed application user interfaces be within a spatial threshold in order to be part of the cluster.
- the proximity criteria include a temporal proximity requirement that respective user interfaces be closed within a threshold time of the last closed user interface in order to be part of the cluster.
- the proximity criteria include the temporal proximity requirement without the spatial proximity requirement, the spatial proximity requirement without the temporal proximity requirement, or both (e.g., with or without one or more other requirements).
- the user interfaces within a cluster change as a function of time, due to closure and/or display of additional user interfaces, and/or movement of user interfaces.
- computer system 101 determines a placement position of home menu user interface 7222 based on respective locations of user interfaces in the cluster without taking into respective locations of user interfaces that are excluded from the cluster.
- Figures 7BW-7BX in contrast to Figures 7BU-7BV, illustrate placement of the home menu user interface in the three-dimensional environment in scenarios in which time intervals between closing successive user interfaces (e.g., and in particular the time interval between closing the penultimately closed user interface and the last closed user interface) do not exceed the delay time threshold.
- time intervals between closing successive user interfaces e.g., and in particular the time interval between closing the penultimately closed user interface and the last closed user interface
- Figure 7BW shows the same viewport as in Figure 7BU (e.g., with user interface 7208, frame 7258-1, and frame 7258-2 at the same respective locations), and user input 7264 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252 associated with close affordance 7212-2.
- user input 7264 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input
- computer system 101 In response to detecting user input 7264, computer system 101 ceases to display user interface 7208, and displays home menu user interface 7222 in three-dimensional environment at the location depicted in Figure 7BX.
- a time interval between the closing of user interface 7208 and the closing of the most recent prior closed user interface 7204, as shown in timing diagram 7217, is less than the delay time threshold.
- the placement of home menu user interface 7222 in Figure 7BX is based on locations of all user interfaces closed within the first time threshold 7260 (e.g., taking into account the respective positions of frames 7258-1 and 7258-2 in addition to the position of user interface 7208, optionally in accordance with a determination that frames 7258-1 and 7258-2 and user interface 7208 satisfy the spatial proximity requirement), in contrast to the scenario illustrated in Figures 7BU and 7BV (e.g., in which home menu user interface 7222 is positioned based on the position of user interface 7208 without taking into account the respective positions of frames 7258-1 and 7258-2).
- the position of home menu user interface 7222 in the viewport as illustrated in Figure 7BX is different from the position of home menu user interface 7222 in the viewport illustrated in Figure 7BV, which is depicted as dashed-line frame 7258-10 in Figure 7BX for comparison.
- dashed-line frames 7258-1 and 7258-2 being offset to the right of and, on average, downward from user interface 7208
- home menu user interface 7222 in Figure 7BX is positioned to the right of and downward from dashed-line frame 7258-10.
- Figures 7BY-7BZ illustrate placement of the home menu user interface in the three-dimensional environment based on whether respective locations of closed user interfaces of respective applications meet proximity criteria that include both a spatial proximity requirement and a temporal proximity requirement.
- Figure 7BY the positions of dashed-line frames 7258-1 and 7258-2 and user interface 7208 are analogous to those depicted in Figure 7BW, and Figure 7BY further includes dashed-line frame 7258-6 indicating the position of user interface 7206 of application App B when closed.
- Figure 7BY also shows user input 7266 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252 associated with close affordance 7212-2, to close associated user interface 7208.
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to option 7252 associated with close affordance 7212-2, to close associated user interface 7208.
- timing diagram 7221 the timing of the respective closings of frames 7258-1 and 7258-2 and user interface 7208 are analogous to those in Figure 7BW, and timing diagram 7221 additionally shows that user interface 7206 is also closed within first time threshold 7260, after the closing of user interface 7202 and before the closing of user interface 7204.
- computer system 101 In response to detecting user input 7266 ( Figure 7BY), computer system 101 ceases to display user interface 7208, and displays home menu user interface 7222 in three-dimensional environment at the location depicted in Figure 7BZ.
- user interface 7206 of App B is located outside of spatial threshold 7215, as depicted in top view 7223, computer system 101 does not take the position of frame 7258-6 into account in determining the location of home menu user interface 7222 shown in Figure 7BZ.
- Figures 7CA-7CB illustrate placement of the home menu user interface in the three-dimensional environment based on whether respective locations of user interfaces of respective applications meet proximity criteria that include attention criteria (e.g., optionally in addition to a spatial proximity requirement and/or a temporal proximity requirement).
- attention criteria e.g., optionally in addition to a spatial proximity requirement and/or a temporal proximity requirement.
- Figure 7CA the positions of dashed-line frames 7258-1, 7258-2, and 7258- 6 and user interface 7208 are analogous to those depicted in Figure 7BY, and Figure 7CA further includes dashed-line frame 7258-4 indicating the position of user interface 7203 of App E when closed.
- Figure 7CA also shows user input 7268 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to option 7252 associated with close affordance 7212-2, to close associated user interface 7208.
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input, and/or another type of user input directed to option 7252 associated with close affordance 7212-2, to close associated user interface 7208.
- timing diagram 7225 the timing of the respective closings of frames 7258-1, 7258-2, and 7258-6 and
- Figure 7CB illustrates that, in response to detecting user input 7268 (Figure 7CA), computer system 101 ceases to display user interface 7208 and displays home menu user interface 7222 in the three-dimensional environment at the location depicted in Figure 7CB.
- Figure 7CB illustrates embodiments in which computer system 101 determines where to display home menu user interface 7222 based on which user interface(s) meet the attention criteria, such as which received the most attention (e.g., based on receiving attention for the longest time among all the user interfaces closed within first time threshold 7260, via a gaze input having a gaze movement that varies below a threshold amount, based on receiving a highest number of user interactions, based on involving a highest amount of cognitive engagement, and/or other factors).
- computer system 101 generates a sorted list of descending amounts of attention, such as gaze time or a number of interactions, for one or more (e.g., all, or a majority) of the application user interfaces.
- the user interface(s) that meet the attention criteria include the first element or top N elements in the sorted list of descending amounts of attention.
- Figure 7CA indicates that the user’s attention 7276 was directed most to user interface 7202 (e.g., represented by dashed-line frame 7258-1 after closing of user interface 7202), of the user interfaces closed within the first time threshold 7260, such that user interface 7202 meets the attention criteria. Because user interface 7202 meets the attention criteria, the location of home menu user interface 7222 depicted in Figure 7CB is determined based on the location of user interface 7208 (e.g., due to being the last remaining user interface in the three-dimensional environment just prior to being closed, in some embodiments even if user interface 7208 does not itself meet the attention criteria) and the location of frame 7258-1 corresponding to user interface 7202 as shown in Figure 7CA.
- the locations of user interfaces 7206 and 7204 are disregarded for not meeting the attention criteria (and user interface 7206 also does not meet the spatial proximity requirement). Accordingly, the placement of home menu user interface 7222 in Figure 7CB is higher than and to the right of the placement of home menu user interface 7222 in Figures 7BN (e.g., Figures 7BN1 and 7BN2), because the position of dashed-line frame 7258-1 is taken into account in the placement of home menu user interface 7222 in Figure 7CB.
- closed user interfaces that would otherwise meet attention criteria are excluded in determining the placement of the home menu user interface if the closed user interfaces do not meet one or more other requirements of the proximity criteria.
- the location of frame 7258-4 corresponding to App E in Figure 7CA is ignored in determining the placement of home menu user interface 7222 in Figure 7CB, because App E was closed outside of the first time threshold 7260 as shown in timing diagram 7225 and/or because user interface 7203 of App E was outside the spatial threshold 7215 as shown in top view 7227 (e.g., even if the user paid the same or more attention to user interface 7203 of application App E as to the user interfaces of the other applications such that user interface 7203 of application App E meets the attention criteria).
- Figures 7CC-7CH illustrate different example scenarios in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) directed to a portion of computer system (e.g., a respective input device or input element) associated with the home menu user interface, based on whether a home menu user interface is within a viewport of the three-dimensional environment at the time the user input is detected.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- Figure 7CC illustrates a viewport of the three-dimensional environment that includes user interface 7208 and a passthrough view that includes representation 7008’ of the floor and a representation 7014’ of physical object 7014.
- the viewport of the three- dimensional environment illustrated in Figure 7CC does not include home menu user interface 7222.
- home menu user interface 7222 is present in a portion of the three-dimensional environment outside the viewport illustrated in Figure 7CC.
- Figure 7CC also illustrates user input 7270 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) detected via button 7273 associated with the home menu user interface.
- display generation component 7100 is provided within a housing of computer system 101, and one or more buttons 7271 and 7273 are provided on the housing that encloses or surrounds display generation component 7100.
- user input 7270 is a respective type of input (e.g., a button press directed to one or more of button 7271 and/or button 7273, a rotational input to button 7271 and/or button 7273, a hand gesture such as an air gesture, a gaze and pinch gesture, a tap gesture, a long press gesture, or other system gesture associated with the home menu user interface).
- a button press directed to one or more of button 7271 and/or button 7273
- a rotational input to button 7271 and/or button 7273 e.g., a hand gesture such as an air gesture, a gaze and pinch gesture, a tap gesture, a long press gesture, or other system gesture associated with the home menu user interface.
- Figure 7CD illustrates that, in response to detecting user input 7270 ( Figure 7CC), computer system 101 displays home menu user interface 7222 in the three-dimensional environment at a characteristic portion relative to the current viewport (e.g., a central portion of the viewport, in a portion of the viewport that is lower, along a y-direction, than a viewpoint of the user, closer to representation 7008’ of the floor, such as lowered by between 1-10° as determined from the viewpoint of the user, and/or a respective amount lower in altitudinal angle from the viewpoint of the user).
- a characteristic portion relative to the current viewport e.g., a central portion of the viewport, in a portion of the viewport that is lower, along a y-direction, than a viewpoint of the user, closer to representation 7008’ of the floor, such as lowered by between 1-10° as determined from the viewpoint of the user, and/or a respective amount lower in altitudinal angle from the viewpoint of the user.
- home menu user interface 7222 is displayed in the scenario of Figure 7CD in accordance with a determination that home menu user interface 7222 was not displayed when user input 7270 was detected (Figure 7CC).
- the home menu user interface 7222 is displayed at the characteristic portion of the viewport in accordance with being displayed in response to an input via button 7273, rather than being displayed based on a characteristic portion of a cluster of one or more application user interfaces in accordance with being displayed automatically in conjunction with the closing of an application user interface.
- Figures 7CE-7CF stand in contrast to Figures 7CC-7CD with respect to the behavior of computer system 101 in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) via button 7273.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the
- Figure 7CE illustrates user input 7272 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) detected via button 7273, which is optionally a same type of input as user input 7270 ( Figure 7CC).
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- button 7273 which is optionally a same type of input as user input 7270 (Figure 7CC).
- the viewport of the three-dimensional environment illustrated in Figure 7CE includes home menu user interface 7222 in addition to user interface 7208 and a passthrough view that includes representation 7008’ of the floor and representation 7014’ of physical object 7014.
- Figure 7CF illustrates that, in response to detecting user input 7272, computer system 101 ceases to display home menu user interface 7222 and in some embodiments also ceases to display user interface 7208 of application App D.
- computer system 101 continues displaying user interface 7208 of application App D in response to detecting user input 7272 until detecting a repetition of user input 7272, in response to which computer system 101 then ceases to display user interface 7208 of application App D such that only the passthrough view is visible.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input
- button 7273 e.g., a tap input, a gaze input, a drag input, and/or another type of user input
- Figures 7CG-7CH illustrate placement of the home menu user interface in the three-dimensional environment in response to detecting a user input (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) (e.g., user input 7274) via button 7273 while multiple application user interfaces and no home menu user interface (or portion thereof) are displayed in the viewport of the three-dimensional environment.
- a user input e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- Figures 7CG-7CH thus differ from Figures 7CC-7CD in that, in Figures 7CC-7CD, a single application user interface is displayed in the viewport when a user input via button 7273 is detected.
- Figures 7CG-7CH also differ from Figures 7BM-7BN and 7BQ-7CB in that, in Figures 7BM-7BN and 7BQ-7CB, the home menu user interface is displayed automatically in conjunction with closing a user interface rather than in response to a user input via button 7273.
- Figure 7CG user interface 7206 of App B, user interface 7208 of App D, and user interface 7204 of App C are displayed within the viewport of the three-dimensional environment, and no home menu user interface is displayed within the viewport. Even though Figure 7CG depicts user interface 7206, user interface 7208, and user interface 7204 as being displayed in the viewport, in some scenarios, one or more of user interface 7206, user interface 7208, and user interface 7204 have been closed (e.g., within first time threshold 7620 or beyond first time threshold 7620 of each other).
- Figure 7CG also illustrates user input 7274 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) (e.g., a user input that is of a same type as user input 7272 ( Figure 7CE) and/or as user input 7270 ( Figure 7CC)) detected via button 7273.
- a direct air gesture such as an air tap or air pinch at a location with which the user is interacting
- an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting
- a tap input, a gaze input, a drag input e.g., a tap input that is of a same type as user input 7272 ( Figure 7CE) and/or as
- computer system 101 detects user input 7274 while attention 7276 (e.g., representing a gaze input that meets attention criteria, and/or a head elevation or other posture of a user that meets attention criteria) of the user is directed to a target location in the three-dimensional environment.
- attention 7276 e.g., representing a gaze input that meets attention criteria, and/or a head elevation or other posture of a user that meets attention criteria
- Figure 7CH illustrates that, in response to detecting user input 7274 ( Figure 7CG), computer system 101 displays home menu user interface 7222 in the three-dimensional environment.
- home menu user interface 7222 in Figure 7CH is displayed in response to user input 7274 via button 7273 (e.g., manually requesting display of home menu user interface 7222 ) rather than automatically in conjunction with closing a user interface
- home menu user interface 7222 in Figure 7CH is displayed at the characteristic portion of the viewport, as described with reference to Figures 7CC-7CD, rather than a characteristic portion (e.g., a centroid or an edge) of a cluster of application user interfaces that satisfy the proximity criteria (e.g., user interfaces 7206, 7208, and 7204, via spatial proximity) as described herein with reference to Figures 7BS-7CB.
- home menu user interface 7222 is displayed at a home location (e.g., in the characteristic portion of the viewport) that is independent of the respective locations of one or more of user interface 7206 of App B, user interface 7208 of App D, and user interface 7204 of App C, and independent of any characteristic portion of the cluster of user interfaces 7206, 7208, and 7204.
- the home location at which home menu user interface 7222 is placed is the target location of attention 7276, or is determined relative to the target location of attention 7276. In some embodiments, the home location is determined relative to the current viewport of the three-dimensional environment when user input 7274 (e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input, a gaze input, a drag input, and/or another type of user input) is detected (e.g., in a substantially central portion of the viewport), as described in reference to Figure 7CD.
- user input 7274 e.g., a direct air gesture such as an air tap or air pinch at a location with which the user is interacting, an indirect air gesture such as an air pinch while attention of the user or gaze of the user is directed toward at a location with which the user is interacting, a tap input
- a centroid of home menu user interface 7222 is within a threshold angular distance (e.g., within 1°, 2°, 3°, 5°, 10°, or other angle) of a head elevation of the user (e.g., within the threshold angular distance of a normal vector originating from a head or eye portion of the user) or of the target location of attention 7276.
- a threshold angular distance e.g., within 1°, 2°, 3°, 5°, 10°, or other angle
- the home location at which home menu user interface 7222 is displayed is at a respective (e.g., default) distance from a viewpoint of the user, such as a distance that is between 0.5 m to 5 m from the viewpoint of the user in a simulated depth dimension of the three-dimensional environment.
- home menu user interface 7222 is displayed closer to the viewpoint of the user, in the simulated depth dimension of the three-dimensional environment, than one or more application user interfaces are to the viewpoint of the user (e.g., home menu user interface 7222 is positioned closer than the closest of the one or more application user interface), as shown in top view 7229 in Figure 7CH.
- FIG 8 is a flow diagram of an exemplary method 800 for conditionally displaying controls for applications, in accordance with some embodiments.
- method 800 is performed at a computer system (e.g., computer system 101 in Figure 1) including a display generation component (e.g., display generation component 120 in Figures 1 A, 3, and 4) (e.g., a heads-up display, a display, a touchscreen, a projector, etc.) and one or more cameras (e.g., a camera (e.g., color sensors, infrared sensors, and other depth-sensing cameras) that points downward at a user’s hand or a camera that points forward from the user’s head).
- a computer system e.g., computer system 101 in Figure 1
- a display generation component e.g., display generation component 120 in Figures 1 A, 3, and 4
- cameras e.g., a camera (e.g., color sensors, infrared sensors, and other depth-sensing cameras) that points
- the method 800 is governed by instructions that are stored in a non-transitory (or transitory) computer-readable storage medium and that are executed by one or more processors of a computer system, such as the one or more processors 202 of computer system 101 (e.g., control 110 in Figure 1 A). Some operations in method 800 are, optionally, combined and/or the order of some operations is, optionally, changed.
- Method 800 describes conditionally displaying controls for an application window in response to detecting a user’s gaze. Automatically displaying a control in response to detecting a user’s gaze directed to an area of the control, without requiring additional user input, enables the user to access a particular control to perform an operation by shifting the user’s gaze, without cluttering the user interface with display of all the available controls.
- the method 800 is performed at a computer system that is in communication with a first display generation component (e.g., a first display generation component of a two-sided display generation component, a heads-up display, a head-mounted display (HMD), a display, a touchscreen, a projector, a standalone display, and/or a display that is enclosed in the same housing as another display generation component of the same type or a different type) and one or more input devices (e.g., cameras, controllers, touch-sensitive surfaces, joysticks, buttons, gloves, watches, motion sensors, and/or orientation sensors).
- the first display generation component is a display component facing the user and provides an XR experience to the user.
- the first display generation component includes two or more display components (e.g., one set for each eye) that display slightly different images to form a stereoscopic view of the three-dimensional environment.
- the first display generation component and a second display generation component form a two-sided display device (e.g., a two-sided HMD) that displays a first user interface on a first side corresponding to the first display generation component, and a second user interface on the second side corresponding to the second display generation component.
- the second display generation component is a display component facing away from the user and toward an external environment of the user and optionally provides status information related to the first display generation component (e.g., displayed content and/or operational state) and/or the user (e.g., movement of the user’s eyes, and/or attention state of the user) to other users in the external environment.
- the computer system is an integrated device with one or more processors and memory enclosed in the same housing as the first and the second display generation components and at least some of the one or more input devices.
- the computer system includes a computing component (e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer) that includes one or more processors and memory that is separate from the display generation component(s) and/or the one or more input devices.
- a computing component e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer
- the display generation component(s) and the one or more input devices are integrated and enclosed in the same housing.
- the computer system displays (802), via the first display generation component, a first object (e.g., a user-interface object that is associated with an application and/or experience (e.g., an application window, a media display area, a viewport into a three- dimensional environment, and/or a user interface control region), an application view, and/or a virtual object) in a first view of a three-dimensional environment, wherein the first object includes at least a first portion of the first object (e.g., corners of the first object, and/or edges of the first object) and a second portion of the first object (e.g., an interior portion of the first object, and/or portions other than the first portion of the first object).
- a first object e.g., a user-interface object that is associated with an application and/or experience (e.g., an application window, a media display area, a viewport into a three- dimensional environment, and/or a user interface control region), an application view, and/or
- the computer system detects (804), via the one or more input devices, a first gaze input that meets first criteria, wherein the first criteria require that the first gaze input is directed to the first portion of the first object (e.g., a preselected point and/or segment along the boundary (e.g., comers and/or edges) of the first object) in order for the first criteria to be met.
- the first object is an application window or a viewport into a three-dimensional environment.
- the first portion of the first object is a corner or an edge of the first object (e.g., a respective one of the corners and/or edges of a rectangular object).
- the first portion of the first object is one or more preselected points on a periphery of the first object (e.g., a respective one of the top, left, right, bottom points along the circumference of a circular or oval object).
- the first criteria require that the first gaze input being maintained on the first portion of the first object for at least a threshold amount of time in order for the first criteria to be met.
- the first criteria are not met if the first gaze input is detected outside of the first portion of the first object, (e.g., in the second portion of the first object). For example, as described with reference to Figures 7D, in response to detecting the user’s attention 710-2 (e.g., a gaze input that meets the first criteria), the computer system displays resize affordance 708-1.
- the computer system In response to detecting that the first gaze input meets the first criteria, the computer system displays (806) a first control element (e.g., an affordance or control object) (e.g., grabber 706-1, resize affordance 708-1, and/or close affordance 7030) that corresponds to a first operation associated with the first object, wherein the first control element was not displayed prior to detecting that the first gaze input met the first criteria (e.g., the first control element is a newly displayed user interface object or an initially displayed user interface object in response to the detection of the first gaze input meeting the first criteria).
- a first control element e.g., an affordance or control object
- the first control element is a newly displayed user interface object or an initially displayed user interface object in response to the detection of the first gaze input meeting the first criteria.
- the first control element is displayed at or next to the first portion of the first object (e.g., a grabber icon appears on or near the respective corner of the first object). In some embodiments, the first control element is selectively presented for display based at least in part on the respective portion of the first object on which the first gaze input has been detected. In some embodiments, the first criteria do not require another user input, such as a gesture input (e.g., an air gesture input) or an input provided via another input device, to be detected in conjunction with the first gaze input, in order for the first criteria to be met. In other words, in some embodiments, the first gaze input alone triggers the display of the first control element, without the accompaniment of other types of inputs from the user. For example, as described with reference to Figures 7D, in response to detecting the user’s attention 710-2 (e.g., a gaze input that meets the first criteria), the computer system displays resize affordance 708-1.
- attention 710-2 e.g., a gaze input that meets the first criteria
- the computer system While displaying the first control element, the computer system detects (808), via the one or more input devices, a first user input (e.g., a gaze and pinch gesture or another selection gesture or input) directed to the first control element.
- a first user input e.g., a gaze and pinch gesture or another selection gesture or input
- Figures 7K e.g., Figures 7K1, 7K2 and 7K3
- a user input is detected via the user’s hand 7020 that is directed to resize affordance 708-5.
- the computer system performs (810) the first operation with respect to the first object (e.g., increases or decreases the size of the first object, increases the length of a respective edge of the first object without changing the length of other edges of the first object, moves (e.g., translates and/or rotates) the first object as a whole, and/or closes the first object).
- the first operation that is performed on the first object is chosen based at least in part on the type of the user input, the characteristic values of the user input, the type of the first user interface object, the location or type of the first portion of the first object, and/or the type of the first object.
- application window 702 is resized in accordance with the movement of the user’s hand 7020.
- virtual object 7028 is resized to a smaller size.
- application window 702 is moved in position within the three-dimensional environment.
- displaying the first object in the first view of the three- dimensional environment includes displaying an application window (e.g., application window 702) of a first application (e.g., a control panel application, a launchpad or home application, a communication application, an application providing a shared experience for multiple users, or another system or user application executable in a three-dimensional environment) in the first view of the three-dimensional environment (e.g., an AR or VR environment).
- the first object is a virtual application window, such as a two- dimensional or pseudo-three-dimensional window that is displayed at a first position in the three-dimensional environment.
- the computer system while displaying the virtual application window in the first view of the three-dimensional environment, the computer system detects, via the one or more input devices, a second gaze input that meets second criteria, wherein the second criteria require that the second gaze input is directed to a first portion of the virtual application window (e.g., one or more corners of the virtual application window or an edge of the virtual application window) (e.g., the first portion of the virtual application window is an instance of one or more example instances of the first portion of the first object) in order for the second criteria to be met; and in response to detecting that the second gaze input meets the second criteria, the computer system displays a second control element (e.g., the second control element is an instance of one or more example instances of the first control element) that corresponds to a second operation associated with the virtual application window (e.g., the second operation is an instance of one or more example instances of the first operation), wherein the second control element was not displayed prior to detecting that the second gaze input met the second criteria.
- a second control element e.
- the second control element is an object resize control (e.g., resize control 708-1, Figures 7F) that is displayed at or near a respective corner of the virtual application window.
- the second control element is an object move control (e.g., grabber bar (e.g., grabber 706-1) or other affordance) for repositioning the virtual application window within the three-dimensional environment.
- the second control element is an object close control (e.g., close affordance 7030, Figure 7D3) for closing the virtual application window, wherein, in response to detecting selection of the object close control, the computer system ceases to display the virtual application window in the three- dimensional environment.
- Dynamically displaying a respective control in response to a user gazing at a corresponding respective portion of an application window in an AR/VR environment enables the user to access additional controls for the application window without displaying additional controls that obscure the user’s view of the AR/VR environment.
- the computer system while displaying the first object, the computer system detects, via the one or more input devices, a third gaze input, directed to the second portion of the first object (e.g., an interior portion of the first object that is not near a corner and/or an edge included in the first portion of the first object) that is distinct from the first portion of the first object, that does not meet the first criteria.
- the third gaze input does not meet the first criteria because the third gaze input is not directed to the first portion of the first object.
- the first and second portions of the first object respectively correspond to distinct control elements that are associated with distinct operations that can be performed with respect to the first object.
- the second portion of the first object does not correspond to a control element (e.g., the computer system does not display a control element for performing an operation associated with the first object).
- a gaze input directed to an interior portion of the first object does not cause the computer system to display a resize control, a move control, a close control.
- the computer system 101 forgoes displaying resize affordance 708-1 and close affordance 7030, and optionally forgoes displaying grabber 706-1.
- the computer system in response to detecting that the third gaze input does not meet the first criteria, forgoes display of the first control element that corresponds to the first operation associated with the first object. For example, in some embodiments, in accordance with a determination that the third gaze input is not directed to the first portion of the first object and/or does not meet the first criteria, the first control element (e.g., resize affordance, close affordance, and/or move affordance) is not displayed.
- the first control element e.g., resize affordance, close affordance, and/or move affordance
- the device in response to detecting that the third gaze input satisfies third criteria, wherein the third criteria require that the third gaze input is directed to the second portion of the first object in order for the third criteria to be met, displays a third control element that corresponds to a third operation associated with the first object.
- the first portion is a lower corner of the first object, and in response to a gaze input on the lower comer, the computer system displays a resize affordance (e.g., the first control element); the second portion of and the second portion is an upper left comer of the first object, and in response to a gaze input on the upper left comer, the computer system displays a close affordance (e.g., a third control element), without displaying the resize affordance.
- a resize affordance e.g., the first control element
- the second portion of and the second portion is an upper left comer of the first object, and in response to a gaze input on the upper left comer, the computer system displays a close affordance (e.g., a third control element), without displaying the resize affordance.
- resize affordance 708-3 ceases to be displayed (e.g., and grabber 706-1 is optionally displayed in Figure 7J).
- Forgoing display of a respective control in response to the user’s gaze shifting away from the respective control provides feedback to the user that the user’s gaze is no longer detected at the position corresponding to the respective control, thereby providing improved visual feedback about the state of the device.
- the first control element is a first resize affordance (e.g., also referred to as an “object resize control”) (e.g., the first portion is a comer of the first object and the resize affordance is displayed at a position corresponding to the first portion of the first object, at or near the respective comer at which the first gaze input is detected).
- object resize control e.g., the first portion is a comer of the first object and the resize affordance is displayed at a position corresponding to the first portion of the first object, at or near the respective comer at which the first gaze input is detected.
- the computer system while displaying the first resize affordance (e.g., at or near the first portion of the first object, and, optionally, while the gaze input is directed to the first resize affordance), the computer system detects, via the one or more input devices, a second user input directed to the first resize affordance (e.g., a pinch and drag gesture, or a tap and hold input followed by movement of the input object) and in response to detecting the second user input directed to the first resize affordance, resizes the first object (e.g., in accordance with the detected second user input, such as a direction and/or amount of movement of the second user input (e.g., the direction and movement of the pinch and drag gesture, or another user input that includes an amount and/or direction of movement of the user input)).
- a second user input directed to the first resize affordance e.g., a pinch and drag gesture, or a tap and hold input followed by movement of the input object
- the computer system detects, via
- application window 702 is resized in accordance with a user input directed to resize affordance 708-5.
- virtual object 7028 is resized in response to a user input directed to resize affordance 714-1.
- Dynamically displaying a resize control for changing a size of a virtual object in response to detecting the user gazing at a corner of the object enables the user to access the resize control without displaying additional controls and reduces the number of inputs required to navigate complex menu hierarchies.
- detecting the second user input directed to the first resize affordance includes detecting a direction of movement of the second user input directed to the first resize affordance.
- resizing the first object includes: in accordance with a determination that the direction of movement of the second user input is a first direction, increasing a size of the first object; and in accordance with a determination that the direction of movement of the second user input is a second direction different from the first direction, decreasing the size of the first object.
- the first object is resized in accordance with the second user input.
- the first direction is a direction that moves away from the first object (e.g., dragging a comer of the first object outward) to increase the size of the first object.
- the second direction is a direction that moves toward the interior of the first object (e.g., dragging a corner of the first object inward) to decrease the size of the first object. For example, as described with reference to Figures 7K-7L, an amount and/or a direction of movement of the user’s hand 7020 while resize affordance 708-5 is selected causes the application window 702 to increase and/or decrease in size.
- Figures 7K For example, moving the user’s hand 7020 up and to the left in Figures 7K (e.g., Figures 7K1, 7K2 and 7K3) causes the application window 702 to decrease in size, while moving the user’s hand 7020 down and to the right causes the application window 702 to increase in size. Allowing a user to change a size of a virtual object by selecting and moving a resize control in a respective direction, wherein the resize control is automatically displayed while the user is gazing at a portion of the virtual object that corresponds to the resize control, reduces the number of inputs needed to resize an object without displaying additional controls.
- detecting the second user input directed to the first resize affordance includes detecting an amount of movement of the second user input (e.g., movement in a first direction, movement in a second direction, and/or net movement) directed to the first resize affordance.
- resizing the first object in response to detecting the second user input directed to the first resize affordance, includes: in accordance with a determination that the amount of movement of the second user input is a first amount of movement, changing a size of the first object to a first size that is selected based on the first amount of movement of the second user input (e.g., increasing by a first amount in a first direction, and/or decreasing by a first amount in a second direction); and in accordance with a determination that the amount of movement of the second user input is a second amount of movement different from the first amount of movement, changing the size of the first object to a second size different from the first size, the second size selected based on the second amount of movement of the second user input (e.g., increasing by a second amount in a first direction, and/or decreasing by a second amount in a second direction).
- the size of the first object is changed by an amount that is proportional to or otherwise a function of the amount of movement of the user input directed to the first resize affordance.
- the size of the first object is changed gradually as the amount of movement of the user input changes. For example, as the user increases an amount of movement away from the initial position of the user input, the size is changed (e.g., increased and/or decreased) by a corresponding amount, and as the user decreases an amount of movement away from the initial position of the user input, the size is changed by a corresponding amount.
- an amount of movement of the user’s hand 7020 causes the application window 702 to increase and/or decrease by a corresponding (e.g., proportional amount). For example, a smaller amount of movement of the user’s hand 7020 relative to its initial position causes a smaller change in size of the application window 702. Allowing a user to change a size of a virtual object by selecting and moving a resize control by a respective amount, wherein the resize control is automatically displayed while the user is gazing at a portion of the virtual object that corresponds to the resize control, reduces the number of inputs needed to resize an object without displaying additional controls.
- detecting the first user input directed to the first control element includes detecting a first air gesture (e.g., a gaze input in combination with a selection input, such as a pinch input or a tap input and/or a drag input; or an air gesture, as described above) directed to the first control element.
- the first object is resized in response to detecting the first air gesture directed to the first control element (and, optionally, in accordance with the direction and/or amount of movement of the first air gesture). For example, in response to detecting the air gesture directed to a resize affordance for an application window, the computer system changes a size of the application window in accordance with a direction and/or an amount of movement of the air gesture.
- the computer system in response to detecting the air gesture directed to a move affordance for the application window, moves a position of the application window within the three-dimensional environment in accordance with a direction and/or an amount of movement of the air gesture.
- the user input detected via the user’s hand 7020 is an air gesture for resizing the application window 702. Allowing a user to perform an operation for a virtual object, such as moving or resizing the virtual object, by performing an air gesture without requiring the user to connect, set up, or use a separate input device to perform the operation, reduces the number of inputs needed.
- the first portion of the first object includes a first corner of the first object, and the first criteria require that the first gaze input is directed to the first corner of the first object in order for the first criteria to be met.
- the first corner is a bottom corner, including a bottom left corner and/or a bottom right comer.
- the first portion of the first object includes two or more comers of the first object (e.g., a bottom left and a bottom right corner), but does not include a space or an edge that is between the two or more comers).
- the first portion of the first object includes a predefined area around the comer of the first object.
- the gaze input is detected as being directed to the predefined area around the corner of the first object (e.g., including an end portion of the edge connected to the first corner, a region inside the first object near the first corner, and/or a region outside the first object near the first corner).
- the user’s attention 710-2 is directed to a bottom right comer of application window 702. Automatically displaying a respective control, such as a resize control, at a comer of a virtual object in response to detecting a user’s gaze at the comer reduces the number of inputs needed to perform an operation using the control for the virtual object without displaying additional controls.
- the first portion of the first object includes a first subportion of the first object (e.g., a first comer of the first object, a left corner, and/or a lower left corner) and a second sub-portion of the first object (e.g., a second corner of the first object, the right corner, and/or the lower right comer), wherein the first sub-portion of the first object and the second sub-portion of the first object are separated by a third sub-portion of the first object (e.g., the central portion of the bottom edge, the central portion of the edge connecting the first corner and the second corner of the first object) that is not included in the first portion of the first object, and the first criteria require that the first gaze input is directed to at least one of the first sub-portion and the second sub-portion of the first object in order for the first criteria to be met.
- a first subportion of the first object e.g., a first comer of the first object, a left corner, and/or a lower left corner
- the computer system displays the first control element (e.g., a resize affordance) at or near at the respective sub-portion that is the target of the first gaze input.
- the first control element is displayed with different visual properties depending on whether the user is gazing at the first sub-portion of the first object or the second sub-portion of the first object.
- the first control element is displayed at a location proximate to the respective corner to which the gaze input is detected (e.g., if the user is gazing at the bottom left comer, the resize affordance is displayed proximate to the bottom left corner of the first object with an “L” shape; and if the user is gazing at the bottom right corner, the resize affordance is displayed proximate to the bottom right comer of the first object with a reversed “L” shape).
- the computer system displays a corresponding resize affordance at the respective corner (e.g., the bottom right and/or bottom left comer).
- Automatically displaying a respective control, such as a resize control, at one of multiple possible comers of a virtual object in response to detecting a user’s gaze at the respective corner reduces the number of inputs needed to perform an operation using the control for the virtual object without displaying additional controls, including only displaying a control at the corner at which the user is gazing without displaying a control in the other corners at which the user is not gazing.
- detecting the first user input while displaying the first control element includes detecting the first user input directed to the first sub-portion of the first object or the second sub-portion of the first object.
- performing the first operation with respect to the first object in response to detecting the first user input directed to the first control element includes: in accordance with a determination that the first user input is directed to the first sub-portion of the first object, changing a size of the first object to a first size (e.g., a larger size than the previous size of the first object) while maintaining a position of a center of the first object; and in accordance with a determination that the first user input is directed to the second sub-portion of the first object, changing the size of the first object to a second size (e.g., a smaller size than the previous size of the first object, and/or a size that is different from the first size) while maintaining the position of the center of the first object in the three-dimensional environment.
- a first size e.g., a larger size than the previous size of the first
- resizing application window 702 includes maintaining a center position of application window 702, for example by moving the edges and comers of application window 702 inwards to decrease the size of the application window 702.
- Changing a size of a virtual object while maintaining a center of the virtual object’s position within an AR/VR environment provides improved visual feedback about the state of the device and reduces the number of inputs required to move the virtual object after resizing by maintaining the center position of the virtual object during and after resizing.
- the first portion of the first object corresponds to a first edge of the first object (e.g., the edges that are connected to the respective corner of the first control element) and does not correspond to a second edge of the first object (e.g., the opposite edges that are not connected to the respective corner of the first control element).
- performing the first operation with respect to the first object in response to detecting the first user input directed to the first control element includes: changing a size of the first object by moving the first edge of the first object while maintaining a position of the second edge of the first object in the three-dimensional environment (e.g., moving the two edges connected to the comer of the first control element, while keeping the opposite edges stationary, while resizing the first object).
- resizing application window 702 includes maintaining the top left comer at its position (e.g., and the top edge and left edge at their respective positions) while moving the bottom right corner inward to decrease a size of application window 702.
- Changing a size of a virtual object while maintaining an edge and/or a corner of the virtual object at a same position within an AR/VR environment provides improved visual feedback about the state of the device and reduces the number of inputs required to move the virtual object after resizing the object by maintaining a position of a corner of the virtual object during and after resizing.
- the computer system while displaying the first object, detects, via the one or more input devices, a fourth gaze input, directed to a respective portion of the first object.
- the computer system in response to detecting the fourth gaze input, the computer system: in accordance with a determination that the fourth gaze input is directed to the first portion of the first object and meets the first criteria with respect to the first portion of the first object, displays the first control element; and in accordance with a determination that the fourth gaze input is directed to a third portion of the first object (a second comer, a third corner, and/or a fourth comer of the first object, and/or central portion of a first edge, a second edge, a third edge, and/or a fourth edge of the first object), different from the first portion of the first object (and, optionally, different from the second portion of the first object), displays a second control element that corresponds to a second operation associated with the first object, wherein the second control element was not displayed prior to detecting that the fourth gaze input
- the computer system while displaying the second control element, the computer system detects a third user input directed to the second control element and in response to detecting the third user input directed to the second control element, performs the second operation with respect to the first object.
- the second operation is the same as the first operation, such as a resize operation.
- the first control element and the second control element are a same type of user interface objects that are displayed at different positions relative to the first object, optionally with different appearances.
- the first object comprises an application window
- the first control element is a resize affordance displayed at a nonzero distance away from a bottom right corner of the application window, displayed with a curvature corresponding to the bottom right corner
- the second control element is a resize affordance displayed at a nonzero distance away from a bottom left comer of the application window, displayed with a curvature corresponding to the bottom left corner.
- the first control element and the second control element correspond to distinct operations with respect to the first object.
- the first control element that corresponds to the first operation includes a resize affordance for resizing the first object and the second control element that corresponds to the second operation includes a close affordance for closing the first object (e.g., the first portion is a bottom corner and the third portion is a top left corner of the first object).
- the respective operation associated with a respective control element is based at least in part on the respective portion of the first object on which the gaze input is detected.
- the computer system displays a close affordance (e.g., also referred to as an “object close control”) for closing, removing, or ceasing display of the first object; and in response to detecting that the gaze input is directed to a bottom left comer of the first object, the computer system displays a resize affordance (e.g., also referred to as an “object resize control”) for changing a size of the first object.
- a close affordance e.g., also referred to as an “object close control”
- a resize affordance e.g., also referred to as an “object resize control”
- the user is enabled to resize application window 702 by gazing at the bottom right corner and/or at the bottom left corner, wherein a respective resize affordance 708-1 is displayed at the respective comer at which the user gazes.
- the user is enabled to gaze at another portion of application window 702, such as the top left corner, to cause the computer system to display close affordance 7030.
- Detecting whether the user’s gaze is directed to a first corner or a second corner and automatically displaying a control option, such as a resize control, for the respective corner that the user’s gaze is detected without displaying a control option at the other comer enables the user to perform an operation on the virtual object, such as resizing the virtual object, without displaying additional controls and provides improved visual feedback about the state of the device to indicate where the user’s gaze is detected.
- the first control element is a close affordance (e.g., also referred to as an “object close control”).
- close affordance 7030 is displayed in response to detecting the user’s attention 710-1 at a top left corner of application window 702.
- the computer system while displaying the close affordance, detects, via the one or more input devices, a fourth user input directed to the close affordance; and in response to detecting the fourth user input directed to the close affordance, closes the first object, including ceasing display of the first object in the first view of the three-dimensional environment, as described with reference to Figures 7R-7S for virtual object 7028.
- the first portion of the first object is a top left corner of the first object, and in response to detecting a gaze input directed to the top left corner of the first object, the computer system displays the close affordance proximate to the top left comer.
- the close affordance is displayed at a non-zero distance away from the first object (e.g., the close affordance is displayed as a separate user interface object than the first object).
- performing the first operation with respect to the first object in response to detecting the first user input includes closing the first object, including ceasing display of the first object in the first view of the three-dimensional environment, along with ceasing display of the first control element (e.g., the closing affordance).
- other virtual objects continue to be displayed in the three- dimensional environment.
- a portion of the three-dimensional environment that was obscured by display of the first object before closing the first object is displayed after the first object is closed. For example, ceasing display of the first object enables a portion of the three-dimensional environment (e.g., including any virtual objects and/or passthrough representations) that was behind the first object to be visible.
- Dynamically displaying a close control for closing a virtual object such that it is no longer displayed in the three-dimensional environment in response to detecting the user gazing at a corner of the object, enables the user to access the close control and close the virtual object without displaying additional controls, and automatically displays the close control in response to the user gazing at a comer associated with the close control, thereby reducing the number of inputs required to access the close control.
- the computer system while displaying the first object and a second object in the first view of the three-dimensional environment (e.g., the first object and the second object are both application windows, both three-dimensional objects, or of the same object type), the computer system detects that a fifth gaze input is directed to the first object.
- the first object and the second object are both application windows, both three-dimensional objects, or of the same object type
- the computer system in response to detecting that the fifth gaze input directed to the first object: in accordance with a determination that the fifth gaze input is directed to a first subportion of the first object (e.g., a top left corner of the first object) and meets the first criteria with respect to the first sub-portion of the first object, the computer system displays a first close affordance for the first object (e.g., at or proximate to the first sub-portion of the first object) and in accordance with a determination that the fifth gaze input is directed to a second sub-portion of the first object (e.g., a lower left corner or lower right comer of the first object) and meets the first criteria with respect to the second sub-portion of the first object, the computer system displays a first resize affordance for the first object (e.g., at or proximate to the second sub-portion of the first object).
- a first subportion of the first object e.g., a top left corner of the first object
- the computer system displays a first close
- the computer system while displaying the first object and the second object in the first view of the three-dimensional environment (e.g., the first object and the second object are both application windows, both three-dimensional objects, or of the same object type), the computer system detects that a sixth gaze input is directed to the second object.
- the computer system in response to detecting that the sixth gaze input directed to the second object: in accordance with a determination that the sixth gaze input is directed to a first sub-portion of the second object (e.g., a top left corner of the second object, and/or a portion of the second object that corresponds to the first sub-portion of the first object) and meets the first criteria with respect to the first sub-portion of the second object, the computer system displays a second close affordance for the second object (e.g., at or proximate to the first sub-portion of the second object) and in accordance with a determination that the sixth gaze input is directed to a second sub-portion of the second object (e.g., a lower left corner or lower right comer of the second object, and/or a portion of the second object that corresponds to the second sub-portion of the first object) and meets the first criteria with respect to the second sub-portion of the second object, the computer system forgoes display of a second resize affordance for the second object (e.g.
- some application windows and/or virtual objects cannot be resized, and a resize affordance is not displayed in response to the user gazing at a comer of the application window and/or virtual object.
- resize affordance 708-1 is not displayed if application window 702 cannot be resized.
- Displaying a close control for closing a virtual object in response to detecting the user gazing at a corner of the object associated with the close control, without displaying a resize control if the object cannot be resized, enables the user to access the close control and close the virtual object without displaying additional controls, and automatically displays the close control in response to the user gazing at a corner associated with the close control, thereby reducing the number of inputs required to close an object without requiring the user to navigate complex menu hierarchies.
- the computer system while displaying the first object and a third object in the first view of the three-dimensional environment (e.g., the first object and the third object are both application windows, both three-dimensional objects, or of the same object type), the computer system detects that a seventh gaze input directed to the first object.
- the first object and the third object are both application windows, both three-dimensional objects, or of the same object type
- the computer system in response to detecting that the seventh gaze input directed to the first object: in accordance with a determination that the seventh gaze input is directed to a first sub- portion of the first object (e.g., a top left corner of the first object) and meets the first criteria with respect to the first sub-portion of the first object, the computer system displays a first close affordance for the first object (e.g., at or proximate to the first sub-portion of the first object); and in accordance with a determination that the seventh gaze input is directed to a third sub-portion of the first object (e.g., a center portion of the bottom edge) and meets the first criteria with respect to the third sub-portion of the first object, the computer system displays a first move affordance (e.g., also referred to as an “object move control”) for the first object (e.g., at or proximate to the third sub-portion of the first object).
- a first move affordance e.g., also referred to as an “object move control” for
- the computer system while displaying the first object and the third object in the first view of the three-dimensional environment (e.g., the first object and the third object are both application windows, both three-dimensional objects, or of the same object type), the computer system detects that an eighth gaze input directed to the third object.
- the computer system in response to detecting that the eighth gaze input directed to the third object: in accordance with a determination that the eighth gaze input is directed to a first sub-portion of the third object (e.g., a top left corner of the second object, and/or a portion of the second object that corresponds to the first sub-portion of the first object) and meets the first criteria with respect to the first sub-portion of the third object, the computer system displays a third close affordance for the third object (e.g., at or proximate to the first sub-portion of the third object); and in accordance with a determination that the eighth gaze input is directed to a second sub-portion of the third object (e.g., a center portion of the bottom edge of the third object, and/or a portion of the third object that corresponds to the third sub-portion of the first object) and meets the first criteria with respect to the second sub-portion of the third object, the computer system forgoes display of a second move affordance for the third object (e.g., in accord
- one or more application windows and/or virtual objects cannot be repositioned within the three- dimensional object, and a grabber is not displayed for the one or more application windows and/or virtual objects.
- grabber 706-1 is optionally not displayed if application window 702 cannot be moved in position within the three-dimensional environment.
- one or more other controls for application window 702 are displayed in response to detecting the user’s attention at respective portions of the application window 702 (e.g., a resize affordance 708-1 is displayed, and/or close affordance 7030 is displayed) without displaying grabber 706-1 for moving the application window.
- Displaying a close control for closing a virtual object in response to detecting the user gazing at a corner of the object associated with the close control, without displaying a move grabber for changing a position of the object, enables the user to access the close control and close the virtual object without displaying additional controls, and automatically displays the close control in response to the user gazing at a corner associated with the close control, thereby reducing the number of inputs required to close an object without requiring the user to navigate complex menu hierarchies.
- the computer system while displaying the first object, displays a first move affordance for repositioning a position of the first object in the first view of the three-dimensional environment, detecting a ninth gaze input directed to a respective portion of the first object that corresponds to a first resize affordance (e.g., the ninth gaze input moved from the location that corresponds to the first move affordance to the location that corresponds to the first resize affordance; or the ninth gaze input is unrelated to the display of the first move affordance).
- the first control element is the first move affordance.
- the second control element is the first move affordance.
- the first move affordance is optionally displayed for the first object in response to detecting a gaze input for the user that is directed to a center region of the first object (e.g., the first portion).
- the computer system while detecting the user is gazing at any portion of the first object, the computer system automatically displays the move affordance for the first object and, optionally maintains display of the move affordance until detecting a gaze input is detected on another respective portion of the first object (e.g., a second portion, such as a corner, of the first object).
- the computer system in response to detecting the ninth gaze input directed to the respective portion of the first object that corresponds to the first resize affordance (e.g., in response to detecting the ninth gaze input including detecting movement of a gaze input from a center of the first object, or from a center of the bottom edge (e.g., the first portion) of the first object, to a corner of the first object (e.g., the respective portion that corresponds to the first resize affordance)), the computer system ceases display of the first move affordance; and displays the first resize affordance for resizing the first object at or proximate to the respective portion of the first object that corresponds to the first resize affordance.
- displaying the first resize affordance includes animating a transition between ceasing display of the first move affordance (e.g., by shifting or animating movement of the first move affordance in a direction toward the first resize affordance and/or gradually decreasing a size of the first move affordance until it disappears) and displaying the first resize affordance (e.g., by shifting or animating movement of the first resize affordance as coming from the first move affordance and/or gradually increasing a size of the first resize affordance).
- the computer system replaces display of grabber 706-1 (e.g., a move affordance) with display of resize affordance 708-1.
- Replacing display of a move grabber for changing a position of a virtual object with display of a resize control in response to detecting the user gazing at a corner of the object associated with the resize control enables the user to access the resize control and resize the virtual object without displaying additional controls, thereby reducing the number of inputs required to access the resize control, and providing improved visual feedback that the user is gazing at the corner of the object associated with the resize control.
- the computer system while displaying the first object, displays a second move affordance (e.g., same as the first move affordance, or different from the first move affordance) for repositioning a position of the first object in the first view of the three-dimensional environment, the computer system detects a tenth gaze input directed to a respective portion of the first object that corresponds to a second resize affordance (e.g., the same as the first resize affordance, or different from the first resize affordance) (e.g., the tenth gaze input moved from the location that corresponds to the second move affordance to the location that corresponds to the second resize affordance; or the tenth gaze input is unrelated to the display of the second move affordance).
- a second move affordance e.g., same as the first move affordance, or different from the first move affordance
- the first control element is the second move affordance. In some embodiments, the second control element is the second move affordance. In some embodiments, the second move affordance is optionally displayed for the first object in response to detecting a gaze input for the user that is directed to a center region of the first object (e.g., the first portion). In some embodiments, while detecting the user is gazing at any portion of the first object, the computer system automatically displays the second move affordance for the first object and, optionally maintains display of the move affordance until detecting a gaze input is detected on another respective portion of the first object (e.g., a second portion, such as a corner, of the first object).
- a gaze input is detected on another respective portion of the first object (e.g., a second portion, such as a corner, of the first object).
- the computer system in response to detecting the tenth gaze input directed to the respective portion of the first object that corresponds to the second resize affordance (e.g., in response to detecting the tenth gaze input including detecting movement of a gaze input from a center of the first object, or from a center of the bottom edge (e.g., the first portion) of the first object, to a comer of the first object (e.g., the respective portion that corresponds to the first resize affordance)), the computer system displays an animated transition between displaying the second move affordance and displaying the second resize affordance, including: moving the second move affordance toward a location corresponding to the second resize affordance (e.g., at or proximate to a respective corner of the first object) (e.g., and ceasing display of the second move affordance before the second move affordance reaches the location corresponding to the second resize affordance, and/or before the second resize affordance is displayed at the location corresponding to the second
- displaying the second resize affordance includes animating a transition between ceasing display of the second move affordance (e.g., by shifting or animating movement of the move affordance in a direction toward the location corresponding to the second resize affordance and/or gradually decreasing a size of the second move affordance until it disappears) and displaying the second resize affordance (e.g., by shifting or animating movement of the resize affordance as coming from the second move affordance and/or gradually increasing a size of the second resize affordance until it is full displayed at the location corresponding to the second resize affordance).
- ceasing display of the second move affordance e.g., by shifting or animating movement of the move affordance in a direction toward the location corresponding to the second resize affordance and/or gradually decreasing a size of the second move affordance until it disappears
- displaying the second resize affordance e.g., by shifting or animating movement of the resize affordance as coming from the second move affordance and
- the animated transition described with reference to Figures 7E-7F includes moving grabber 706-1 to the right as it morphs into resize affordance 708-1.
- Displaying an animated transition that shifts the move grabber in a first direction, such that the move grabber is moved over until it is animated into display of a resize control enables the user to access the resize control and resize the virtual object without displaying additional controls, and provides improved visual feedback that the user is gazing at the corner of the object associated with the resize control.
- the computer system while displaying a respective move affordance (e.g., the first move affordance, and/or the second move affordance) for moving a position of the first object in the three-dimensional environment, the computer system detects a fifth user input directed to the respective move affordance for moving the position of the first object; and in response to detecting the fifth user input directed to the respective move affordance, updates display of the first object from being displayed at a first object position in the first view of the three-dimensional environment to being displayed at a second object position, distinct from the first object position, in the first view of the three-dimensional environment.
- a respective move affordance e.g., the first move affordance, and/or the second move affordance
- the second object position is determined based on the fifth user input directed to the respective move affordance (e.g., a pinch and drag gesture, wherein the user input drags the first object to the second object position).
- the computer system displays the first object as gradually moving in accordance with a speed, distance, and/or direction of movement of the fifth user input directed to the respective move affordance.
- the computer system maintains display of the respective move affordance as the first object is moved to the second object position.
- the respective move affordance is displayed concurrently with the first object at the second object position (e.g., at an end of the fifth user input).
- the application window 702 is moved in response to a user input directed to grabber 706-3.
- Displaying a move grabber associated with a virtual object, and allowing a user to reposition the virtual object within a three-dimensional environment by selecting the move grabber reduces the number of inputs required access a move control without requiring the user to navigate complex menu hierarchies.
- updating the display of the first object from being displayed at the first object position in the first view of the three-dimensional environment to being displayed at the second object position includes updating the position of the first object in three different dimensions (e.g., x, y, z, directions in a Cartesian coordinate system; r, theta, phi in a polar coordinate system; or a horizontal dimension, a vertical dimension, and a visual depth dimension (e.g., a change in distance relative to a viewpoint of the user)) in the three-dimensional environment.
- the user input for repositioning the first object from the first object position to the second object position includes movement of an input object in three different dimensions in the physical environment.
- the user is enabled to move grabber 706-4 and associated application window 702 in at least three dimensions.
- Displaying a move grabber associated with a virtual object, and allowing a user to reposition the virtual object in three- dimensions, including a horizontal position, a vertical position, and a depth of the object within a three-dimensional environment by selecting the move grabber reduces the number of inputs required to navigate complex menu hierarchies, and provides improved visual feedback for repositioning the virtual object within the three-dimensional environment.
- the computer system while displaying the first object, displays a third move affordance for repositioning a position of the first object in the first view of the three-dimensional environment (e.g., the same as the first move affordance, and/or the second move affordance), and detects an eleventh gaze input directed to a respective portion of the first object (e.g., the eleventh gaze input moved from the location that corresponds to the third move affordance to the location that corresponds to a resize affordance; or the eleventh gaze input is unrelated to the display of the third move affordance).
- the first control element is the third move affordance.
- the second control element is the third move affordance.
- the third move affordance is optionally displayed for the first object in response to detecting a gaze input for the user that is directed to a center region of the first object (e.g., the first portion).
- the computer system while detecting the user is gazing at any portion of the first object, the computer system automatically displays the third move affordance for the first object and, optionally maintains display of the third move affordance until detecting a gaze input is detected on another respective portion of the first object (e.g., a second portion, such as a corner, of the first object).
- the computer system in response to detecting the eleventh gaze input directed to the respective portion of the first object (e.g., in response to detecting the eleventh gaze input including detecting movement of a gaze input from a center of the first object, or from a center of the bottom edge (e.g., the first portion) of the first object, to a comer of the first object (e.g., the respective portion that corresponds to the first resize affordance)): in accordance with a determination that the respective portion of the first object corresponds to a first type of control element for the first object (e.g., the first type of control element includes a resize affordance), the computer system ceases display of the third move affordance and displays a respective instance of the first type of control element (e.g., displaying a third resize affordance for resizing the first object at or proximate to the respective portion of the first object that corresponds to the first type of control element).
- displaying the respective instance of the first type of control element includes animating a transition between ceasing display of the third move affordance (e.g., by shifting or animating movement of the third move affordance in a direction toward the respective instance of the first type of control element and/or gradually decreasing a size of the third move affordance until it disappears) and displaying the respective instance of the first type of control element (e.g., by shifting or animating movement of the respective instance of the first type of control element as coming from the third move affordance and/or gradually increasing a size of the respective instance of the first type of control element).
- the animated transition described with reference to Figures 7D-7F includes gradually shrinking grabber 706-1 until it is no longer displayed and gradually displaying resize affordance 708-4.
- Replacing display of a move grabber for changing a position of a virtual object with display of another type of control, such as a resize control in response to detecting the user gazing at a comer of the object associated with the resize control, enables the user to access the resize control and resize the virtual object without displaying additional controls, reduces the number of inputs required to navigate complex menu hierarchies, and provides improved visual feedback that the user is gazing at the comer of the object associated with the resize control.
- the computer system in response to detecting the eleventh gaze input directed to the respective portion of the first object (e.g., in response to detecting the eleventh gaze input including detecting movement of a gaze input from a center of the first object, or from a center of the bottom edge (e.g., the first portion) of the first object, to a corner of the first object (e.g., the respective portion that corresponds to the first resize affordance)): in accordance with a determination that the respective portion of the first object corresponds to a second type of control element for the first object (e.g., the second type of control element includes a close affordance) that is different from the first type of control element, the computer system maintains display of the third move affordance and displaying a respective instance of the second type of control element (e.g., displaying a close affordance for closing the first object at or proximate to the respective portion of the first object that corresponds to the second type of control element).
- grabber 706-1 continues to be displayed. Maintaining display of a move grabber for changing a position of a virtual object with display of another type of control, such as a resize control, in response to detecting the user gazing at a corner of the object associated with the other type of control, enables the user to access additional controls without requiring additional user input to navigate complex menu hierarchies.
- another type of control such as a resize control
- displaying the first control element that corresponds to the first operation associated with the first object in response to detecting that the first gaze input meets the first criteria with respect to the first portion of the first object includes: in accordance with a determination that the first object is displayed as a two-dimensional object in the three-dimensional environment, displaying the first control element (e.g., the resize affordances) at a first position with a first spatial relationship to the first object (e.g., the first control element is displayed at a position corresponding to the first portion of the first object).
- the first control element e.g., the resize affordances
- an object is considered a two-dimensional object even if the two-dimensional object has some thickness (e.g., a perceived depth such that the object is actually more than two-dimensional) but is treated as a two-dimensional object (e.g., in accordance with having a relatively small amount of thickness and/or appears mostly flat to the user).
- a resize affordance is displayed proximate to the bottom right corner of the application window; and in response to detecting the gaze input directed to a bottom left comer of the application window, a resize affordance is displayed proximate to the bottom left corner of the application window.
- resize affordance 708-1 in Figures 7F is displayed at a corner of application window 702.
- Displaying a resize affordance at a corner of a two-dimensional object in response to detecting the user gazing at the corner of the object, enables the user to access additional controls and provides improved visual feedback.
- displaying the first control element that corresponds to the first operation associated with the first object in response to detecting that the first gaze input meets the first criteria with respect to the first portion of the first object includes: in accordance with a determination that the first object is displayed as a three-dimensional object in the three-dimensional environment, displaying the first control element (e.g., the resize affordances) at a second position with a second spatial relationship to the first object (e.g., the first control element is displayed at a position corresponding a platter supporting the first object (e.g., comers of the platter), and not at a position of the first portion of the first object), wherein the first spatial relationship is different from the second spatial relationship.
- the first control element e.g., the resize affordances
- the computer system displays a virtual platter at the bottom of the first object to support the first object, and the resize affordance(s) are displayed at the corners of the virtual platter.
- the virtual platter is translucent or invisible in the first view of the three-dimensional environment, and the user perceives the presence of the virtual platter by the appearances of the control elements (e.g., resize affordances, move affordance, or other control affordances) that have been displayed for the first object.
- the control elements e.g., resize affordances, move affordance, or other control affordances
- the computer system in response to detecting that the first gaze input meets the first criteria with respect to the first portion of the first object (e.g., the bottom portion of the first object, and/or within a threshold distance from the bottom portion of the first object): in accordance with a determination that the first object is displayed as a three-dimensional object in the three-dimensional environment, the computer system displays, via the first display generation component, a second object in the three-dimensional environment, wherein the second object is displayed as a three-dimensional application object with a third spatial relationship to the first object (e.g., at the bottom of the first object, supporting the first object).
- platter 7029 is displayed below virtual object 7028. Displaying a platter underneath and/or above a three-dimensional virtual object enables the user to more easily interact with the three-dimensional object, and provides improved visual feedback reflecting the dimension of the virtual object, including a depth of the virtual object.
- displaying e.g., in response to detecting a gaze input directed to the bottom of the second object (e.g., a virtual platter supporting the first object) or below the second object, or automatically displayed when the second object is displayed) a second control element (e.g., a close affordance for closing the first object that is a three- dimensional object) below the second object.
- a second control element e.g., a close affordance for closing the first object that is a three- dimensional object
- close affordance 717 is displayed below platter 7029. Automatically displaying a close affordance at a corner of a three-dimensional object in response to detecting the user gazing at the corner of the object, enables the user to access additional controls and provides improved visual feedback.
- the computer system prior to displaying the first object in the first view of the three-dimensional environment, detects a sixth user input that corresponds to a request to display the first object in the first view of the three-dimensional environment; and in response to detecting the sixth user input that corresponds to the request to display the first object in the first view of the three-dimensional environment: concurrently displays the first object in the first view of the three-dimensional environment with a first set of control elements for the first object, including the first control element (e.g., including the first control element described above that is displayed in response to detecting that the first gaze input meets the first criteria (e.g., the first resize affordance, the close affordance and/or the first move affordance)(e.g., to educate the user of the existence of the one or more control elements, and their corresponding locations); and after concurrently displaying the first object with the first set of control elements for the first object for a threshold amount of time, ceases display of the first set of control elements for the first object while maintaining
- the first control element e.
- the computer system after ceasing display of the first set of control elements, the computer system detects a gaze input directed to a respective portion of the first object, wherein the respective portion of the first object corresponds to a respective control element of the first set of control elements.
- the first control element corresponds to the respective control element of the first set of control elements.
- the computer system displays, upon display of the first object, a close affordance as the first control element (e.g., along with other control elements, such as the resize affordances, and/or move affordance); and after ceasing display of the close affordance, in response to detecting a gaze input on a first portion of the first object, the computer system redisplays the close affordance (e.g., the first control element).
- a close affordance as the first control element
- the close affordance e.g., the first control element
- Automatically displaying available control options before ceasing display of the control options, such that the user is enabled to invoke a respective control of the available control options provides the user with improved visual feedback such that the user is aware of the available controls without displaying additional instructions and without cluttering the view of the three-dimensional environment by maintaining display the available control options.
- the first set of control elements includes a close affordance for ceasing display of the first object in the three-dimensional environment.
- the computer system detects a gaze input that is directed to a respective portion of the first object that corresponds to the display location of the close affordance; and in response to detecting that the gaze input that is directed to the respective portion of the first object that corresponds to the display location of the close affordance, redisplaying the close affordance for the first object.
- Automatically displaying available control options at respective positions, including a close control, before ceasing display of the control options, such that the user is enabled to invoke the close control by gazing at the respective position corresponding to the close control provides the user with improved visual feedback such that the user is aware of the available controls without displaying additional instructions and without cluttering the view of the three-dimensional environment by maintaining the available control options.
- the computer system ceasing display of the first set of control elements for the first object while maintaining display of the first object includes: moving the first set of control elements toward the first object while changing one or more visual properties of the first set of control elements (e.g., gradually reducing visual prominence of the first set of control elements by reducing the size, brightness, color saturation, and/or opacity of the first set of control elements, as the set of control elements are animated toward the first object from their respective display locations); and ceasing display of the first set of control elements after moving the first set of control elements and changing the one or more visual properties of the first set of control elements.
- moving the first set of control elements toward the first object while changing one or more visual properties of the first set of control elements e.g., gradually reducing visual prominence of the first set of control elements by reducing the size, brightness, color saturation, and/or opacity of the first set of control elements, as the set of control elements are animated toward the first object from their respective display locations
- ceasing display of the first set of control elements after moving the
- the available controls are animated as moving toward application window 720 as the available controls disappear.
- Animating available control options as moving toward a virtual object before ceasing display of the control options, such that the control options gradually disappear, provides the user with improved visual feedback.
- the computer system while displaying the first object, displays a fourth move affordance (e.g., same as the first, second, and third move affordance, or a different move affordance) for repositioning a position of the first object in the first view of the three-dimensional environment (e.g., the fourth move affordance is displayed without requiring a prior gaze input, or the fourth move affordance is displayed in response to a prior gaze input directed to a respective portion of the first object that corresponds to the reactive region for displaying the fourth move affordance).
- a fourth move affordance e.g., same as the first, second, and third move affordance, or a different move affordance
- the computer system while displaying the fourth move affordance with the first object, the computer system detects a twelfth gaze input (e.g., same as the first gaze input, the second gaze input, or is another gaze input) directed to a respective portion of the three-dimensional environment that corresponds to the first object (e.g., a respective portion that corresponds to a preview of the close affordance displayed next to the fourth move affordance, to a region proximate to the first object, or to a portion of the first object).
- a twelfth gaze input e.g., same as the first gaze input, the second gaze input, or is another gaze input
- a twelfth gaze input directed to a respective portion of the three-dimensional environment that corresponds to the first object (e.g., a respective portion that corresponds to a preview of the close affordance displayed next to the fourth move affordance, to a region proximate to the first object, or to a portion of the first object).
- the computer system in response to detecting the twelfth gaze input directed to the respective portion of the three-dimensional environment that corresponds to the first object, the computer system: in accordance with a determination that the twelfth gaze input corresponds to a request to display a close affordance for the first object (e.g., the twelfth gaze input meets the first criteria, optionally, while the gaze input is directed to any portion of the first object, a portion of the first object that corresponds to a preview of the close affordance, or a portion of the first object that does not correspond to the reactive region for displaying the resizing affordance and does not correspond to the reactive region for dragging the fourth move affordance), displays a respective close affordance adjacent to the fourth move affordance (e.g., to the right and/or to the left of the fourth move affordance).
- the close affordance and the move affordance are displayed concurrently below (e.g., along a Y-axis in three-dimensional space) the first object.
- the first control element is a first preview of the close affordance (e.g., a dot) that, in response to detecting another gaze input directed to the first preview of the close affordance or a continuation of the first gaze input that caused the display of the first preview of the close affordance, the computer system transforms the first preview of the close affordance to the actual close affordance that, when activated, closes the first object.
- the first control element is a close affordance
- the preview of the close affordance e.g., an indicator in the shape of a dot, a miniature version of the close affordance, or another appearance
- the preview of the close affordance is displayed adjacent to the move affordance of the first object before any gaze input is detected to transform the preview of the close affordance into the actual close affordance adjacent to the move affordance.
- the computer system in response to detecting the twelfth gaze input directed to the respective portion of the three-dimensional environment that corresponds to the first object: in accordance with a determination that the twelfth gaze input does not correspond to a request to display a close affordance for the first object (e.g., the twelfth gaze input does not meet the first criteria, or the gaze input is not directed to the first object, or a portion of the first object that corresponds to a preview of the close affordance), the computer system forgoes displaying a respective close affordance adjacent to the fourth move affordance (e.g., to the right and/or to the left of the fourth move affordance).
- a respective close affordance adjacent to the fourth move affordance e.g., to the right and/or to the left of the fourth move affordance
- user interface object 705 e.g., a close affordance and/or one or more other controls (e.g., a control menu)
- grabber 706-1 e.g., the fourth move affordance
- Dynamically displaying a close control for closing an application window proximate to a grabber for moving the application window in response to a user gazing at a minimized version of the close control displayed proximate to the grabber in an AR/VR environment enables the user to access additional controls for the application window, including an expanded close control, without displaying additional controls that obscure the user’s view of the AR/VR environment and reduces the number of inputs required to navigate complex menu hierarchies.
- the computer system prior to detecting the twelfth gaze input directed to the respective portion of the three-dimensional environment that corresponds to the first object, displays a first preview of the respective close affordance (e.g., a dot with a first size that does not include text or an icon, or another indicator appearance for the close affordance) adjacent to the fourth move affordance, wherein the determination that the twelfth gaze input corresponds to a request to display a close affordance for the first object includes a determination that the twelfth gaze input is directed to the first preview of the respective close affordance.
- a first preview of the respective close affordance e.g., a dot with a first size that does not include text or an icon, or another indicator appearance for the close affordance
- the first control element is a close affordance
- the first preview of the close affordance e.g., an indicator in the shape of a dot, a miniature version of the close affordance, or another appearance
- the user interface object 705 is displayed next to grabber 706-1 before detecting the user’s attention directed to the bottom portion of application window 702.
- a respective control such as a minimized dot that can be expanded into a close control, before detecting a user’s gaze to access the close control, reduces the number of inputs needed to access the close control without obscuring the user’s view of the AR/VR environment by continuing to display the minimized dot without requiring user input.
- the computer system while displaying the respective close affordance (e.g., displayed in response to detecting the twelfth gaze input, and/or displayed adjacent to the fourth move affordance), the computer system detects, via the one or more input devices, a seventh user input directed to the respective close affordance (e.g., an air tap gesture, an air pinch gesture, or another activation or selection gesture, optionally while a gaze input is directed to the respective close affordance); and in response to detecting the seventh user input directed to the respective close affordance, in accordance with a determination that the seventh user input meets selection criteria (e.g., the seventh user input includes a selection gesture, such as an air tap gesture or an air pinch gesture, at the location of the respective close affordance or the seventh user input includes a selection gesture detected in conjunction with a gaze input directed to the respective close affordance), the computer system closes the first object, including ceasing display of the first object in the first view of the three- dimensional environment (along with the control elements, such as the respective close
- the computer system in response to detecting the user input directed to the close affordance 7030-2, ceases display of the application window 702 (e.g., closes the application window 7020). Allowing a user to perform an operation for a virtual object, such as closing the virtual object such that the virtual object ceases to be displayed, by performing an air gesture without requiring the user to connect, set up, or use a separate input device to perform the operation, reduces the number of inputs needed to close the virtual object.
- the computer system while displaying the first object in the first view of the three-dimensional environment, the computer system detects, via the one or more input devices, a thirteenth gaze input moving relative to a respective portion of the first object that corresponds to a respective control element (e.g., the first control element, or another control element) of the first object (e.g., while the respective control element is not yet displayed, or while the respective control element is currently displayed).
- a respective control element e.g., the first control element, or another control element
- the computer system determines whether user attention is directed to the respective portion of the first object that corresponds to the respective control element of the first object, including: in accordance with a determination that the thirteenth gaze input moving relative to the respective portion of the first object has moved into a first respective portion of the first object that corresponds to the respective control element of the first object, wherein the first respective portion of the first object that corresponds to the respective control element is of a first size, determining that user attention is directed to the respective portion of the first object that corresponds to the respective control element (and optionally, performing a respective operation (e.g., performing the first operation, or performing another operation different from the first operation) with respect to the respective control element (e.g., if the respective control element is not displayed yet, displaying the respective control element of the first object with the first object in the first view of the three-dimensional environment; or activating the respective control element
- the computer system determines that user attention is not directed to the respective portion of the first object that corresponds to the respective control element.
- the computer system determines whether user attention is directed to the respective portion of the first object that corresponds to the respective control element of the first object, including in accordance with a determination that the thirteenth gaze input moving relative to the respective portion of the first object has moved from within the first respective portion of the first object to a region outside of the first respective portion of the first object and within a second respective portion of the first object that corresponds to the respective control element of the first object, wherein the second respective portion of the first object that corresponds to the respective control element is of a second size, larger than the first size, determining that user attention is remaining on the respective portion of the first object that corresponds to the respective control element; and in accordance with a determination that the thirteenth gaze input moving relative to the respective portion of the first object has moved from within the first respective portion of the first object to a region outside of the second respective portion of the first object that corresponds to the respective control element;
- the second respective portion of the first object encloses the first respective region of the first object (e.g., the second respective portion is the entire first object and the first respective portion is a portion, less than all, of the first object; or the second respective portion includes a portion of the first object and optionally a portion of the three-dimensional environment immediately outside of the first object, and the first respective portion includes a portion, less than all, of the second respective portion (e.g., the portion within the first object and/or the portion within the environment immediately outside of the first object)).
- the user’s attention 710-2 is directed to a first area having a first size that corresponds to a respective portion of the application window 702 (e.g., to determine that the user’s attention is directed to the respective portion of the application window 702) and detecting that the user’s attention is no longer directed to the bottom right comer of the application window 702 includes determining that the user’s attention is directed outside of a second area having a second size (e.g., distinct from the first area having the first size that corresponds to a respective portion of the application window 702 used to determine that the user’s attention satisfies attention criteria).
- a second area having a second size e.g., distinct from the first area having the first size that corresponds to a respective portion of the application window 702 used to determine that the user’s attention satisfies attention criteria.
- Using a first sized region to determine that the user’s gaze is directed to a respective control in an AR/VR environment, and dynamically displaying the control in response to detecting that the user’s gaze is directed to the first sized region, while using a second sized region that is different from the first sized region to determine that the user’s gaze is no longer directed to the respective control reduces the number of inputs needed for the user to specify whether the user intends to access the control, thereby providing improved feedback for the user and providing the user with access to additional control options by detecting the user’s gaze directed to respective controls without requiring additional user input.
- the computer system while displaying the first object in the first view of the three-dimensional environment, the computer system detects an eighth user input while user attention is directed to the first portion of the first object; and in response to detecting the eighth user input while user attention is directed to the first portion of the first object, the computer system initiates performance of one or more operations corresponding to the first portion of the first object (e.g., displaying one or more control elements of the first object, activating one or more control elements of the first object, and/or using one or more control elements of the first object to interact with the first object) in accordance with the eighth user input.
- the eighth user input e.g., displaying one or more control elements of the first object, activating one or more control elements of the first object, and/or using one or more control elements of the first object to interact with the first object
- the computer system while performing the one or more operations corresponding to the first portion of the first object, the computer system detects that user attention has moved away from the first portion of the first object (e.g., user’s gaze has moved outside of the first respective portion and/or the second respective portion of the first object); and after (e.g., in response to) detecting that user attention has moved away from the first portion of the first object and in accordance with a determination that the one or more operations corresponding to the first portion of the first object is ongoing, the computer system continues to perform the one or more operations corresponding to the first portion of the first object (in accordance with one or more inputs such as a continuation of an air gesture that started before the user attention moved away from the first portion of the first object).
- one or more inputs such as a continuation of an air gesture that started before the user attention moved away from the first portion of the first object.
- the user can continue to move the first object in accordance with the movement gesture that was detected while the user’s attention was directed to the move affordance and that is still continuing, even if the user’s gaze has moved away from the move affordance.
- the user can continue to resize the first object in accordance with the resize gesture that was detected while the user’s attention was directed to the resize affordance and that is still continuing, even if the user’s gaze has moved away from the resize affordance.
- the user is not enabled to interact with (e.g., via a gaze and air gesture input) the resize affordance.
- the user input detected via the user’s hand 7020 is enabled to continue (e.g., the user is enabled to continue resizing the window).
- Enabling a user to continue the user’s interaction with a virtual object even after detecting that the user’s attention has shifted away from the virtual object, enables the user to continue to perform operations for the virtual object without displaying additional controls and without interrupting the user’s interaction, thereby reducing the number of inputs required to continue performing the operations.
- the first criteria require that the first gaze input is maintained on the first portion of the first object for at least a first threshold amount of time in order for the first criteria to be met; and, displaying the first control element includes displaying the first control element after the first gaze input is directed to the first portion of the first object for at least the first threshold amount of time (e.g., display of the first control element is delayed until the first time threshold has been met).
- the resize affordance 708-1 is displayed in accordance with a determination that the user’s attention 710-2 is detected for at least a threshold amount of time.
- the display of resize affordance 708-1 is delayed until the threshold amount of time has been met.
- Dynamically displaying a respective control in response to a user gazing at a corresponding respective portion of an application window for at least a threshold amount of time in an AR/VR environment enables the user to access additional controls for the application window without displaying additional controls that obscure the user’s view of the AR/VR environment.
- the computer system detects a start of a first portion of the first user input while the first gaze input is directed to the first portion of the first object and before the first control element is displayed.
- the computer system in response to detecting the start of the first portion of the first user input while the first gaze input is directed to the first portion of the first object and before the first control element is displayed, the computer system: in accordance with a determination that operation performance criteria are met (e.g., operation performance criteria are met by the start of the first user input in accordance with a determination that the first gaze input has met the first criteria, and that the first portion of the first object corresponds to the first control element that is a first type of control element (e.g., a move affordance, a resize affordance, and/or a close affordance)), initiates performance of the first operation (e.g., performs an initial portion of the first operation, such as starting to resize the first object in accordance with the first portion of the first user input, or dragging
- operation performance criteria e.g
- a second type of control element e.g., a close affordance, or another type of affordance
- the computer system enables the user to access functionality and/or perform operations associated with the resize affordance 708-1.
- Enabling a user to initiate performing an operation, such as changing a size of a virtual object, before displaying an affordance corresponding to the operation enables the user to access additional functionality for the virtual object without displaying additional controls, and provides improved visual feedback by dynamically displaying the affordance while the user continues to perform the operation.
- aspects/operations of methods 900, 1000, 1100, 1200, 1300, 1400, 1500, 1600,1800, and 20000 may be interchanged, substituted, and/or added between these methods.
- the first object (e.g., application window or virtual object) in the method 800 has characteristics of the first user interface object (e.g., application window or virtual) in the method 900, and the first control elements that are displayed may be replaced by, or concurrently displayed with, other user interface elements (e.g., the first control element, as in the method 800 above, can be displayed as a title bar, as in the method 1000 below). For brevity, these details are not repeated here.
- FIG 9 is a flow diagram of an exemplary method 900 for updating visual properties of controls in response to user interactions, in accordance with some embodiments.
- method 900 is performed at a computer system (e.g., computer system 101 in Figure 1) including a display generation component (e.g., display generation component 120 in Figures 1 A, 3, and 4 or display generation component 7100) (e.g., a headsup display, a display, a touchscreen, a projector, etc.) and one or more cameras (e.g., a camera (e.g., color sensors, infrared sensors, and other depth-sensing cameras) that points downward at a user’s hand or a camera that points forward from the user’s head).
- a computer system e.g., computer system 101 in Figure 1
- a display generation component e.g., display generation component 120 in Figures 1 A, 3, and 4 or display generation component 7100
- a headsup display e.g., a headsup display, a display,
- the method 900 is governed by instructions that are stored in a non-transitory (or transitory) computer-readable storage medium and that are executed by one or more processors of a computer system, such as the one or more processors 202 of computer system 101 (e.g., control 110 in Figure 1 A). Some operations in method 900 are, optionally, combined and/or the order of some operations is, optionally, changed.
- Method 900 is a method for displaying a control with a first appearance in response to the user gazing at the control, and displaying the control with a second appearance after detecting the user is interacting with the control to perform an operation. Automatically updating an appearance of the control when the user is gazing at the control and further updating the appearance of the move control when the user is interacting with the control provides the user with improved visual feedback of the user’s interactions.
- method 900 is performed at a computer system that is in communication with a first display generation component (e.g., a first display generation component of a two-sided display generation component, a heads-up display, a head-mounted display (HMD), a display, a touchscreen, a projector, a standalone display, and/or a display that is enclosed in the same housing as another display generation component of the same type or a different type) and one or more input devices (e.g., cameras, controllers, touch- sensitive surfaces, joysticks, buttons, gloves, watches, motion sensors, and/or orientation sensors).
- the first display generation component is a display component facing the user and provides an XR experience to the user.
- the first display generation component includes two or more display components (e.g., one set for each eye) that display slightly different images to form a stereoscopic view of the three-dimensional environment.
- the first display generation component and a second display generation component form a two-sided display device (e.g., a two-sided HMD) that displays a first user interface on a first side corresponding to the first display generation component, and a second user interface on the second side corresponding to the second display generation component.
- the second display generation component is a display component facing away from the user and toward an external environment of the user and optionally provides status information related to the first display generation component (e.g., displayed content and/or operational state) and/or the user (e.g., movement of the user’s eyes, and/or attention state of the user) to other users in the external environment.
- the computing system is an integrated device with one or more processors and memory enclosed in the same housing as the first and the second display generation components and at least some of the one or more input devices.
- the computing system includes a computing component (e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer) that includes one or more processors and memory that is separate from the display generation component(s) and/or the one or more input devices.
- a computing component e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer
- the display generation component s) and the one or more input devices are integrated and enclosed in the same housing.
- the computer system displays (902), via the first display generation component, a first user interface object (e.g., the first user interface object is an application window, a media display area, a viewport into a three-dimensional environment, an application view, or a virtual object) and a first control element (e.g., a respective one of one or more grabbers or handles for dragging, resizing, adjusting a shape and/or size, adjusting a visual property of a corresponding object) that is associated with performing a first operation (e.g., dragging, resizing, adjusting a shape and/or size of, and/or adjusting a visual property of) with respect to the first user interface object, in a first view of a three-dimensional environment (e.g., a passthrough view of a physical environment, an AR environment, or a VR environment), wherein the first control element is spaced apart from (e.g., separate from, not attached to, changes size and/or location separately from, is spaced
- application window 702 is displayed with close affordance 7030 and grabber 706-1
- application window 702 is displayed with resize affordance 708-1
- the first control element is close affordance 7030, grabber 706-1, and/or resize affordance 708-1.
- the computer system detects (904), via the one or more input devices, a first gaze input that is directed to the first control element. For example, in Figures 7F-7G, the user’s attention 710- 4 is directed to resize affordance 708-1.
- the computer system updates (906) an appearance of the first control element from the first appearance to a second appearance that is different from the first appearance (e.g., the first control element is displayed with a second set of values for the first set of visual properties, wherein the second set of values are different from the first set of values) (e.g., the first control element changes brightness, color, luminance, translucency, opacity, color saturation, simulated optical property or lighting effect, size, and/or animated movements, when a gaze input is directed to the first control element).
- a color of resize affordance 708-2 is updated in response to detecting the user’s attention 710-4 directed to resize affordance 708- 1.
- the computer system detects (908), via the one or more input devices, a first user input directed to the first control element (e.g., the user input is a gesture input or another type of input detected via an input device, where the first user input is detected in conjunction with the first gaze input).
- a first user input directed to the first control element e.g., the user input is a gesture input or another type of input detected via an input device, where the first user input is detected in conjunction with the first gaze input.
- a user input via the user’s hand 7020 is detected.
- the computer system updates (910) the appearance of the first control element from the second appearance to a third appearance that is different from the first appearance and the second appearance and that indicates that additional movement associated with the first user input (e.g., a movement of the input object, and/or a drag gesture following the air tap or pinch gesture or while maintaining the pinch gesture) will cause the first operation (e.g., dragging, resizing, adjusting a shape and/or size, adjusting a visual property of a corresponding object (e.g., the first user interface object)) associated with the first control element to be performed (e.g., the first control element is displayed with a third set
- the gaze of the user on the first control element with the first appearance causes the computer system to display the first control element with a bigger size, a greater brightness, and/or a different color; while the first user input directed to the first control element with the second appearance causes the computer system to further change the size, the shape and/or the simulated optical property and lighting effect of the first control element.
- the resize affordance 708-5 is updated in a second manner by changing a size of the resize affordance 708-5.
- grabber 706-4 is updated in two manners, a first manner in response to detecting the user’s attention 710-20 and a second manner in response to detecting a user input via user’s hand 7020.
- the computer system while displaying the first control element with the third appearance, the computer system detects a second user input that includes additional movement associated with the first user input directed to the first control element (e.g., the second user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the input object).
- the second user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the input object.
- the computer system in response to detecting the second user input, performs the first operation with respect to the first user interface object in accordance with the additional movement of the second user input (e.g., resizes the first user interface object, moving the first user interface object, and/or adjusts a property of the first user interface object, e.g., by an amount and/or by a respective type of change, in accordance with a direction, distance, duration, speed, and/or other characteristics of the additional movement of the second user input).
- the additional movement of the second user input e.g., resizes the first user interface object, moving the first user interface object, and/or adjusts a property of the first user interface object, e.g., by an amount and/or by a respective type of change, in accordance with a direction, distance, duration, speed, and/or other characteristics of the additional movement of the second user input.
- application window 702 is resize in response to the user input directed to resize affordance 708-5, and in Figures 7X-7Z, application window 702 is moved in response to the user input directed to grabber 706-4. Allowing the user to perform an operation by providing an input that selects and moves a control that is dynamically displayed while the user is gazing at the control reduces the number of inputs needed to perform the operation without displaying additional controls.
- the computer system while displaying the first control element with the second appearance (e.g., before the second user input that includes the additional movement associated with the first user input is detected, and before the first control element takes on the third appearance), the computer system detects, via the one or more input devices, that the first gaze input is no longer directed to the first control element (e.g., the first gaze input is directed to a second control element, or to an area in the three-dimensional environment (e.g., a portion of the representation of the physical environment, the interior of the first user interface object, or other objects or controls in the three-dimensional environment) that does not correspond to the first control element; or no gaze input is detected).
- the first gaze input is no longer directed to the first control element
- the first gaze input is directed to a second control element, or to an area in the three-dimensional environment (e.g., a portion of the representation of the physical environment, the interior of the first user interface object, or other objects or controls in the three-dimensional environment) that does not correspond to the first control element; or no gaze
- the computer system in response to detecting that the first gaze input is no longer directed to the first control element while displaying the first control element with the second appearance, restores the appearance of the first control element from the second appearance to the first appearance.
- the computer system detects the first gaze input moving away from the first control element, and in response to detecting the first gaze input moving away from the first control element, the computer system ceases to display the first control element with the changed visual appearance (e.g., cease the visual emphasis of the first control element). For example, while the first control element is no longer the subject of the user’s attention, the first control element is restored to its normal appearance (e.g., is not visually emphasized, and/or its first appearance).
- the computer system while displaying the first control element with the third appearance (e.g., after detecting the first user input directed to the first control element, and while detecting the second user input directed to the first control element), the computer system detects, via the one or more input devices, that the first gaze input is no longer directed to the first control element (e.g., the first gaze input is directed to a second control element, or to an area in the three-dimensional environment (e.g., a portion of the representation of the physical environment, the interior of the first user interface object, or other objects or controls in the three-dimensional environment) that does not correspond to the first control element; or no gaze input is detected).
- the first gaze input is no longer directed to the first control element
- the first gaze input is directed to a second control element, or to an area in the three-dimensional environment (e.g., a portion of the representation of the physical environment, the interior of the first user interface object, or other objects or controls in the three-dimensional environment) that does not correspond to the first control element; or no gaze input is
- the computer system in response to detecting that the first gaze input is no longer directed to the first control element while displaying the first control element with the third appearance, maintaining display of the first control element with the third appearance (e.g., after the first control element has been selected by the user with the first user input, the computer system maintains the third appearance of the first control element even if the user’s gaze is no longer directed to the first control element, optionally until the user deselects the first control element (e.g., until the first user input ceases to be detected, and/or until the release of a tap-hold input or release of the pinch posture of the hand)).
- the resize affordance 708-5 continues to be displayed with the updated visual properties, including a change in color and a change in size as compared to resize affordance 708-1. Maintaining an appearance of a control, after the user has selected the control, even if the user looks away from the control after selecting the control, provides the user with improved visual feedback that the control is still selected and reduces a number of user inputs required to reselect the control by maintaining selection of the control even while the user has looked away.
- the first user interface object is a first application window (e.g., a window associated with a first application (e.g., a user application, a system application, a viewport of a three-dimensional experience, a control panel, a settings user interface, and/or a communication application providing a communication user interface or shared experience for multiple users)) displayed in the first view of the three-dimensional environment, and the first control element is associated with performing the first operation with respect to the first application window.
- the first application window is displayed as a two-dimensional object or pseudo-three-dimensional object in the three-dimensional environment.
- application window 702 is considered a two- dimensional application window and virtual object 7028 is considered a three-dimensional object.
- updating the appearance of the first control element from the first appearance to the second appearance includes updating a color of the first control element from a first color to a second color that is different from the first color.
- updating the appearance of the first control element includes changing a translucency and/or opacity of the first control element.
- the first control element is updated from being displayed with a first color, such as grey, to a second color, such as white. For example, as described with reference to Figure 7G, the color of resize affordance 708-2 is updated in response to detecting the user’s attention is directed to the resize affordance.
- grabber 706-3 is displayed with an updated color in response to detecting the user’s attention is directed to grabber 706-2.
- Automatically updating an appearance of a control by changing a color of the control in response to detecting that the user is looking at the control provides the user with improved visual feedback of the user’s interactions.
- updating the appearance of the first control element from the second appearance to the third appearance includes updating a size of the first control element from a first size to a second size that is different from the first size.
- the first control element decreases or increases in size to indicate that the user’s gaze is detected on the first control (e.g., in addition to changing its color and/or translucency).
- the first control element increases in size to a third size larger than the first size and the second size indicate that the first operation is about to be performed with respect to the first user interface object if additional movement associated with the first user input is detected.
- a size of resize affordance 708-5 is updated in response to detecting a user input selecting resize affordance 708-5.
- a size of grabber 706-4 is updated in response to detecting a user input selecting grabber 706-4. Automatically updating an appearance of a control by changing a size of the control in response to detecting that the user has selected the control provides the user with improved visual feedback of the user’s interactions, including visual feedback that the control is currently selected.
- updating the size of the first control element from the first size to the second size includes decreasing the size of the first control element from the first size to the second size that is smaller than the first size (e.g., the second size has a smaller bounding box, or a smaller size in a first dimension and a smaller or same size in other dimensions).
- the first control element decreases in size and is maintained at the decreased size while the user’s gaze is continued to be detected as directed to the first control element (e.g., before detecting user selection of the first control element by the first user input).
- grabber 706-4 is displayed with a smaller size than grabber 706-3. Automatically updating an appearance of a control by decreasing a size of the control in response to detecting that the user has selected the control provides the user with improved visual feedback of the user’s interactions, including visual feedback that the control is currently selected.
- decreasing the size of the first control element from the first size to the second size that is smaller than the first size includes decreasing a first dimension of the first control element by a first amount, and decreasing a second dimension of the first control element by a second amount that is different than the first amount (e.g., shrink the first control element in the horizontal dimension more than the vertical dimension, decrease the width more than decreasing the length, or changing an aspect ratio of the first control element while reducing the size of at least one dimension of the first control element).
- changing a size of resize affordance 708-5 includes changing a width, or thickness of resize affordance 708-5 without changing a length of resize affordance 708-5.
- changing the size of resize affordance 708-5 changes a first dimension of resize affordance 708-5 without changing a second dimension of resize affordance 708-5.
- Automatically updating an appearance of a control by changing a size of the control in a first dimension without changing the size of the control in a second dimension in response to detecting that the user has selected the control, provides the user with improved visual feedback of the user’s interactions.
- the first control element is an object move control (e.g., a grabber) for moving the first user interface object (e.g., translating as a whole, and/or rotating around a pivot point or axis located within or outside of the first user interface object) within the three-dimensional environment.
- the object move control is displayed at a first position relative to the first user interface object.
- the object move control includes a grabber bar displayed at a nonzero distance (e.g., a gap through which content displayed behind the first user interface object and the grabber bar can become visible in the first view of the three-dimensional environment) from a bottom center edge of the first user interface object, in accordance with some embodiments.
- grabber 706-1 is a move control for application window 702, and as described with reference to Figures 7V-7Y, the visual properties of grabber 706-1 are updated in accordance with the user’s attention and a user input directed to grabber 706-1.
- Automatically updating an appearance of a move control for changing a position of an object when the user is gazing at the move control, and further updating the appearance of the move control when the user is interacting with the move control reduces the number of inputs needed to reposition the object and provides the user with improved visual feedback of the user’s interactions to move the object.
- the computer system while displaying the object move control with the third appearance, the computer system detects a third user input that includes first additional movement associated with the first user input directed to the object move control (e.g., the third user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the input object).
- the third user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the input object.
- the computer system moves the first user interface object in the three-dimensional environment in accordance with the first additional movement of the third user input (e.g., moving the first user interface object by an amount in accordance with a direction, distance, duration, speed, and/or other characteristics of the additional movement of the third user input), as described with reference to Figures 7W-7Z in response to selection of grabber 706-4.
- the third user input is detected while displaying the object move control (e.g., the first control element) with the third appearance.
- the object move control changes color in response to detecting the first gaze input directed to the object move control, and the object move control shrinks down in size in response to detecting the first user input selecting the object move control (optionally while maintaining its change in color).
- Allowing the user to move a position of a virtual object by providing an input that selects and moves a move control, and updating an appearance of the move control while the user is interacting with the move control reduces the number of inputs needed to move the object without displaying additional controls and improves visual feedback.
- the first control element is an object resize control for changing a size of the first user interface object within the three-dimensional environment.
- the first control is an object resize control displayed at a nonzero distance from a respective characteristic point or location (e.g., a respective one of two or four corners of a rectangular window, a respective apex of a polygonal object, or a set of cardinal points on a circumference of a circular or oval object or on a surface of a three- dimensional objects) on the first user interface object, as described with reference to Figures 7K-7L.
- a respective characteristic point or location e.g., a respective one of two or four corners of a rectangular window, a respective apex of a polygonal object, or a set of cardinal points on a circumference of a circular or oval object or on a surface of a three- dimensional objects
- the resize control is an L shape with a curvature that matches a curvature of the respective comer of the first user interface object at the characteristic point or location on the first user interface object. Allowing the user to change a size of a virtual object by providing an input that selects and moves a resize control, and updating an appearance of the resize control while the user is interacting with the resize control reduces the number of inputs needed to resize the object without displaying additional controls and improves visual feedback.
- the computer system while displaying the object resize control with the third appearance, the computer system detects a fourth user input that includes second additional movement of the first user input directed to the object resize control (e.g., the fourth user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the input object); and in response to detecting the fourth user input directed to the object resize control, changes the size of the first user interface object in accordance with the second additional movement of the fourth user input (e.g., changing the size and/or aspect ratio of the first user interface object by an amount in accordance with a direction, distance, duration, speed, and/or other characteristics of the second additional movement of the fourth user input).
- the fourth user input is an extension of the first user input, for example a pinch gesture followed by a drag gesture while maintaining the pinched posture of the hand; or movement of the input object followed by a tap or tap hold input of the
- the fourth user input is detected while displaying the object resize control (e.g., the first control element) with the third appearance.
- the object resize control changes color in response to detecting the first gaze input directed to the object resize control, and the object resize control shrinks down in size in response to detecting the first user input selecting the object resize control (optionally while maintaining its change in color).
- resize affordance 708-5 is moved in accordance with movement of the user’s hand 7020 to resize application window 702.
- Allowing the user to change a size of a virtual object by providing an input that selects and moves a resize control, and updating an appearance of the resize control while the user moves the user input to resize the object reduces the number of inputs needed to resize the object without displaying additional controls and improves visual feedback.
- the first control element is an object close control for ceasing display of the first user interface object in the three-dimensional environment.
- the object close control is an “x” or other exit control that is displayed at a nonzero distance from a respective characteristic point or location (e.g., a respective one of two or four comers of a rectangular window, a respective apex of a polygonal object, or a set of cardinal points on a circumference of a circular or oval object or on a surface of a three- dimensional objects) on the first user interface object.
- the computer system in response to detecting a user input directed to the object close control, ceases display of (or otherwise closes) the first user interface object in the three-dimensional environment.
- close affordance 7030 is a close control for application window 702
- close affordance 717 is a close control for virtual object 7028. Allowing the user to close a virtual object by providing an input that selects a close control, and updating an appearance of the close control while the user is interacting with the close control reduces the number of inputs needed to close the object without displaying additional controls and improves visual feedback.
- the computer system while displaying the first control element with a first respective appearance (e.g., the first appearance, the second appearance, or the third appearance), the computer system detects, via the one or more input devices, a second gaze input directed to a portion of the three-dimensional environment (e.g., the second gaze input is a continuation of the first gaze input following a movement of the user’s gaze away from an initial location of the first gaze input on the first control element) (e.g., the second gaze input optionally is on a different portion of the first control element or has moved away from the first control element onto another object (e.g., from the corner of the first user interface object to the center of the bottom edge of the first user interface object)); and in response to detecting the second gaze input directed to the portion of the three-dimensional environment: in accordance with a determination that the second gaze input is directed to a first location relative to the first control element in the three-dimensional environment (e.g., the first location has a first spatial relationship with the first control element (
- the computer system displays the first control element and updates the appearance of the first control element to indicate that the user’s gaze is detected as being directed to the first control element, if the user’s gaze moves while the first control element is displayed, the first control element changes its appearance to indicate the movement of the gaze input but remains displayed as long as the gaze has not moved too far away from the first control element; and if the user’s gaze moved to the location associated with another control element, such as a different corner of the first user interface object, or at a center portion of the first user interface object, the computer system ceases to display the first control element and, optionally, displays the second control element at or near the new location of the user’s gaze.
- FIG. 7G the user’s attention 710-4 is directed to resize affordance 708-2, which is illustrated with an updated color (e.g., as compared to resize affordance 708-1) in response to the user gazing at the resize affordance.
- the color of resize affordance 708-2 is maintained as the updated color, in accordance with a determination that the user’s attention 710-5 has shifted, or moved, away from the resize affordance 708-2 by less than a threshold amount (e.g., the user’s gaze is still within a threshold proximity of the resize affordance 708-2 and/or is directed to application window 702).
- a threshold amount e.g., the user’s gaze is still within a threshold proximity of the resize affordance 708-2 and/or is directed to application window 702
- the color of resize affordance is not maintained (e.g., as illustrated by resize affordance 708- 3, Figure 7H) as the updated color in accordance with a determination that the user’s attention 710-5 has shifted, or moved, away from the resize affordance 708-2 by at least the threshold amount (e.g., the user’s gaze is not within the threshold proximity of the resize affordance and/or is not directed to application window 702).
- Automatically updating an appearance of a control when the user has looked away from the control including changing the appearance of the control in a first manner if the user looks at a first position, such as far away from the control, and changing the appearance of the control in a second manner if the user looks at a second position, such as close to the control, provides the user with improved visual feedback of the user’s interactions to indicate where the user’s gaze is detected.
- the computer system in response to detecting the second gaze input directed to the portion of the three-dimensional environment: in accordance with a determination that the second gaze input is directed to a second location relative to the first control element in the three-dimensional environment (e.g., the second location has a second spatial relationship with the first control element (e.g., is outside a threshold distance of the first control element, is no longer on the first control element or the portion of the first user interface object that corresponds to the first control element); and/or the second location has the first spatial relationship with a location associated with a second control element (e.g., is within a threshold distance of the display position of the second control element (e.g., center of the bottom edge of the first user interface object, a lower right comer of the first user interface object, or the upper left comer of the first user interface object))), the computer system displays a second control element that is associated with performing a second operation (e.g., dragging, resizing, adjusting a shape and/or size
- a second operation
- the appearance of the first control element changes to indicate the movement of the gaze; as the gaze moves to the location associated with the second control element, the computer system ceases to display the first control element and displays the second control element or changes the appearance of the second control element to indicate that the location of the gaze input is now on the second control element.
- the appearance of the second control element changes to indicate the movement of the gaze; as the gaze moves to the location associated with the first control element, the computer system ceases to display the second control element and displays the first control element or changes the appearance of the first control element to indicate that the location of the gaze input is now on the first control element.
- the first control element is an object resize control displayed proximate to a first comer of the first user interface object
- the second control element is an object move control displayed proximate to the center of the bottom edge of the first user interface object, as described with reference to Figure 71.
- the user’s attention 710-6 has moved away from resize affordance 708-3, and, as the user’s attention 710-6 moves to a center of the bottom edge of application window 702, resize affordance 708-4 is displayed as shifting over in a same direction as if to follow the user’s attention 710-6, until grabber 706-1 replaces display of resize affordance 708-4 in Figure 7J.
- the second control element has some or all of the interactions, appearances, and properties described herein with respect to the first control element, and is an example of other types of control elements that can be concurrently displayed with the first user interface object.
- Determining a position of where the user is currently gazing, and automatically ceasing display of a first control when the user has looked away from the first control and displaying a second control based on the currently detected position of the user’s gaze reduces the number of inputs required to access various controls and provides the user with improved visual feedback of the user’s interactions to indicate where the user’s gaze is detected.
- the first view of the three-dimensional environment corresponds to a first viewpoint of the user
- displaying the first control element with the first appearance includes: in accordance with a determination that the first user interface object is displayed within the three-dimensional environment at a first position that is a first distance away from the first viewpoint of the user, displaying the first control element with a first simulated size (e.g., at or near the first distance away from the first viewpoint) that corresponds to the first distance; and in accordance with a determination that the first user interface object is displayed within the three-dimensional environment at a second position that is a second distance away from the first viewpoint of the user, displaying the first control element with a second simulated size (e.g., at or near the second distance away from the first viewpoint) that corresponds to the second distance.
- a first simulated size e.g., at or near the first distance away from the first viewpoint
- the computer system increases the simulated size (e.g., an intrinsic size, as opposed to the size on the display) of the first control element, such that the first control element still appears roughly the same displayed size (e.g., angular extent) to the user as before, rather than becoming too small to see or manipulate for the user.
- the computer system decreases the simulated size (e.g., an intrinsic size, as opposed to the size on the display) of the first control element, such that the first control element is not blown up on the display by too much to obscure other content in the first view.
- grabber 706-2 in Figure 7M is displayed with a smaller size than grabber 706-1 after application window 702 is resized to a smaller size, such that the grabber is proportional to the size of the application window.
- grabber 706-2 is displayed with a proportional size to application window 702. Displaying a control associated with a virtual object with a simulated physical size that is based on a simulated distance between the virtual object and the user within the three-dimensional environment provides the user with improved visual feedback of the user’s position within the three-dimensional environment.
- the computer system detects movement of the first user interface object (and correspondingly, detecting movement of the first control element) from the first position to a second position that is a second distance away from the first viewpoint of the user (e.g., in response to and in accordance with the additional movement associated with the first user input; or in accordance with other events occurring in the three-dimensional environment), wherein the second distance is greater than the first distance; and in response to the detecting the movement of the first user interface object from the first position to the second position that is farther away from the first viewpoint of the user than the first position (and in accordance with a determination that the first control element is moved together with the first user interface object from the first distance to the second distance away from the first viewpoint of the user), the computer system: displays the first control element at or near the second
- the increase in simulated size of the first control element counteracts and/or offsets at least a portion of the scale down of the displayed size of the first control element due to the increased display depth away from the first viewpoint, so the first control element appears to the user with roughly the same displayed size, or slightly reduced displayed size that is larger than the normal scaled-down size of the first control element due to the increase in distance from the first viewpoint.
- resize affordance 708-5 appears to change in size by an amount based on (i) being currently selected by a user input and (ii) a current size of application window 702.
- Increasing a simulated physical size of a control, while the control is selected, that is associated with a virtual object while the virtual object is displayed as decreasing in size based on an increased distance between the virtual object and the user within the three-dimensional environment provides the user with improved visual feedback of the user’s position within the three-dimensional environment.
- the first control element is displayed with a first size.
- the computer system detects movement of the first user interface object (and correspondingly, detecting movement of the first control element) from the first position to a third position that is a third distance away from the first viewpoint of the user (e.g., in response to and in accordance with the additional movement associated with the first user input; or in accordance with other events occurring in the three-dimensional environment), wherein the third distance is smaller than the first distance; and in response to the detecting the movement of the first user interface object from the first position to the third position that is closer to the first viewpoint of the user than the first position (and, optionally, in accordance with a determination that the first control element is moved together with the first user interface object from the first distance to the third distance away from the first viewpoint of the user): displays the first control element at or near the third distance away from the first viewpoint of the user, with a decreased simulated size as compared to when the first control element was displayed at or near the first distance away from the first viewpoint of the user.
- the decrease in simulated size of the first control element counteracts and/or offsets at least a portion of the scale up of the displayed size of the first control element due to the decreased display depth to the first viewpoint, so the first control element appears to the user with roughly the same displayed size, or slightly increased displayed size that is smaller than the normal scaled-up size of the first control element due to the decrease in distance from the first viewpoint.
- resize affordance 708-5 appears to change in size by an amount based on (i) being currently selected by a user input and (ii) a current size of application window 702.
- Decreasing a simulated physical size of a control associated with a virtual object while the virtual object is displayed as increasing in size based on an decreased distance between the virtual object and the user within the three-dimensional environment provides the user with improved visual feedback of the user’s position within the three-dimensional environment.
- the first view of the three-dimensional environment corresponds to a first viewpoint of the user
- displaying the first user interface object includes: in accordance with a determination that the first user interface object is displayed within the three-dimensional environment at a first position that is a first distance away from the first viewpoint of the user, displaying the first user interface object with a third simulated size that corresponds to the first distance; and in accordance with a determination that the first user interface object is displayed within the three-dimensional environment at a second position that is a second distance away from the first viewpoint of the user, displaying the first control element with a second simulated size that corresponds to the second distance.
- the computer system increases the simulated size (e.g., an intrinsic size, as opposed to the size on the display) of the first user interface object, such that the first user interface object does not becoming too small to see or manipulate for the user.
- the computer system decreases the simulated size (e.g., an intrinsic size, as opposed to the size on the display) of the first user interface object such that the first user interface object is not blown up on the display by too much to obscure too much of other content in the first view.
- application window 702 decreases in size, and if application window 702 (e.g., and its associated controls, such as grabber 706-1) is moved in position towards the user, application window (e.g., and its associated controls, such as grabber 706-1) increases in size in accordance with the closer position to the user.
- Displaying a virtual object with a simulated physical size that is based on a simulated distance between the virtual object and the user within the three-dimensional environment provides the user with improved visual feedback of the user’s position within the three-dimensional environment.
- aspects/operations of methods 800, 1000, 1100, 1200, 1300, 1400, 1500, 1600, 1800 and 20000 may be interchanged, substituted, and/or added between these methods.
- the first object (e.g., application window or virtual object) in the method 800 has characteristics of the first user interface object (e.g., application window or virtual) in the method 900, and the first control elements that are displayed may be replaced by, or concurrently displayed with, other user interface elements (e.g., the first control element, as in the method 900 above, can be displayed as a title bar, as in the method 1000 below).
- the first object e.g., application window or virtual object
- the first control elements that are displayed may be replaced by, or concurrently displayed with, other user interface elements (e.g., the first control element, as in the method 900 above, can be displayed as a title bar, as in the method 1000 below).
- FIG 10 is a flow diagram of an exemplary method 1000 for displaying a title bar with information for an application, in accordance with some embodiments.
- method 1000 is performed at a computer system (e.g., computer system 101 in Figure 1) including a display generation component (e.g., display generation component 120 in Figures 1 A, 3, and 4 or display generation component 7100) (e.g., a heads-up display, a display, a touchscreen, a projector, etc.) and one or more cameras (e.g., a camera (e.g., color sensors, infrared sensors, and other depth-sensing cameras) that points downward at a user’s hand or a camera that points forward from the user’s head).
- a computer system e.g., computer system 101 in Figure 1
- a display generation component e.g., display generation component 120 in Figures 1 A, 3, and 4 or display generation component 7100
- a heads-up display e.g., a heads-up display, a
- the method 1000 is governed by instructions that are stored in a non-transitory (or transitory) computer-readable storage medium and that are executed by one or more processors of a computer system, such as the one or more processors 202 of computer system 101 (e.g., control 110 in Figure 1 A). Some operations in method 1000 are, optionally, combined and/or the order of some operations is, optionally, changed.
- Method 1000 is a method for displaying a title bar for an application window that dynamically increases in size, in response to detecting a user’s attention is directed to the title bar, to automatically provide the user with access to additional controls, thereby reducing a number of inputs required to access additional controls for the application window and providing visual feedback about a state of the device.
- method 1000 is performed at a computer system that is in communication with a first display generation component (e.g., a first display generation component of a two-sided display generation component, a heads-up display, a head-mounted display (HMD), a display, a touchscreen, a projector, a standalone display, and/or a display that is enclosed in the same housing as another display generation component of the same type or a different type) and one or more input devices (e.g., cameras, controllers, touch- sensitive surfaces, joysticks, buttons, gloves, watches, motion sensors, and/or orientation sensors).
- the first display generation component is a display component facing the user and provides an XR experience to the user.
- the first display generation component includes two or more display components (e.g., one set for each eye) that display slightly different images to form a stereoscopic view of the three-dimensional environment.
- the first display generation component and a second display generation component form a two-sided display device (e.g., a two-sided HMD) that displays a first user interface on a first side corresponding to the first display generation component, and a second user interface on the second side corresponding to the second display generation component.
- the second display generation component is a display component facing away from the user and toward an external environment of the user and optionally provides status information related to the first display generation component (e.g., displayed content and/or operational state) and/or the user (e.g., movement of the user’s eyes, and/or attention state of the user) to other users in the external environment.
- the computing system is an integrated device with one or more processors and memory enclosed in the same housing as the first and the second display generation components and at least some of the one or more input devices.
- the computing system includes a computing component (e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer) that includes one or more processors and memory that is separate from the display generation component(s) and/or the one or more input devices.
- a computing component e.g., a server, a mobile electronic device such as a smart phone or tablet device, a wearable device such as a watch, wristband, or earphones, a desktop computer, or a laptop computer
- the display generation component s) and the one or more input devices are integrated and enclosed in the same housing.
- the computer system concurrently displays (1002), via the first display generation component, a first application window and a first title bar of the first application window, wherein the first title bar of the first application window is separate from the first application window (e.g., changes size and/or location separately from the first application window and/or is spaced apart from the first application window by a gap (e.g., content (e.g., passthrough content, and/or other objects and windows in the environment) behind the first application window and the first title bar is visible to the user through the gap)) on a first side of the first application window (e.g., displayed above the top edge, below the bottom edge, left of the left edge, right of the right edge, and/or external to a respective corner of the first application window) and displays a respective identifier of the first application window (e.g., a respective identifier (e.g., a name and/or an application icon for an application, a name and/or avatar of a user or group engaged in an ongoing communication session
- the computer system While displaying the first application window with the first title bar separate from the first application window, the computer system detects (1004), via the one or more input devices, that a user’s attention is directed to the first title bar (e.g., detecting a gaze and/or gesture input directed to the first title bar, optionally, for at least a threshold amount of time and/or with sufficient specificity with respect to the first title bar). For example, in Figure 7Z, user’s attention 710-23 is directed to title bar 716.
- the computer system expands (1008) the first title bar (e.g., in a direction away from the first application window, in a direction toward the interior of the first application window, and/or in a predetermined direction (e.g., downward, upward, to the left, to the right, or in another direction)) to display one or more first selectable controls for interacting with a first application corresponding to the first application window, wherein the one or more first selectable controls were not displayed prior to expanding the first title bar.
- first criteria e.g., the gaze input is maintained on the first title bar for at least a threshold amount of time, and/or a predetermined gesture or interaction is detected in conjunction with the gaze input directed to the first title bar
- the computer system expands (1008) the first title bar (e.g., in a direction away from the first application window, in a direction toward the interior of the first application window, and/or in a predetermined direction (e.g., downward, upward, to the left, to the right, or in another direction)) to display
- the expanded first title bar includes one or more selectable controls for interacting with the content of the first application window (e.g., performing an operation that changes the content of the first application window, updating at least a portion of the content of the first application window, changing the appearance of the content displayed in the first application window, and/or displaying different content in the first application window).
- the one or more first selectable controls include selectable controls for interacting with the first application corresponding to the first application window (e.g., opening a new window of the first application, and/or navigating to a different user interface in the first application window).
- the first title bar before the first title bar is expanded, the first title bar is separate from the first application window by a gap; and after expanding the first title bar, the expanded first title bar visually obscure a portion of the first application window (e.g., expanded first title bar is displayed in front of the first application window, overlaying or covering at least a portion of the first application window).
- the expanded first title bar expands in a direction away from the first application window and does not visually obscure the first application window. For example, in Figure 7AA, title bar 716-2 expands to display additional controls.
- the computer system displays (e.g., concurrently with the first application window and the first title bar), via the first display generation component, a second application window (e.g., of a distinct application than the first application window or of the same application as the first application window) and a second title bar of the second application window, wherein the second title bar of the second application window is separate from the second application window (e.g., changes size and/or location separately from the second application window and or is spaced apart from the second application window, e.g., in a manner analogous to that of the first application window and the first title bar) on a first side of the second application window (e.g., displayed above the top edge, below the bottom edge, left of the left edge, right of the right edge, and/or external to a respective comer of the second application window) and displays a respective identifier of the second application window (e.g., a respective identifier (e.g., a name and/or an application icon for an application
- application window 720 is also displayed with its own title bar 722.
- the computer system detects, via the one or more input devices, that the user’s attention is directed to the second title bar of the second application window (e.g., detecting a gaze and/or gesture input directed to the second title bar, optionally, for at least a threshold amount of time and/or with sufficient specificity with respect to the second title bar); and in response to detecting that the user’s attention is directed to the second title bar of the second application window: in accordance with a determination that the user’s attention meets the first criteria with respect to the second title bar of the second application window (e.g., the gaze input is maintained on the second title bar for at least a threshold amount of time, and/or a predetermined gesture or interaction is detected in conjunction with the gaze input directed to the second title bar), expands the
- the second application is distinct from the first application corresponding to the first application window.
- the second application is the same application as the first application (e.g., each application window represents a distinct tab and/or window of the same application).
- the expanded second title bar includes one or more selectable controls for interacting with the content of the second application window (e.g., performing an operation that changes the content of the second application window, updating at least a portion of the content of the second application window, changing the appearance of the content displayed in the second application window, and/or displaying different content in the second application window).
- the one or more second selectable controls include selectable controls for interacting with the second application corresponding to the second application window (e.g., opening a new window of the second application, and/or navigating to a different user interface in the second application window).
- the second title bar before the second title bar is expanded, the second title bar is separate from the second application window by a gap; and after expanding the second title bar, the expanded second title bar visually obscure a portion of the second application window (e.g., expanded second title bar is displayed in front of the second application window, overlaying or covering at least a portion of the second application window).
- the expanded second title bar expands in a direction away from the second application window and does not visually obscure the second application window.
- title bar 722-2 optionally expands to partially obscure application window 720.
- the computer system in response to detecting that the user’s attention is directed to the first title bar of the first application window: in accordance with a determination that the user’s attention does not meet the first criteria with respect to the first title bar of the first application window (e.g., the gaze input is maintained on the first title bar for less than the threshold amount of time, and/or a predetermined gesture or interaction is not detected in conjunction with the gaze input directed to the first title bar), the computer system forgoes expanding the first title bar (e.g., maintaining the first title bar at a same size without expanding the first title bar to display one or more first selectable controls). In some embodiments, if the user’s attention is not directed to the first title bar, the first title bar is not expanded.
- the computer system in response to detecting that the user’s attention is directed to the first title bar, updates one or more visual properties of the first title bar (e.g., changes a color, opacity, or other visual property of the first title bar) while detecting that the user’s attention is directed to the first title bar. For example, in Figure 7Z, in accordance with a determination that the user’s attention 710-23 is not maintained for a threshold amount of time before the user’s attention is directed elsewhere (e.g., not to title bar 716), expanded title bar 716-2 is not displayed.
- one or more visual properties of the first title bar e.g., changes a color, opacity, or other visual property of the first title bar
- expanding the first title bar to display the one or more first selectable controls includes: expanding the first title bar to overlap at least a portion of the first application window; and displaying the one or more first selectable controls over the portion of the first application window.
- the one or more first selectable controls include controls for the first application associated with the first application window.
- the one or more first selectable controls include one or more of controls for: displaying details of an account, controlling media playback, changing one or more settings, closing the current or all documents or windows of the first application, starting or stopping sharing, zooming, renaming the current document or window, exporting the current document, duplicating the current document, performing a search, printing the content of the first application window, and/or other operations of the first application or operations for controlling the first application window and/or its content).
- the sets of selectable controls displayed in the expanded title bars of different application windows and/or for different applications are the same.
- the sets of selectable controls displayed in the expanded title bars of different application windows and/or for different applications are different for the different application windows and/or applications.
- title bar 716-2 is expanded to display additional controls that overlap with application window 702. Displaying additional controls in an expanded title bar in response to detecting the user’s gaze is directed to the title bar for an application window, reduces a number of inputs required to access additional controls for the respective application window.
- the first title bar is one of a set of one or more application control affordances (e.g., title bar, privacy indicator, tab indicator, sharing indicator, media playback indicator, location indicator, and other window control affordances) that are displayed (e.g., side by side, in a column, or in a row) separate from the first application window on a respective side of the first application window (e.g., some of the set of application control affordances are displayed above a top side of the first application window, some of the set of application control affordances are displayed below a bottom side of the first application window, some of the set of application control affordances are displayed left of a left side of the first application window, and/or some of the set of application control affordances are displayed right of a right side of the first application window).
- application control affordances e.g., title bar, privacy indicator, tab indicator, sharing indicator, media playback indicator, location indicator, and other window control affordances
- the computer system while displaying the first application window with the set of one or more application affordances separate from the first application window, the computer system detects, via the one or more input devices, that a user’s attention is directed to a respective application control affordance of the set of one or more application control affordances; and in response to detecting that the user’s attention is directed to the respective application control affordance for the first application window: in accordance with a determination that the user’s attention meets the first criteria with respect to the respective application control affordance of the first application window (e.g., the gaze input is maintained on the respective application control affordance for at least a threshold amount of time, and/or a predetermined gesture or interaction is detected in conjunction with the gaze input directed to the respective application control affordance), expands the respective application control affordance to display additional information and/or controls that are not displayed prior to detecting that the user’s attention meets the first criteria with respect to the respective application control affordance; and in accordance with a determination that the user’s attention does not meet the first criteria with respect to the
- interactions and properties of the first title bar described herein are also applicable to one or more of other application control affordances of the set of application control affordances, and are not repeated in an interest of brevity.
- additional controls e.g., sharing indication 719, privacy indicator 718-1, and/or additional control option 721
- the additional control expands (e.g., privacy indicator 718-1 expands to privacy indicator 718-2, as described with reference to Figure 7V).
- the respective application control affordance includes information indicating content that is displayed in the first application window. For example, a document name is displayed in the respective application control affordance (e.g., the title bar, or another application control affordance) for a word processing and/or document viewer application to indicate the document that is displayed in the first application window.
- the respective application control affordance includes an application icon indicating the first application associated with the first application window.
- the information indicating content that is displayed in the first application window is displayed in an application control affordance separate from the title bar.
- the title bar indicates a name of the first application associated with the first application window and a separate application control affordance indicates the document name (or website name) for the content displayed within the first application window.
- title bar 716 in Figure 7S displays information about the content displayed in application window 702. Displaying a title bar that includes information about content that is currently displayed in the application window associated with the title bar enables the user to quickly view information about and navigate between application windows without requiring the user to navigate through complex menu hierarchies.
- the set of application control affordances includes a first set of one or more privacy indicators for the first application, wherein the first set of one or more privacy indicators have been displayed in accordance with a determination that the first application is accessing one or more sensors of the computer system (e.g., using, collecting, and/or accessing data of one or more cameras, microphones, geolocation sensors, biometric sensors, and/or other sensors that collect personal or privacy-related information of the user), as described in more detail with respect to method 1100.
- the computer system e.g., using, collecting, and/or accessing data of one or more cameras, microphones, geolocation sensors, biometric sensors, and/or other sensors that collect personal or privacy-related information of the user
- the computer system while displaying the first application window of the first application with the first set of one or more privacy indicators, the computer system detects that the first application associated with the first application window is no longer accessing the one or more sensors of the computer system; and in accordance with a determination that the first application is no longer accessing the one or more sensors of the computer system, ceases to display the first set of privacy indicators with the first application window.
- a respective privacy indicator is displayed with a set of visual properties (e.g., a color, an opacity, and/or a size) that are selected in accordance with the sensor that is used to collect information.
- the privacy indicator is displayed with a first set of visual properties, including a first color, in accordance with a determination that a microphone is being collected by the first application, and the privacy indicator is displayed with a second set of visual properties, including a second color distinct from the first color, in accordance with a determination that location information is being collected by the first application).
- more than one privacy indicator are, optionally, concurrently displayed at a given moment in time, each corresponding to a respective type of sensor that is being accessed by the first application, as described with reference to privacy indicator 718-1 in Figures 7T (e.g., Figures 7T1 and 7T2). Determining that an application is accessing sensitive user data and displaying a privacy indicator near an application window of the application that is accessing sensitive user data improves security and privacy of the system by providing real-time information about the specific application window that is accessing the sensitive user data.
- the first set of one or more privacy indicators includes a first privacy indicator that has been displayed in accordance with a determination that the first application is collecting audio information through at least one of the one or more sensors of the computer system (e.g., using, collecting, and/or accessing data of one or more cameras, and/or microphones).
- the first set of one or more privacy indicators includes a single indicator and it is displayed with a first appearance that corresponds to the collection of audio information.
- the first set of one or more privacy indicators include multiple indicators, the first privacy indicator that corresponds to the collection of audio information is one of the first set of one or more privacy indicators, and at least one other privacy indicator of the first set of privacy indicators corresponds to the collection of other types of privacy information (e.g., visual information, location information, and/or biometric information).
- privacy indicator 718-1 is displayed with a first color indicating that audio information is being collected. Determining that an application is accessing a microphone to collect audio information, and displaying a privacy indicator that indicates audio information is being collected near an application window of the application that is accessing the audio information improves security and privacy of the system by providing real-time information about the specific application window that is accessing the microphone.
- the first set of one or more privacy indicators includes a second privacy indicator that has been displayed in accordance with a determination that the first application is collecting location information through at least one of the one or more sensors of the computer system (e.g., using, collecting, and/or accessing data of one or more GPS devices, WiFi devices, gyros, telephony devices, antennas, cellular phone component, and/or other location-determination devices of the computer system).
- the first set of one or more privacy indicators includes a single indicator and it is displayed with a second appearance that corresponds to the collection of location information (e.g., different from the appearance that corresponds to collection of audio information and from the appearance that corresponds to collection of video information).
- the first set of one or more privacy indicators include multiple indicators, the second privacy indicator that corresponds to the collection of location information is one of the first set of one or more privacy indicators, and at least one other privacy indicator of the first set of privacy indicators corresponds to the collection of other types of privacy information (e.g., audio information, visual information, and/or biometric information).
- privacy indicator 718-1 is displayed with a different color if location information is being accessed (e.g., as compared to a color representing if other types of information are being accessed), as described with reference to Figures 7T.
- Determining that an application is accessing location data of the computer system, and displaying a privacy indicator that indicates location data is being collected near an application window of the application that is accessing the location data improves security and privacy of the system by providing real-time information about the specific application window that is accessing the location data.
- the set of application control affordances includes a first sharing indicator for the first application, wherein the first sharing indicator has been displayed with a first appearance in accordance with a determination that the first application is sharing content with another device (e.g., the content of the first application window is shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is sharing content with another user and/or device and the first application window is the active window of the first application).
- another device e.g., the content of the first application window is shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is sharing content with another user and/or device and the first application window is the active window of the first application.
- the computer system while concurrently displaying the first application window and the first sharing indicator with the first appearance, the computer system detects that the first application is no longer sharing content with another device (e.g., the user has stopped sharing the first application window or the first application); and in accordance with a determination that the first application is no longer sharing content with another device, ceases to display the first sharing indicator with the first appearance (e.g., ceasing to display the first sharing indicator completely, or changing the appearance of the first sharing indicator from the first appearance to a second appearance different from the first appearance, to indicate that the first application is now private, no longer sharing the first application window, and/or no longer sharing content with another user).
- the first sharing indicator e.g., ceasing to display the first sharing indicator completely, or changing the appearance of the first sharing indicator from the first appearance to a second appearance different from the first appearance, to indicate that the first application is now private, no longer sharing the first application window, and/or no longer sharing content with another user.
- the first sharing indicator optionally includes identifiers (e.g., avatar(s), name(s), initial(s) and/or picture(s)) associated with devices with which the first application window or content of the first application is being shared.
- the computer system detects that the first application window and/or content of the first application is not being shared with another device, and in response to detecting that the first application window and/or content of the first application is not being shared with another device, the computer system displays the first sharing indicator with a second appearance (e.g., a dimmed user and/or device indication that indicates the first application window and/or the first application is only displayed for the user without being shared with another device).
- a second appearance e.g., a dimmed user and/or device indication that indicates the first application window and/or the first application is only displayed for the user without being shared with another device.
- sharing indication 719 displays information about the users that have access to shared application window 702, and sharing indication 719 is updated to indicate that application window 702 is private in accordance with a determination that application window 702 is not part of a shared session. Determining whether an application window is being shared across multiple devices, and displaying a sharing indicator that indicates content of the application window is currently being shared across devices associated with other users improves security and privacy of the system by providing real-time information about the specific application window that is being shared and provides improved visual feedback to the user.
- displaying the first sharing indicator includes displaying the first sharing indicator proximate to (and optionally separate from and/or adjacent to) the first title bar on the first side of the first application window.
- the computer system concurrently displays the first application window, the first title bar, the first set of one or more privacy indicators, and the first sharing indicator for the first application window, wherein the first sharing indicator is displayed proximate to the first privacy indicator and/or proximate to the first title bar.
- sharing indicator 719 is a separate user interface element displayed at a nonzero distance away from application window 702 in Figure 7Z.
- Displaying a sharing indicator that indicates content of an application window is currently being shared across devices associated with other users near the title bar that indicates content of the application window improves security and privacy of the system by providing real-time information about the application window that is being shared and provides improved visual feedback to the user.
- the computer system detects (e.g., while the first sharing indicator is displayed with a second appearance to indicate that the first application is not currently sharing content with another device, or while the first sharing indicator is not displayed) that the first application is sharing content with another device (e.g., the content of the first application window is shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is sharing content with another user and/or device and the first application window is the active window of the first application).
- the first application is sharing content with another device
- the content of the first application window is shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is sharing content with another user and/or device and the first application window is the active window of the first application.
- the computer system in response to detecting that the first application sharing content with another device, the computer system: in accordance with a determination that said another device is associated with a first other user, displays, in the first sharing indicator, a first identifier corresponding to the first other user (e.g., the name and/or avatar, username, telephone number, ID number, and/or other identifiers of the first other user); and in accordance with a determination that said other device is associated with a second other user different from the first other user, displays, in the first sharing indicator, a second identifier corresponding to the second other user (e.g., the name and/or avatar, username, telephone number, ID number, and/or other identifiers of the second other user), the second identifier different from the first identifier.
- a first identifier corresponding to the first other user e.g., the name and/or avatar, username, telephone number, ID number, and/or other identifiers of the first other user
- sharing indicator 719 includes an indication of the user that has access to shared application window 702. Determining whether an application window is being shared across multiple devices, and displaying a sharing indicator that indicates the user of another device that is sharing the content of the application window with the computer system improves security and privacy of the system by providing real-time information about the specific application window that is being shared and provides improved visual feedback to the user about who is sharing the application window without requiring additional user input.
- the computer system detects (e.g., while the first sharing indicator is displayed with the first appearance to indicate that the first application is currently sharing content with another device) that the first application is no longer sharing content with another device (e.g., the content of the first application window is no longer shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is no longer sharing content with another user and/or device).
- the computer system in response to detecting that the first application is no longer sharing content with another device, displaying the first sharing indicator with a second appearance that is different from the first appearance to indicate that the first application is not sharing content with another device (e.g., the first application window and any other windows of the first application are private to the user).
- the computer system displays a lock symbol indicating the first application window and/or the content of the first application are private and not being shared by, or with, other devices. For example, sharing indication 719 in Figure 7Z displays an indication that application window 702 is private.
- Determining whether an application window is being shared across multiple devices, and displaying a private indicator that indicates the content of the application window is not currently being shared with other devices improves security and privacy of the system by providing real-time information about the sharing status of the application window and provides improved visual feedback to the user indicating the application window is private without requiring additional user input.
- the computer system detects (e.g., while the first sharing indicator is displayed with the second appearance to indicate that the first application is not currently sharing content with another device) that the first application is sharing content with another device (e.g., the content of the first application window is being shared with another user or made public via the first application and/or via another application, such as a videoconferencing or other sharing application; or the first application is sharing content with another user and/or device).
- the computer system in response to detecting that the first application is sharing content with another device, the computer system displays the first sharing indicator with the first appearance to indicate that the first application is sharing content with another device.
- the computer system displays an indication that the computer system is sharing the first application window, and/or displays indications of the other users (e.g., avatars, initials, and/or other identifiers) that are accessing (or provided with access to) the shared first application window.
- the sharing indication is updated in real-time to indicate which users are currently participating in the shared session of the first application window. For example, as described with reference to Figure 7Z, sharing indication 719 includes avatars or other user identifiers indicating that users that have access to the shared application window 702.
- Determining whether an application window is being shared across multiple devices, and displaying a sharing indicator that indicates the users of the devices that are viewing the shared content of the application window improves security and privacy of the system by providing real-time information about the who is accessing content shared by the computer system and provides improved visual feedback to the user about the other users of the devices accessing the application window without requiring user input.
- the first application window is displayed as a two- dimensional object in a three-dimensional environment (e.g., a passthrough view of the physical environment, an AR environment, an XR environment, or a virtual three- dimensional environment), and a respective application control affordance of the set of application control affordances associated with the first application (e.g., the first title bar, the one or more privacy indicators, the sharing indicator, or other application control affordances) is displayed on a first side of the two-dimensional object (e.g., above the top edge of the two-dimensional application window, or left of the left edge of the two- dimensional application window).
- a respective application control affordance of the set of application control affordances associated with the first application e.g., the first title bar, the one or more privacy indicators, the sharing indicator, or other application control affordances
- the computer system displays, via the first display generation component, a first three-dimensional object (e.g., a three- dimensional model of a physical object, a viewport into a three-dimensional XR experience, a three-dimensional user interface object, or other objects associated with the first application or another application or environment) and a second set of application control affordances that is associated with the first three-dimensional object, in the three-dimensional environment, wherein: the first three-dimensional object is associated with a third application (e.g., same as the first application, or different from the first application), the second set of application control affordances (e.g., including a third title bar and/or a set of other application control affordances described herein with respect to the first application window and the second application window) that is associated with the first three-dimensional object is separate from the first three-dimensional object on a second side (e.g., bottom side, or right side) of the first three-dimensional object (e.g., the third title bar displays a respective identifie
- the respective side from which the title bar is offset away from the application object is determined based at least in part on whether the application object is displayed as a two-dimensional or a three-dimensional application object.
- title bar 716 and/or other control affordances e.g., close affordance 7030, privacy indicator 718-1, or another control
- first respective positions relative to application window 702 which is a two-dimensional object
- virtual object 7028, a title bar, and/or other control affordance are displayed at different respective positions relative to virtual object 7028 (e.g., close affordance 717 is optionally displayed below virtual object 7028).
- Displaying a title bar of a two-dimensional object, such as an application window, on a first side of the two-dimensional object and displaying a title bar for a three-dimensional object on a different side of the three-dimensional object provides improved visual feedback to the user that differentiates between two-dimensional and three- dimensional objects.
- the computer system concurrently displays, via the first display generation component, the first application window and a third application control affordance (and optionally concurrently displaying one or more other application control affordances (e.g., the first title bar, the first set of privacy indicators, and/or the first sharing indicator)) associated with the first application window, wherein the third application control affordance is separate from the first application window and on a third side of the first application window (e.g., the third side is the same as the first side, or different from the first side) (e.g., the third application control affordance is an affordance for opening a new tab of the first application window).
- a third application control affordance e.g., the first title bar, the first set of privacy indicators, and/or the first sharing indicator
- the third application control affordance is a distinct control element that is displayed next to (e.g., to the left and/or right of) the first title bar of the first application window.
- the computer system detects, via the one or more input devices, a user input (e.g., a user’s gaze directed to the third application control affordance, an air tap or air pinch gesture directed to the third application control affordance, and/or a selection input provided by a touch gesture or cursor click) directed to the third application control affordance; and in response to detecting the user input directed to the third application control affordance associated with the first application window, in accordance with a determination that the user input meets selection criteria with respect to the third application control affordance, displays a third application window (e.g., a new window of the first application, or a new window of a selection user interface for selecting an application, or a new window of a different application) (e.g., at a location
- the third application window includes content of the first application window and additional new window content in separate display areas (e.g., tabbed pages, or tabbed display areas) that are selectively displayed in the third application window by selection of their respective tabs or title bars.
- the first title bar of the first application window is displayed adjacent to the title bar of the third application window, and selection of the first title bar causes the first application window to be redisplayed and causing the third window to be pushed behind the first application window.
- the title bars of the hidden window(s) are displayed with a smaller size than the title bar of the top layer window.
- control option 721 is a control for adding a new tab and/or window, and in response to detecting a user input selecting control option 721, a new application window is displayed and/or new content is displayed in application window 702. Displaying a control for opening an additional tab for an application window, reduces the number of inputs required to open a new tab by enabling the user to select the control for opening the additional tab without requiring the user to navigate through complex menu hierarchies.
- the computer system in response to detecting that the user’s attention is directed to the first title bar and in accordance with the determination that the user’s attention meets the first criteria with respect to the first title bar, expands the first title bar to display an application icon of the first application associated with the first application window (e.g., wherein the application icon was not displayed prior to expanding the first title bar).
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Computer Hardware Design (AREA)
- User Interface Of Digital Computer (AREA)
Abstract
Description
Claims
Priority Applications (5)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| KR1020257008502A KR20250048786A (en) | 2022-09-23 | 2023-09-22 | Devices, methods and graphical user interfaces for interacting with window controls in three-dimensional environments |
| EP23797920.8A EP4540684A1 (en) | 2022-09-23 | 2023-09-22 | Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments |
| CN202510949133.1A CN120803316A (en) | 2022-09-23 | 2023-09-22 | Apparatus, method, and graphical user interface for interacting with window controls in a three-dimensional environment |
| CN202380068172.XA CN119998764A (en) | 2022-09-23 | 2023-09-22 | Device, method and graphical user interface for interacting with window controls in a three-dimensional environment |
| JP2025515861A JP2025533474A (en) | 2022-09-23 | 2023-09-22 | DEVICE, METHOD, AND GRAPHICAL USER INTERFACE FOR INTERACTING WITH WINDOW CONTROLS IN A THREE-DIMENSIONAL ENVIRONMENT - Patent application |
Applications Claiming Priority (16)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US202263409600P | 2022-09-23 | 2022-09-23 | |
| US63/409,600 | 2022-09-23 | ||
| US202363470909P | 2023-06-04 | 2023-06-04 | |
| US63/470,909 | 2023-06-04 | ||
| US202363535012P | 2023-08-28 | 2023-08-28 | |
| US63/535,012 | 2023-08-28 | ||
| US202363539566P | 2023-09-20 | 2023-09-20 | |
| US63/539,566 | 2023-09-20 | ||
| US18/371,368 | 2023-09-21 | ||
| US18/371,378 | 2023-09-21 | ||
| US18/371,372 | 2023-09-21 | ||
| US18/371,368 US20240152245A1 (en) | 2022-09-23 | 2023-09-21 | Devices, Methods, and Graphical User Interfaces for Interacting with Window Controls in Three-Dimensional Environments |
| US18/371,374 | 2023-09-21 | ||
| US18/371,372 US20240103681A1 (en) | 2022-09-23 | 2023-09-21 | Devices, Methods, and Graphical User Interfaces for Interacting with Window Controls in Three-Dimensional Environments |
| US18/371,374 US20240104873A1 (en) | 2022-09-23 | 2023-09-21 | Devices, Methods, and Graphical User Interfaces for Interacting with Window Controls in Three-Dimensional Environments |
| US18/371,378 US12449948B2 (en) | 2022-09-23 | 2023-09-21 | Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments |
Publications (2)
| Publication Number | Publication Date |
|---|---|
| WO2024064373A1 true WO2024064373A1 (en) | 2024-03-28 |
| WO2024064373A8 WO2024064373A8 (en) | 2025-02-13 |
Family
ID=88584945
Family Applications (1)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| PCT/US2023/033530 Ceased WO2024064373A1 (en) | 2022-09-23 | 2023-09-22 | Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments |
Country Status (5)
| Country | Link |
|---|---|
| EP (1) | EP4540684A1 (en) |
| JP (1) | JP2025533474A (en) |
| KR (1) | KR20250048786A (en) |
| CN (2) | CN119998764A (en) |
| WO (1) | WO2024064373A1 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| CN119272984A (en) * | 2024-09-18 | 2025-01-07 | 北京联物瑞达信息技术有限公司 | Computer room energy consumption optimization method and system based on temperature monitoring |
| EP4660755A1 (en) * | 2024-06-09 | 2025-12-10 | Apple Inc. | Techniques for rendering adjacent ornaments |
Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150130740A1 (en) * | 2012-01-04 | 2015-05-14 | Tobii Technology Ab | System for gaze interaction |
| US20150309570A1 (en) * | 2009-04-09 | 2015-10-29 | Dynavox Systems Llc | Eye tracking systems and methods with efficient text entry input features |
| US20210191600A1 (en) * | 2019-12-23 | 2021-06-24 | Apple Inc. | Devices, Methods, and Graphical User Interfaces for Displaying Applications in Three-Dimensional Environments |
| US20220214743A1 (en) * | 2021-01-04 | 2022-07-07 | Apple Inc. | Devices, Methods, and Graphical User Interfaces for Interacting with Three-Dimensional Environments |
-
2023
- 2023-09-22 KR KR1020257008502A patent/KR20250048786A/en active Pending
- 2023-09-22 JP JP2025515861A patent/JP2025533474A/en active Pending
- 2023-09-22 WO PCT/US2023/033530 patent/WO2024064373A1/en not_active Ceased
- 2023-09-22 EP EP23797920.8A patent/EP4540684A1/en active Pending
- 2023-09-22 CN CN202380068172.XA patent/CN119998764A/en active Pending
- 2023-09-22 CN CN202510949133.1A patent/CN120803316A/en active Pending
Patent Citations (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20150309570A1 (en) * | 2009-04-09 | 2015-10-29 | Dynavox Systems Llc | Eye tracking systems and methods with efficient text entry input features |
| US20150130740A1 (en) * | 2012-01-04 | 2015-05-14 | Tobii Technology Ab | System for gaze interaction |
| US20210191600A1 (en) * | 2019-12-23 | 2021-06-24 | Apple Inc. | Devices, Methods, and Graphical User Interfaces for Displaying Applications in Three-Dimensional Environments |
| US20220214743A1 (en) * | 2021-01-04 | 2022-07-07 | Apple Inc. | Devices, Methods, and Graphical User Interfaces for Interacting with Three-Dimensional Environments |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| EP4660755A1 (en) * | 2024-06-09 | 2025-12-10 | Apple Inc. | Techniques for rendering adjacent ornaments |
| CN119272984A (en) * | 2024-09-18 | 2025-01-07 | 北京联物瑞达信息技术有限公司 | Computer room energy consumption optimization method and system based on temperature monitoring |
Also Published As
| Publication number | Publication date |
|---|---|
| CN119998764A (en) | 2025-05-13 |
| EP4540684A1 (en) | 2025-04-23 |
| WO2024064373A8 (en) | 2025-02-13 |
| CN120803316A (en) | 2025-10-17 |
| JP2025533474A (en) | 2025-10-07 |
| KR20250048786A (en) | 2025-04-10 |
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US12449948B2 (en) | Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments | |
| US20240152245A1 (en) | Devices, Methods, and Graphical User Interfaces for Interacting with Window Controls in Three-Dimensional Environments | |
| US20230336865A1 (en) | Device, methods, and graphical user interfaces for capturing and displaying media | |
| US20240152256A1 (en) | Devices, Methods, and Graphical User Interfaces for Tabbed Browsing in Three-Dimensional Environments | |
| US20240103716A1 (en) | Methods for interacting with user interfaces based on attention | |
| EP4591140A1 (en) | Devices, methods, and graphical user interfaces for interacting with extended reality experiences | |
| US12182391B2 (en) | Devices, methods, and graphical user interfaces for interacting with three-dimensional environments | |
| WO2024253979A1 (en) | Methods for moving objects in a three-dimensional environment | |
| WO2024064373A1 (en) | Devices, methods, and graphical user interfaces for interacting with window controls in three-dimensional environments | |
| US20240428539A1 (en) | Devices, Methods, and Graphical User Interfaces for Selectively Accessing System Functions and Adjusting Settings of Computer Systems While Interacting with Three-Dimensional Environments | |
| US20240385858A1 (en) | Methods for displaying mixed reality content in a three-dimensional environment | |
| US20250298470A1 (en) | Devices, Methods, and Graphical User Interfaces for Navigating User Interfaces within Three-Dimensional Environments | |
| US20250355485A1 (en) | Devices, Methods, And Graphical User Interfaces For Interacting With System User Interfaces Within Three-Dimensional Environments | |
| US20250355555A1 (en) | Devices, Methods, and Graphical User Interfaces for Interacting with System User Interfaces within Three-Dimensional Environments | |
| US20250377758A1 (en) | Devices, Methods, And Graphical User Interfaces For Interacting With Volumetric Application User Interfaces Within Three-Dimensional Environments | |
| US20250378629A1 (en) | Methods of interacting with content in a virtual environment | |
| WO2025198713A1 (en) | Devices, methods, and graphical user interfaces for navigating user interfaces within three-dimensional environments | |
| WO2025240643A1 (en) | Devices, methods, and graphical user interfaces for interacting with system user interfaces within three-dimensional environments | |
| WO2025255162A1 (en) | Devices, methods, and graphical user interfaces for interacting with volumetric application user interfaces within three-dimensional environments | |
| WO2025259436A1 (en) | Methods of facilitating multiview display of content items in a three-dimensional environment | |
| WO2025259409A1 (en) | Methods of interacting with content in a virtual environment | |
| WO2024026024A1 (en) | Devices and methods for processing inputs to a three-dimensional environment | |
| WO2024064231A1 (en) | Devices, methods, and graphical user interfaces for interacting with three-dimensional environments | |
| WO2024064278A1 (en) | Devices, methods, and graphical user interfaces for interacting with extended reality experiences | |
| EP4591143A1 (en) | Devices, methods, and graphical user interfaces for interacting with three-dimensional environments |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| 121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 23797920 Country of ref document: EP Kind code of ref document: A1 |
|
| DPE1 | Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101) | ||
| WWE | Wipo information: entry into national phase |
Ref document number: 2023797920 Country of ref document: EP |
|
| ENP | Entry into the national phase |
Ref document number: 2023797920 Country of ref document: EP Effective date: 20250115 |
|
| ENP | Entry into the national phase |
Ref document number: 20257008502 Country of ref document: KR Kind code of ref document: A Ref document number: 2025515861 Country of ref document: JP Kind code of ref document: A |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 1020257008502 Country of ref document: KR Ref document number: 2025515861 Country of ref document: JP |
|
| WWE | Wipo information: entry into national phase |
Ref document number: 202380068172.X Country of ref document: CN |
|
| WWP | Wipo information: published in national office |
Ref document number: 1020257008502 Country of ref document: KR |
|
| WWP | Wipo information: published in national office |
Ref document number: 2023797920 Country of ref document: EP |
|
| NENP | Non-entry into the national phase |
Ref country code: DE |
|
| WWP | Wipo information: published in national office |
Ref document number: 202380068172.X Country of ref document: CN |