US20150095846A1 - Pan and selection gesture detection - Google Patents
Pan and selection gesture detection Download PDFInfo
- Publication number
- US20150095846A1 US20150095846A1 US14/231,917 US201414231917A US2015095846A1 US 20150095846 A1 US20150095846 A1 US 20150095846A1 US 201414231917 A US201414231917 A US 201414231917A US 2015095846 A1 US2015095846 A1 US 2015095846A1
- Authority
- US
- United States
- Prior art keywords
- canvas
- pan
- gesture
- selecting control
- dimension
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/332—Query formulation
- G06F16/3329—Natural language query formulation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/24—Querying
- G06F16/242—Query formulation
- G06F16/2428—Query predicate definition using graphical user interfaces, including menus and forms
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/017—Gesture based interaction, e.g. based on a set of recognized hand gestures
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04817—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance using icons
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/0482—Interaction with lists of selectable items, e.g. menus
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/04842—Selection of displayed objects or displayed text elements
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0484—Interaction techniques based on graphical user interfaces [GUI] for the control of specific functions or operations, e.g. selecting or manipulating an object, an image or a displayed text element, setting a parameter value or selecting a range
- G06F3/0485—Scrolling or panning
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/10—Text processing
- G06F40/103—Formatting, i.e. changing of presentation of documents
- G06F40/117—Tagging; Marking up; Designating a block; Setting of attributes
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F8/00—Arrangements for software engineering
- G06F8/30—Creation or generation of source code
- G06F8/34—Graphical or visual programming
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/451—Execution arrangements for user interfaces
- G06F9/452—Remote windowing, e.g. X-Window System, desktop virtualisation
-
- G—PHYSICS
- G06—COMPUTING OR CALCULATING; COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/46—Multiprogramming arrangements
- G06F9/48—Program initiating; Program switching, e.g. by interrupt
- G06F9/4806—Task transfer initiation or dispatching
- G06F9/4843—Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system
- G06F9/485—Task life-cycle, e.g. stopping, restarting, resuming execution
- G06F9/4856—Task life-cycle, e.g. stopping, restarting, resuming execution resumption being on a different machine, e.g. task migration, virtual machine migration
Definitions
- a current paradigm for navigating through various information contexts is windows based.
- a classic example of this is the web browser experience.
- a user might begin with a home page that occupies the entire browser space. The user might then select a hyperlink, whereupon a new window appears. However, the previous window either disappears or, in the case of exercising an option to open the new page in a new window, the previous window is fully, or at least partially, hidden.
- a pan detection mechanism is configured to distinguish between an intent to pan the canvas and an intent to select the particular selectable element.
- the pan detection mechanism is configured to distinguish between an intent to pan the canvas along its extendible direction (such as horizontally), and an intent to scroll a user interface element perpendicular (such as vertically) to the extendible direction of the canvas.
- the pan detection mechanism is configured to distinguish between all three intents; that is 1) an intent to pan the canvas in its extendible direction, 2) an intent to scroll a user interface element perpendicular to the extendible direction of the canvas, and 3) an intent to select a user interface element.
- FIG. 1 abstractly illustrates an example computing system in which the principles described herein may be employed
- FIG. 2 abstractly illustrates an environment in which a canvas which has several selectable user interface elements and is pannable in the extendible direction (e.g., horizontal) of the canvas, but several selectable elements are pannable (or scrollable) also in the direction orthogonal (e.g., vertical to the extendible direction) to the extendible direction of the canvas;
- a canvas which has several selectable user interface elements and is pannable in the extendible direction (e.g., horizontal) of the canvas, but several selectable elements are pannable (or scrollable) also in the direction orthogonal (e.g., vertical to the extendible direction) to the extendible direction of the canvas;
- FIG. 3 illustrates a supporting architecture for the mechanism for distinguishing between a panning and selection gestures in accordance with the principles described herein;
- FIG. 4 illustrates a much more detailed user interface that includes a canvas populated by a large number of selectable elements
- FIG. 5 illustrates a user interface that is similar to the user interface of FIG. 4 , except that the arc associated with pointer movement is illustrated;
- FIG. 6 illustrates a user interface in which an example pan start region is illustrated with respect to the pan-start position
- FIG. 7 illustrates a flowchart of a method for permitting user interaction with a canvas that extends in an extendable dimension on the user interface, and which is be populated with selectable elements.
- At least some embodiments described herein relate to a user interface that has a canvas that is extendable in an extendable dimension and on which multiple selectable elements may be placed.
- a pan detection mechanism is configured to distinguish between an intent to pan the canvas and an intent to select the particular selectable element.
- the pen detection mechanism may be distinguish between an intent to pan the canvas and an intent to scroll the orthogonally scrollable element.
- Computing systems are now increasingly taking a wide variety of forms. Computing systems may, for example, be handheld devices, appliances, laptop computers, desktop computers, mainframes, distributed computing systems, or even devices that have not conventionally been considered a computing system.
- the term “computing system” is defined broadly as including any device or system (or combination thereof) that includes at least one physical and tangible processor, and a physical and tangible memory capable of having thereon computer-executable instructions that may be executed by the processor.
- the memory may take any form and may depend on the nature and form of the computing system.
- a computing system may be distributed over a network environment and may include multiple constituent computing systems.
- a computing system 100 typically includes at least one processing unit 102 and memory 104 .
- the memory 104 may be physical system memory, which may be volatile, non-volatile, or some combination of the two.
- the term “memory” may also be used herein to refer to non-volatile mass storage such as physical storage media. If the computing system is distributed, the processing, memory and/or storage capability may be distributed as well.
- the term “executable module” or “executable component” can refer to software objects, routines, or methods that may be executed on the computing system. The different components, modules, engines, and services described herein may be implemented as objects or processes that execute on the computing system (e.g., as separate threads).
- embodiments are described with reference to acts that are performed by one or more computing systems. If such acts are implemented in software, one or more processors of the associated computing system that performs the act direct the operation of the computing system in response to having executed computer-executable instructions.
- such computer-executable instructions may be embodied on one or more computer-readable media that form a computer program product.
- An example of such an operation involves the manipulation of data.
- the computer-executable instructions (and the manipulated data) may be stored in the memory 104 of the computing system 100 .
- Computing system 100 may also contain communication channels 108 that allow the computing system 100 to communicate with other message processors over, for example, network 110 .
- the computing system 100 also includes a display 112 on which a user interface, such as the user interfaces described herein, may be rendered.
- a user interface such as the user interfaces described herein
- Such user interfaces may be generated in computer hardware or other computer-represented form prior to rendering.
- the presentation and/or rendering of such user interfaces may be performed by the computing system 100 by having the processing unit(s) 102 execute one or more computer-executable instructions that are embodied on one or more computer-readable media.
- Such computer-readable media may form all or a part of a computer program product.
- Embodiments described herein may comprise or utilize a special purpose or general-purpose computer including computer hardware, such as, for example, one or more processors and system memory, as discussed in greater detail below.
- Embodiments described herein also include physical and other computer-readable media for carrying or storing computer-executable instructions and/or data structures.
- Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer system.
- Computer-readable media that store computer-executable instructions are physical storage media.
- Computer-readable media that carry computer-executable instructions are transmission media.
- embodiments of the invention can comprise at least two distinctly different kinds of computer-readable media: computer storage media and transmission media.
- Computer storage media includes RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other tangible medium which can be used to store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
- a “network” is defined as one or more data links that enable the transport of electronic data between computer systems and/or modules and/or other electronic devices.
- a network or another communications connection can include a network and/or data links which can be used to carry or desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer. Combinations of the above should also be included within the scope of computer-readable media.
- program code means in the form of computer-executable instructions or data structures can be transferred automatically from transmission media to computer storage media (or vice versa).
- computer-executable instructions or data structures received over a network or data link can be buffered in RAM within a network interface module (e.g., a “NIC”), and then eventually transferred to computer system RAM and/or to less volatile computer storage media at a computer system.
- a network interface module e.g., a “NIC”
- NIC network interface module
- computer storage media can be included in computer system components that also (or even primarily) utilize transmission media.
- Computer-executable instructions comprise, for example, instructions and data which, when executed at a processor, cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions.
- the computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, or even source code.
- the invention may be practiced in network computing environments with many types of computer system configurations, including, personal computers, desktop computers, laptop computers, message processors, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, mobile telephones, PDAs, pagers, routers, switches, and the like.
- the invention may also be practiced in distributed system environments where local and remote computer systems, which are linked (either by hardwired data links, wireless data links, or by a combination of hardwired and wireless data links) through a network, both perform tasks.
- program modules may be located in both local and remote memory storage devices.
- FIG. 2 abstractly illustrates an environment 200 in which a portion 201 A of a canvas 201 is displayed on a displayed area 210 of a display, and a portion 201 B of the canvas 201 is not displayed in the displayed area 210 of the display.
- the display might be, for instance, the display 112 of the computing system 100 of FIG. 1 .
- the display 200 might alternatively be a window within a display 112 , so as not to occupy the entire area of the display 112 .
- the canvas 201 extends in an extendable direction 221 along a single extendable dimension 222 . For instance, in FIG. 2 , the extendable direction 221 is rightward, and the extendable dimension 222 is the horizontal dimension.
- the extendable dimension 222 of the canvas 201 may be configurable to be vertical or horizontal according to the tastes of the user. In other cases, the limitations of the display 210 or container might restrict extendable dimension 222 to the horizontal dimension, or to the vertical dimension.
- the extendable direction 221 may be configurable. For instance, if the extendable dimension 222 is vertical, the user might configure the canvas to extend downward, or upward, according to the user's preference. If the extendable dimension 222 is horizontal (as is the case of FIG. 2 ), the user might configure the canvas to extend leftward, or rightward (as is the case of FIG. 2 ), according to the user's preference.
- the canvas 201 is illustrated as including multiple selectable elements 202 .
- the selectable elements 202 in this particular case are illustrated as including selectable elements 202 A through 202 F.
- Some of the selectable elements 202 A and 202 B are within the displayed area 201 A of the canvas 201 , and some of the selectable elements 202 D through 202 F are in the non-displayed area 201 B of the canvas 201 .
- One of the selectable elements 202 C is illustrated as spanning a boundary between the displayed portion 201 A and the undisplayed portion 201 B of the canvas 201 .
- the portion of the canvas that contains elements will be referred to as the “populated portion” of the canvas 201 .
- the selectable element might also include navigation controls.
- the user interface may have a panning mechanism that permits the user to pan the canvas 201 along the extensible dimension.
- Two of the selectable elements 202 B and 202 E are illustrated as having a greater extent in a dimension orthogonal to the extendable direction (hereinafter an “orthogonal dimension”).
- the orthogonal dimension is the vertical dimension represented by bi-directional arrow 223 .
- a pan detection mechanism distinguishes between a canvas pan gesture and an element selection gesture through subsequent interaction with the selectable element. For instance, in the canvas 201 , perhaps the user may have the experience of being able to pan the canvas by dragging the canvas 201 along the extendable dimension 222 . Yet, there are selectable elements 202 throughout the populated portion of the canvas 201 .
- the pan detection mechanism allows the user to pan the canvas 201 by activating the selecting control on any portion of the canvas 201 even if that portion happens to be occupied by a selectable element.
- the user need not be concerned about avoiding selection of elements. Instead, the user may provide a natural panning gesture on any portion of the canvas.
- the user may provide a selection gesture while the selecting control is associated with that particular selectable element.
- the pan detection mechanism distinguishes between a canvas pan gesture and an element scroll gesture. For instance, in the canvas 201 , perhaps the user may have the experience of being able to pan the canvas by dragging the canvas 201 along the extendable dimension 222 . Yet, there are orthogonally scrollable elements 202 B and 202 E throughout the populated portion of the canvas 201 . The pan detection mechanism distinguishes between an intent to pan the canvas 201 in the extendible dimension, and an intent to scroll the orthogonally scrollable element 202 B or 202 E associated with the selecting control.
- a selecting control such as a mouse cursor
- the panning control distinguishes between intents to pan the canvas in the extendible dimension, scroll the orthogonally scrollable element in the orthogonal dimension, or select the selectable element.
- FIG. 3 illustrates a supporting architecture 300 for a user interface 301 .
- the user interface 301 is an example of the user interface 200 of FIG. 2 .
- the supporting architecture 300 is illustrated as including a selecting control 311 that is capable of selecting any of the selectable elements within the canvas.
- a pan detection mechanism 312 may be used to distinguish between a pan gesture, an orthogonally scrollable element scroll gesture, and an element selection gesture.
- An activation module 321 may be used to respond to the selection gesture by selecting the appropriate selectable element.
- a pan module 322 may be used to pan the canvas in response to the pan detection mechanism 312 detecting a pan gesture and/or scroll the orthogonally scrollable element in response to the orthogonally scrollable element scroll gesture.
- the user may provide a canvas pan gesture if the user desires to pan the canvas 201 along the extendable dimension 222 , or the user may provide an element selection gesture if the user desires to select the selectable element 202 A.
- the user may additionally provide an orthogonal pan gesture (also called herein an orthogonal scroll gesture) if the user desires to pan the selectable element 202 B in the orthogonal direction.
- an orthogonal pan gesture also called herein an orthogonal scroll gesture
- This option is in addition to the possibility of a canvas pan gesture if the user desires to pan the canvas 201 along the extendable dimension 222 , and an element selection gesture if the user desires to select the selectable element 202 B.
- the pan detection mechanism 312 may be used to distinguish between the element selection gesture, the canvas pan gesture, and the orthogonal pan gesture when the selecting control 311 is associated with a selectable element. If the pan detection mechanism 312 detects an element selection gesture, the activation module 321 may be used to respond to the selection gesture by selecting the appropriate selectable element. If the pan detection mechanism 312 detects a canvas pan gesture, the pan module 322 may be used to pan the canvas in the extendable dimension. If the pan detection mechanism 312 detects an orthogonal pan gesture, the pan module 322 may be used to pan the selectable element in the orthogonal dimension.
- the pan detection mechanism 312 may operate in the same manner to detect the canvas pan gesture and the element selection gesture regardless of which of the selectable elements 202 are associated with the selecting control. Furthermore, the pan detection mechanism 312 may operate to pan the canvas in the extendable dimension when the selecting control is activated and associated with an area of the canvas that is not associated with any of the selectable elements 202 . Accordingly, the principles described herein allow the user to pan the canvas 201 through a simple drag operation without having to worry about avoiding selectable elements when grabbing hold of the canvas with the selecting control. Furthermore, when the user desires to select a selectable element, the user may use the same selecting control to do that also.
- FIG. 4 illustrates a much more detailed user interface 400 that includes a canvas populated by a large number of selectable elements. Only a few of the selectable elements are labeled as they are more relevant for the discussion herein.
- the user has a selecting control in the form of a pointer 401 .
- the pointer 401 is activated at the position illustrated in FIG. 4 such that the pointer 401 is associated with the selectable control 402 A at the beginning of either a pan or selection gesture (depending on movement of the pointer 401 thereafter while selected).
- the pan detection mechanism 312 disambiguates the element selection gesture (and interaction gesture as the parts can have hotspots) versus the canvas pan gesture by analyzing the pointer (e.g., mouse) events (e.g., mouse-down, mouse-up and mouse-move events). If the user mouses-down and then moves more than a certain number (X) of pixels (e.g. 4 pixels) then the pan detection mechanism detects a pan gesture, rather than an element selection gesture. If the user mouses-down, then mouses-up, without moving more than X pixels, the pan detection mechanism 312 identifies an element selection gesture.
- the region defined by the X pixels may be referred to as a “pan-start diamond”, an area that the user may interface with to express their intent, without them ever needing to cognitively recognize that they are expressing this intent.
- the pan detection mechanism identifies an element selection gesture.
- the pan detection mechanism identifies at least one of (and potentially both of) the orthogonal pan gesture and the canvas pan gesture pans in accordance with movement of the selecting control.
- a selection gesture is no longer possible.
- the pan selection mechanism distinguishes between a canvas pan gesture or an orthogonal scroll gesture based on where the selecting control exited the pan start region.
- the dimensions of the pan-start diamond may be adjusted by simply changing the width and height of the pan-start diamond. Further, the pan-start diamond may be adjusted to polish the core panning experience. When users pan the canvas, they often do not move their mouse in a perfectly straight line. Rather, they pan in an arc. This arc is more exaggerated when panning left and right because the desired travel distance is often farther.
- FIG. 5 illustrates a user interface 500 that is similar to the user interface 400 of FIG. 4 , except that the arc associated with pointer movement is illustrated.
- the pointer has moved continuously from its original position to position 501 , 502 , 503 , 504 and 505 in succession, to create a drag motion 510 that includes an orthogonal dip at or near the pointer pan start position. If this arc is not accounted for, an intent to orthogonally scroll may be falsely inferred.
- an orthogonally scrollable element might orthogonally scroll slightly before the canvas begins to pan in the extendable dimension. This might reduce the user's feeling of the quality of the experience.
- the pan detection mechanism 312 may use a pan start region that is a diamond that is elongated rather than square.
- the pan start region has a greater span in the orthogonal dimension than in the extendable dimension of the canvas. More concretely, in the case of FIG. 5 , the pan start region is taller than it is wider. This means that the user has to pan a bit more orthogonally than they would in the extendable dimension before the pan detection mechanism detects an orthogonal pan gesture. In doing so, this greatly reduces chance that a user will see an orthogonal pan of a selectable element when the user really intended a canvas scan in the extendable dimension.
- FIG. 6 illustrates a user interface 600 in which an example pan start region 601 is illustrated with respect to the pan-start position 401 .
- the horizontal half-width of the pan-start diamond region is 3 pixels
- the vertical half-height of the pan-start diamond region is 6 pixels.
- the 3 pixel and 6 pixel values are for illustrative purposes only.
- FIG. 7 illustrates a flowchart of a method 700 for permitting user interaction with a canvas that extends in an extendable dimension on the user interface, and which is be populated with selectable elements.
- the method 700 is initiated upon detecting activation of a selecting control (event 701 ). For instance, the user might activate a pointer (e.g., press the left mouse button).
- a selecting control For instance, the user might activate a pointer (e.g., press the left mouse button).
- the method 700 defines the pan start region (act 702 ).
- the pan start region surrounds the area of the user interface where the user first activated the selection control in act 701 .
- the exact shape of the pan start region will define whether or not the pan detection mechanism concludes that the user intended to select the element (if the selecting control stays within the pan start area while the selecting control is active), or intended to pan the canvas or an orthogonally scrollable element (if the selecting control exits the pan start area at any point while activated). Accordingly, the shape of the pan start area should be carefully selected to as to reduce false detections of intent. In some embodiments, the shape of the pan start area may vary by user.
- the computing system can detect when it has wrongly detected intent for a given user, and adjust 1) the shape and/or size of the pan start area accordingly, and/or 2) the placement of the pan start area around the initial point that the activation control was initially activated. For instance, perhaps the pan start area is initially a diamond that has the initial point of activation at its center. However, with subsequent monitoring of user behavior, a month later, the system may have adjusted the pan start area to have a different shape (perhaps rounded slightly at the tips, or perhaps bulging slightly on one end), and may have perhaps shifted the pan start area rightward slightly with respect to the initial activation point. The computing system may thus experiment to improve prediction of user intent on a per-user basis.
- the user may do any one of a number of things.
- the pan detection mechanism actively monitors whether the selection control exits the pan start region (decision block 703 ) and whether or not the selection control is deactivated (decision block 704 ). These two decisions may occur quite frequently and thus may be considered as a single consolidated decision.
- the selection control is still activated (“No” in decision block 704 ) and the selection control does not exit the pan start region (“No” in decision block 703 )
- there is no significant change in state of the method 700 if the selection control is still activated (“No” in decision block 704 ) and the selection control does not exit the pan start region (“No” in decision block 703 ).
- the pan detection mechanism determines whether or not the selection control was activated at a selectable control within the user interface (decision block 705 ). If the selection control was activated at a selectable element (“Yes” in decision block 705 ), then the pan detection mechanism causes the selectable element to be activated (act 706 ). Otherwise, if the selection control was activated at an area of the user interface that is not selectable (“No” in decision block 705 ), then the method 700 simply ends (act 707 ). For instance, there might have just been a stray mouse click.
- panning occurs (act 708 ).
- the canvas may be panned as directed by the selection control.
- the pan detection mechanism may either pan the canvas or scroll the orthogonally scrollable element (or perhaps both), depending on where the selection control exited the pan start region.
- This panning continues as directed by any movement in the selection control, so long as the selection control continues to be activated (“No” in decision block 709 ).
- the selection control is deactivated (“Yes” in decision block 709 )
- the method 700 ends.
- pan start region there is no need for keeping track of the pan start region once the pan detection mechanism determines that the selection control has exited the pan start region.
- a pan start region will thus be generated each time the selection control is activated.
- the shape of the pan start region may differ according to the circumstances. For instance, as described above, the pan start region may have a different shape, size, and position relative to the point of activation, depending on the user.
- the pan start region might also differ by any one or more of the following circumstances: 1) whether the canvas is already visualizing the whole populated portion of the canvas, 2) whether the canvas is already panned all the way to one end or the other of the population portion of the canvas at one extent of the extendible dimension, 3) whether the selectable control was activated over a selectable element, 4) whether the selectable control was activated over an orthogonally scrollable element, and if so, the scrolled position of the orthogonally scrollable element in the orthogonal dimension, and so forth.
- the selecting of a selectable control results in an additional selectable control being created. For instance, selecting a selectable control might create a new user interface element that might be added to the populated portion of the canvas in the extendable direction, and which itself might include one or more selectable elements. Such further selectable elements in the new element might be further selected to add yet more user interface elements to the populated portion of the canvas.
- the canvas might have originally included only a favorites area 410 that includes multiple selectable elements including selectable element 411 .
- selectable element 411 When the selectable element 411 is selected, a new user interface element 420 appears.
- the user interface element may be thought of as a slice (or blade) of the canvas that, within given dimensions in the extendible direction of the canvas, extends along all or substantially all of the canvas in the orthogonal dimension. For instance, if the extendible direction is horizontal as in the case of FIG. 4 , a blade is a vertical slice of the canvas.
- the user interface element 420 might further include selectable elements such as selectable element 421 (and 402 A previously mentioned).
- a further user interface element 430 (e.g., a blade) appears and as well as being a selectable element itself, might further include selectable elements such as selectable element 431 .
- a further user interface element 440 (e.g., a blade) appears and as well as being a selectable element itself, might further include selectable elements such as selectable element 441 . This may continue to allow the user to engage in a journey building up a canvas of history showing the path taken to get to where the user is presently. Although the journey may begin in a manner that fits well in the displayed region, as the journey adds additional user interface elements, and the populated area of the canvas extends, the journey may quickly exceed the capability of the display to simultaneously visualize.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- User Interface Of Digital Computer (AREA)
- Stored Programmes (AREA)
- Woven Fabrics (AREA)
- Image Analysis (AREA)
Abstract
Description
- This application claims the benefit of each of the following provisional patent applications, and each of the following provisional patent applications are incorporated herein by reference in their entirety:
-
- 1. U.S. Provisional Application Ser. No. 61/905,119, filed Nov. 15, 2013;
- 2. U.S. Provisional Application Ser. No. 61/884,743, filed Sep. 30, 2013;
- 3. U.S. Provisional Application Ser. No. 61/905,111, filed Nov. 15, 2013;
- 4. U.S. Provisional Application Ser. No. 61/905,243, filed Nov. 17, 2013;
- 5. U.S. Provisional Application Ser. No. 61/905,114, filed Nov. 15, 2013;
- 6. U.S. Provisional Application Ser. No. 61/905,116, filed Nov. 15, 2013;
- 7. U.S. Provisional Application Ser. No. 61/905,129, filed Nov. 15, 2013;
- 8. U.S. Provisional Application Ser. No. 61/905,105, filed Nov. 15, 2013;
- 9. U.S. Provisional Application Ser. No. 61/905,247, filed Nov. 17, 2013;
- 10. U.S. Provisional Application Ser. No. 61/905,101, filed Nov. 15, 2013; and
- 11. U.S. Provisional Application Ser. No. 61/905,128, filed Nov. 15, 2013.
- A current paradigm for navigating through various information contexts is windows based. A classic example of this is the web browser experience. A user might begin with a home page that occupies the entire browser space. The user might then select a hyperlink, whereupon a new window appears. However, the previous window either disappears or, in the case of exercising an option to open the new page in a new window, the previous window is fully, or at least partially, hidden.
- The subject matter claimed herein is not limited to embodiments that solve any disadvantages or that operate only in environments such as those described above. Rather, this background is only provided to illustrate one exemplary technology area where some embodiments described herein may be practiced.
- At least some embodiments described herein relate to a user interface that has a canvas that is extendable in an extendable dimension and on which multiple selectable elements may be placed. In this context, when a selecting control is associated with a particular selectable element, and further user gestures are provided, a pan detection mechanism is configured to distinguish between an intent to pan the canvas and an intent to select the particular selectable element. In another embodiment, the pan detection mechanism is configured to distinguish between an intent to pan the canvas along its extendible direction (such as horizontally), and an intent to scroll a user interface element perpendicular (such as vertically) to the extendible direction of the canvas. In yet another embodiment, the pan detection mechanism is configured to distinguish between all three intents; that is 1) an intent to pan the canvas in its extendible direction, 2) an intent to scroll a user interface element perpendicular to the extendible direction of the canvas, and 3) an intent to select a user interface element.
- This summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used as an aid in determining the scope of the claimed subject matter.
- In order to describe the manner in which the above-recited and other advantages and features of the invention can be obtained, a more particular description of the invention briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only typical embodiments of the invention and are not therefore to be considered to be limiting of its scope, the invention will be described and explained with additional specificity and detail through the use of the accompanying drawings in which:
-
FIG. 1 abstractly illustrates an example computing system in which the principles described herein may be employed; -
FIG. 2 abstractly illustrates an environment in which a canvas which has several selectable user interface elements and is pannable in the extendible direction (e.g., horizontal) of the canvas, but several selectable elements are pannable (or scrollable) also in the direction orthogonal (e.g., vertical to the extendible direction) to the extendible direction of the canvas; -
FIG. 3 illustrates a supporting architecture for the mechanism for distinguishing between a panning and selection gestures in accordance with the principles described herein; -
FIG. 4 illustrates a much more detailed user interface that includes a canvas populated by a large number of selectable elements; -
FIG. 5 illustrates a user interface that is similar to the user interface ofFIG. 4 , except that the arc associated with pointer movement is illustrated; -
FIG. 6 illustrates a user interface in which an example pan start region is illustrated with respect to the pan-start position; and -
FIG. 7 illustrates a flowchart of a method for permitting user interaction with a canvas that extends in an extendable dimension on the user interface, and which is be populated with selectable elements. - At least some embodiments described herein relate to a user interface that has a canvas that is extendable in an extendable dimension and on which multiple selectable elements may be placed. In this context, when a selecting control is associated with a particular selectable element, and further user gestures are provided, a pan detection mechanism is configured to distinguish between an intent to pan the canvas and an intent to select the particular selectable element. Alternatively or in addition, when a selecting control is associated with a particular orthogonally scrollable element that may be scrolled orthogonal to the extendible direction of the canvas, the pen detection mechanism may be distinguish between an intent to pan the canvas and an intent to scroll the orthogonally scrollable element. Some introductory discussion of a computing system will be described with respect to
FIG. 1 . Then, example user interfaces, methods and supporting architectures will be described with respect to subsequent figures. - Computing systems are now increasingly taking a wide variety of forms. Computing systems may, for example, be handheld devices, appliances, laptop computers, desktop computers, mainframes, distributed computing systems, or even devices that have not conventionally been considered a computing system. In this description and in the claims, the term “computing system” is defined broadly as including any device or system (or combination thereof) that includes at least one physical and tangible processor, and a physical and tangible memory capable of having thereon computer-executable instructions that may be executed by the processor. The memory may take any form and may depend on the nature and form of the computing system. A computing system may be distributed over a network environment and may include multiple constituent computing systems.
- As illustrated in
FIG. 1 , in its most basic configuration, acomputing system 100 typically includes at least oneprocessing unit 102 andmemory 104. Thememory 104 may be physical system memory, which may be volatile, non-volatile, or some combination of the two. The term “memory” may also be used herein to refer to non-volatile mass storage such as physical storage media. If the computing system is distributed, the processing, memory and/or storage capability may be distributed as well. As used herein, the term “executable module” or “executable component” can refer to software objects, routines, or methods that may be executed on the computing system. The different components, modules, engines, and services described herein may be implemented as objects or processes that execute on the computing system (e.g., as separate threads). - In the description that follows, embodiments are described with reference to acts that are performed by one or more computing systems. If such acts are implemented in software, one or more processors of the associated computing system that performs the act direct the operation of the computing system in response to having executed computer-executable instructions. For example, such computer-executable instructions may be embodied on one or more computer-readable media that form a computer program product. An example of such an operation involves the manipulation of data. The computer-executable instructions (and the manipulated data) may be stored in the
memory 104 of thecomputing system 100.Computing system 100 may also contain communication channels 108 that allow thecomputing system 100 to communicate with other message processors over, for example,network 110. - The
computing system 100 also includes adisplay 112 on which a user interface, such as the user interfaces described herein, may be rendered. Such user interfaces may be generated in computer hardware or other computer-represented form prior to rendering. The presentation and/or rendering of such user interfaces may be performed by thecomputing system 100 by having the processing unit(s) 102 execute one or more computer-executable instructions that are embodied on one or more computer-readable media. Such computer-readable media may form all or a part of a computer program product. - Embodiments described herein may comprise or utilize a special purpose or general-purpose computer including computer hardware, such as, for example, one or more processors and system memory, as discussed in greater detail below. Embodiments described herein also include physical and other computer-readable media for carrying or storing computer-executable instructions and/or data structures. Such computer-readable media can be any available media that can be accessed by a general purpose or special purpose computer system. Computer-readable media that store computer-executable instructions are physical storage media. Computer-readable media that carry computer-executable instructions are transmission media. Thus, by way of example, and not limitation, embodiments of the invention can comprise at least two distinctly different kinds of computer-readable media: computer storage media and transmission media.
- Computer storage media includes RAM, ROM, EEPROM, CD-ROM or other optical disk storage, magnetic disk storage or other magnetic storage devices, or any other tangible medium which can be used to store desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer.
- A “network” is defined as one or more data links that enable the transport of electronic data between computer systems and/or modules and/or other electronic devices. When information is transferred or provided over a network or another communications connection (either hardwired, wireless, or a combination of hardwired or wireless) to a computer, the computer properly views the connection as a transmission medium. Transmissions media can include a network and/or data links which can be used to carry or desired program code means in the form of computer-executable instructions or data structures and which can be accessed by a general purpose or special purpose computer. Combinations of the above should also be included within the scope of computer-readable media.
- Further, upon reaching various computer system components, program code means in the form of computer-executable instructions or data structures can be transferred automatically from transmission media to computer storage media (or vice versa). For example, computer-executable instructions or data structures received over a network or data link can be buffered in RAM within a network interface module (e.g., a “NIC”), and then eventually transferred to computer system RAM and/or to less volatile computer storage media at a computer system. Thus, it should be understood that computer storage media can be included in computer system components that also (or even primarily) utilize transmission media.
- Computer-executable instructions comprise, for example, instructions and data which, when executed at a processor, cause a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, or even source code. Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the described features or acts described above. Rather, the described features and acts are disclosed as example forms of implementing the claims.
- Those skilled in the art will appreciate that the invention may be practiced in network computing environments with many types of computer system configurations, including, personal computers, desktop computers, laptop computers, message processors, hand-held devices, multi-processor systems, microprocessor-based or programmable consumer electronics, network PCs, minicomputers, mainframe computers, mobile telephones, PDAs, pagers, routers, switches, and the like. The invention may also be practiced in distributed system environments where local and remote computer systems, which are linked (either by hardwired data links, wireless data links, or by a combination of hardwired and wireless data links) through a network, both perform tasks. In a distributed system environment, program modules may be located in both local and remote memory storage devices.
-
FIG. 2 abstractly illustrates anenvironment 200 in which aportion 201A of acanvas 201 is displayed on a displayedarea 210 of a display, and a portion 201B of thecanvas 201 is not displayed in the displayedarea 210 of the display. As an example, the display might be, for instance, thedisplay 112 of thecomputing system 100 ofFIG. 1 . However, thedisplay 200 might alternatively be a window within adisplay 112, so as not to occupy the entire area of thedisplay 112. Thecanvas 201 extends in anextendable direction 221 along a singleextendable dimension 222. For instance, inFIG. 2 , theextendable direction 221 is rightward, and theextendable dimension 222 is the horizontal dimension. - In some embodiments, the
extendable dimension 222 of thecanvas 201 may be configurable to be vertical or horizontal according to the tastes of the user. In other cases, the limitations of thedisplay 210 or container might restrictextendable dimension 222 to the horizontal dimension, or to the vertical dimension. Also, theextendable direction 221 may be configurable. For instance, if theextendable dimension 222 is vertical, the user might configure the canvas to extend downward, or upward, according to the user's preference. If theextendable dimension 222 is horizontal (as is the case ofFIG. 2 ), the user might configure the canvas to extend leftward, or rightward (as is the case ofFIG. 2 ), according to the user's preference. - The
canvas 201 is illustrated as including multiple selectable elements 202. For instance, the selectable elements 202 in this particular case are illustrated as includingselectable elements 202A through 202F. Some of theselectable elements 202A and 202B are within the displayedarea 201A of thecanvas 201, and some of the selectable elements 202D through 202F are in the non-displayed area 201B of thecanvas 201. One of the selectable elements 202C is illustrated as spanning a boundary between the displayedportion 201A and the undisplayed portion 201B of thecanvas 201. Hereinafter, the portion of the canvas that contains elements (whether selectable or otherwise) will be referred to as the “populated portion” of thecanvas 201. The selectable element might also include navigation controls. - When the populated portion of the
canvas 201 is too large to display all at one time, the user interface may have a panning mechanism that permits the user to pan thecanvas 201 along the extensible dimension. Two of the selectable elements 202B and 202E are illustrated as having a greater extent in a dimension orthogonal to the extendable direction (hereinafter an “orthogonal dimension”). InFIG. 2 , the orthogonal dimension is the vertical dimension represented bybi-directional arrow 223. When a particular element has a greater extent in the orthogonal dimension than the display is capable of rendering at one time, the particular element itself may be pannable in the orthogonal dimension. - In accordance with the principles described herein, after a selecting control (such as a mouse cursor) is associated with a selectable element (such as by pressing the left mouse button of the mouse while the mouse cursor is displayed over the selectable element), a pan detection mechanism distinguishes between a canvas pan gesture and an element selection gesture through subsequent interaction with the selectable element. For instance, in the
canvas 201, perhaps the user may have the experience of being able to pan the canvas by dragging thecanvas 201 along theextendable dimension 222. Yet, there are selectable elements 202 throughout the populated portion of thecanvas 201. The pan detection mechanism allows the user to pan thecanvas 201 by activating the selecting control on any portion of thecanvas 201 even if that portion happens to be occupied by a selectable element. Thus, if the user wants to pan, the user need not be concerned about avoiding selection of elements. Instead, the user may provide a natural panning gesture on any portion of the canvas. On the other hand, if the user desires to select a particular selectable element, the user provides a selection gesture while the selecting control is associated with that particular selectable element. - Additionally, after a selecting control (such as a mouse cursor) is associated with an orthogonally scrollable element that is scrollable perpendicular to the extendible dimension of the canvas, the pan detection mechanism distinguishes between a canvas pan gesture and an element scroll gesture. For instance, in the
canvas 201, perhaps the user may have the experience of being able to pan the canvas by dragging thecanvas 201 along theextendable dimension 222. Yet, there are orthogonally scrollable elements 202B and 202E throughout the populated portion of thecanvas 201. The pan detection mechanism distinguishes between an intent to pan thecanvas 201 in the extendible dimension, and an intent to scroll the orthogonally scrollable element 202B or 202E associated with the selecting control. Furthermore, in this case, if the selecting control is associated with a selectable element within that orthogonally scrollable element, the panning control distinguishes between intents to pan the canvas in the extendible dimension, scroll the orthogonally scrollable element in the orthogonal dimension, or select the selectable element. -
FIG. 3 illustrates a supportingarchitecture 300 for a user interface 301. For instance, the user interface 301 is an example of theuser interface 200 ofFIG. 2 . The supportingarchitecture 300 is illustrated as including a selectingcontrol 311 that is capable of selecting any of the selectable elements within the canvas. Apan detection mechanism 312 may be used to distinguish between a pan gesture, an orthogonally scrollable element scroll gesture, and an element selection gesture. Anactivation module 321 may be used to respond to the selection gesture by selecting the appropriate selectable element. Apan module 322 may be used to pan the canvas in response to thepan detection mechanism 312 detecting a pan gesture and/or scroll the orthogonally scrollable element in response to the orthogonally scrollable element scroll gesture. - For instance, suppose the selecting control is over the
selectable element 202A on thecanvas 201 when the selecting control is activated. Depending on the movement of the selecting control during this activation, the user may provide a canvas pan gesture if the user desires to pan thecanvas 201 along theextendable dimension 222, or the user may provide an element selection gesture if the user desires to select theselectable element 202A. - As mentioned above, there are even more possibilities for user intent if the selecting control is over the selectable element 202B on the
canvas 201 when the selecting control is activated. In this case, depending on the movement of the selecting control during this activation, the user may additionally provide an orthogonal pan gesture (also called herein an orthogonal scroll gesture) if the user desires to pan the selectable element 202B in the orthogonal direction. This option is in addition to the possibility of a canvas pan gesture if the user desires to pan thecanvas 201 along theextendable dimension 222, and an element selection gesture if the user desires to select the selectable element 202B. - Referring to
FIG. 3 , thepan detection mechanism 312 may be used to distinguish between the element selection gesture, the canvas pan gesture, and the orthogonal pan gesture when the selectingcontrol 311 is associated with a selectable element. If thepan detection mechanism 312 detects an element selection gesture, theactivation module 321 may be used to respond to the selection gesture by selecting the appropriate selectable element. If thepan detection mechanism 312 detects a canvas pan gesture, thepan module 322 may be used to pan the canvas in the extendable dimension. If thepan detection mechanism 312 detects an orthogonal pan gesture, thepan module 322 may be used to pan the selectable element in the orthogonal dimension. - The
pan detection mechanism 312 may operate in the same manner to detect the canvas pan gesture and the element selection gesture regardless of which of the selectable elements 202 are associated with the selecting control. Furthermore, thepan detection mechanism 312 may operate to pan the canvas in the extendable dimension when the selecting control is activated and associated with an area of the canvas that is not associated with any of the selectable elements 202. Accordingly, the principles described herein allow the user to pan thecanvas 201 through a simple drag operation without having to worry about avoiding selectable elements when grabbing hold of the canvas with the selecting control. Furthermore, when the user desires to select a selectable element, the user may use the same selecting control to do that also. -
FIG. 4 illustrates a much moredetailed user interface 400 that includes a canvas populated by a large number of selectable elements. Only a few of the selectable elements are labeled as they are more relevant for the discussion herein. InFIG. 4 , the user has a selecting control in the form of apointer 401. Thepointer 401 is activated at the position illustrated inFIG. 4 such that thepointer 401 is associated with the selectable control 402A at the beginning of either a pan or selection gesture (depending on movement of thepointer 401 thereafter while selected). - In one embodiment, the
pan detection mechanism 312 disambiguates the element selection gesture (and interaction gesture as the parts can have hotspots) versus the canvas pan gesture by analyzing the pointer (e.g., mouse) events (e.g., mouse-down, mouse-up and mouse-move events). If the user mouses-down and then moves more than a certain number (X) of pixels (e.g. 4 pixels) then the pan detection mechanism detects a pan gesture, rather than an element selection gesture. If the user mouses-down, then mouses-up, without moving more than X pixels, thepan detection mechanism 312 identifies an element selection gesture. The region defined by the X pixels may be referred to as a “pan-start diamond”, an area that the user may interface with to express their intent, without them ever needing to cognitively recognize that they are expressing this intent. - For instance, if the selecting control stays within the pan start region until the selecting control is no longer activated, the pan detection mechanism identifies an element selection gesture. On the other hand, if the selecting control goes outside of the pan start region while the selecting control is still activated, the pan detection mechanism identifies at least one of (and potentially both of) the orthogonal pan gesture and the canvas pan gesture pans in accordance with movement of the selecting control. Upon the first exit of the selecting control outside of the pan start region while the selection control remains activated, a selection gesture is no longer possible. Furthermore, if the selecting control was activated over an orthogonally scrollable element, the pan selection mechanism distinguishes between a canvas pan gesture or an orthogonal scroll gesture based on where the selecting control exited the pan start region.
- In order to accommodate users of different mouse-skill and touch experiences, the dimensions of the pan-start diamond may be adjusted by simply changing the width and height of the pan-start diamond. Further, the pan-start diamond may be adjusted to polish the core panning experience. When users pan the canvas, they often do not move their mouse in a perfectly straight line. Rather, they pan in an arc. This arc is more exaggerated when panning left and right because the desired travel distance is often farther.
-
FIG. 5 illustrates a user interface 500 that is similar to theuser interface 400 ofFIG. 4 , except that the arc associated with pointer movement is illustrated. For instance, the pointer has moved continuously from its original position to position 501, 502, 503, 504 and 505 in succession, to create a drag motion 510 that includes an orthogonal dip at or near the pointer pan start position. If this arc is not accounted for, an intent to orthogonally scroll may be falsely inferred. Thus, an orthogonally scrollable element might orthogonally scroll slightly before the canvas begins to pan in the extendable dimension. This might reduce the user's feeling of the quality of the experience. - In order to compensate for this user tendency, the
pan detection mechanism 312 may use a pan start region that is a diamond that is elongated rather than square. In the illustrated case, the pan start region has a greater span in the orthogonal dimension than in the extendable dimension of the canvas. More concretely, in the case ofFIG. 5 , the pan start region is taller than it is wider. This means that the user has to pan a bit more orthogonally than they would in the extendable dimension before the pan detection mechanism detects an orthogonal pan gesture. In doing so, this greatly reduces chance that a user will see an orthogonal pan of a selectable element when the user really intended a canvas scan in the extendable dimension. -
FIG. 6 illustrates auser interface 600 in which an example pan start region 601 is illustrated with respect to thepan-start position 401. In this embodiment, the horizontal half-width of the pan-start diamond region is 3 pixels, and the vertical half-height of the pan-start diamond region is 6 pixels. However, the 3 pixel and 6 pixel values are for illustrative purposes only. -
FIG. 7 illustrates a flowchart of a method 700 for permitting user interaction with a canvas that extends in an extendable dimension on the user interface, and which is be populated with selectable elements. The method 700 is initiated upon detecting activation of a selecting control (event 701). For instance, the user might activate a pointer (e.g., press the left mouse button). - Next, the method 700 defines the pan start region (act 702). Essentially, the pan start region surrounds the area of the user interface where the user first activated the selection control in
act 701. The exact shape of the pan start region will define whether or not the pan detection mechanism concludes that the user intended to select the element (if the selecting control stays within the pan start area while the selecting control is active), or intended to pan the canvas or an orthogonally scrollable element (if the selecting control exits the pan start area at any point while activated). Accordingly, the shape of the pan start area should be carefully selected to as to reduce false detections of intent. In some embodiments, the shape of the pan start area may vary by user. - Furthermore, perhaps the computing system can detect when it has wrongly detected intent for a given user, and adjust 1) the shape and/or size of the pan start area accordingly, and/or 2) the placement of the pan start area around the initial point that the activation control was initially activated. For instance, perhaps the pan start area is initially a diamond that has the initial point of activation at its center. However, with subsequent monitoring of user behavior, a month later, the system may have adjusted the pan start area to have a different shape (perhaps rounded slightly at the tips, or perhaps bulging slightly on one end), and may have perhaps shifted the pan start area rightward slightly with respect to the initial activation point. The computing system may thus experiment to improve prediction of user intent on a per-user basis.
- In any case, while the selection control is active, the user may do any one of a number of things. The pan detection mechanism actively monitors whether the selection control exits the pan start region (decision block 703) and whether or not the selection control is deactivated (decision block 704). These two decisions may occur quite frequently and thus may be considered as a single consolidated decision. At any given time, if the selection control is still activated (“No” in decision block 704) and the selection control does not exit the pan start region (“No” in decision block 703), then there is no significant change in state of the method 700. However, if the selection control is deactivated (“Yes” in decision block 704) while the selection control still has not exited the pan start region (still “No” in decision block 703), the pan detection mechanism determines whether or not the selection control was activated at a selectable control within the user interface (decision block 705). If the selection control was activated at a selectable element (“Yes” in decision block 705), then the pan detection mechanism causes the selectable element to be activated (act 706). Otherwise, if the selection control was activated at an area of the user interface that is not selectable (“No” in decision block 705), then the method 700 simply ends (act 707). For instance, there might have just been a stray mouse click.
- On the other hand, if while the selection control is still activated (from act 701), and the pan detection mechanism detects that the selection control has exited the pan start region (“Yes” in decision block 703), then panning occurs (act 708). For instance, the canvas may be panned as directed by the selection control. As another more complicated example, if the user activated the selection mechanism at an orthogonally scrollable element, the pan detection mechanism may either pan the canvas or scroll the orthogonally scrollable element (or perhaps both), depending on where the selection control exited the pan start region. This panning (act 708) continues as directed by any movement in the selection control, so long as the selection control continues to be activated (“No” in decision block 709). However, when the selection control is deactivated (“Yes” in decision block 709), the method 700 ends.
- Note that there is no need for keeping track of the pan start region once the pan detection mechanism determines that the selection control has exited the pan start region. A pan start region will thus be generated each time the selection control is activated. In some circumstances, the shape of the pan start region may differ according to the circumstances. For instance, as described above, the pan start region may have a different shape, size, and position relative to the point of activation, depending on the user. However, the pan start region might also differ by any one or more of the following circumstances: 1) whether the canvas is already visualizing the whole populated portion of the canvas, 2) whether the canvas is already panned all the way to one end or the other of the population portion of the canvas at one extent of the extendible dimension, 3) whether the selectable control was activated over a selectable element, 4) whether the selectable control was activated over an orthogonally scrollable element, and if so, the scrolled position of the orthogonally scrollable element in the orthogonal dimension, and so forth.
- In one example, the selecting of a selectable control results in an additional selectable control being created. For instance, selecting a selectable control might create a new user interface element that might be added to the populated portion of the canvas in the extendable direction, and which itself might include one or more selectable elements. Such further selectable elements in the new element might be further selected to add yet more user interface elements to the populated portion of the canvas.
- For instance, in
FIG. 4 , the canvas might have originally included only afavorites area 410 that includes multiple selectable elements includingselectable element 411. When theselectable element 411 is selected, a newuser interface element 420 appears. The user interface element may be thought of as a slice (or blade) of the canvas that, within given dimensions in the extendible direction of the canvas, extends along all or substantially all of the canvas in the orthogonal dimension. For instance, if the extendible direction is horizontal as in the case ofFIG. 4 , a blade is a vertical slice of the canvas. As well as being a selectable element itself, theuser interface element 420 might further include selectable elements such as selectable element 421 (and 402A previously mentioned). When theselectable element 421 is selected, a further user interface element 430 (e.g., a blade) appears and as well as being a selectable element itself, might further include selectable elements such asselectable element 431. When theselectable element 431 is selected, a further user interface element 440 (e.g., a blade) appears and as well as being a selectable element itself, might further include selectable elements such asselectable element 441. This may continue to allow the user to engage in a journey building up a canvas of history showing the path taken to get to where the user is presently. Although the journey may begin in a manner that fits well in the displayed region, as the journey adds additional user interface elements, and the populated area of the canvas extends, the journey may quickly exceed the capability of the display to simultaneously visualize. - The present invention may be embodied in other specific forms without departing from its spirit or essential characteristics. The described embodiments are to be considered in all respects only as illustrative and not restrictive. The scope of the invention is, therefore, indicated by the appended claims rather than by the foregoing description. All changes which come within the meaning and range of equivalency of the claims are to be embraced within their scope.
Claims (21)
Priority Applications (17)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US14/231,917 US20150095846A1 (en) | 2013-09-30 | 2014-04-01 | Pan and selection gesture detection |
| SG11201601888UA SG11201601888UA (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| CA2922725A CA2922725A1 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| EP14786398.9A EP3053030A1 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| RU2016111604A RU2679540C2 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| MX2016004113A MX2016004113A (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection. |
| SG10201802632SA SG10201802632SA (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| BR112016004551A BR112016004551A8 (en) | 2013-09-30 | 2014-09-29 | method for allowing user interaction with a backdrop that extends to an extensible dimension in a user interface, system, and computer program product |
| HK16110904.6A HK1222731A1 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| JP2016518686A JP6465870B2 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| AU2014324620A AU2014324620A1 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| PCT/US2014/057940 WO2015048602A1 (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| KR1020167008352A KR20160063340A (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| CN201480054049.3A CN105593812A (en) | 2013-09-30 | 2014-09-29 | Pan and selection gesture detection |
| PH12016500256A PH12016500256A1 (en) | 2013-09-30 | 2016-02-05 | Pan and selection gesture detection |
| IL244368A IL244368A0 (en) | 2013-09-30 | 2016-03-01 | Pan and selection gesture detection |
| CL2016000729A CL2016000729A1 (en) | 2013-09-30 | 2016-03-30 | A method for detecting a panoramic and selection gesture in a user interface program. |
Applications Claiming Priority (12)
| Application Number | Priority Date | Filing Date | Title |
|---|---|---|---|
| US201361884743P | 2013-09-30 | 2013-09-30 | |
| US201361905111P | 2013-11-15 | 2013-11-15 | |
| US201361905116P | 2013-11-15 | 2013-11-15 | |
| US201361905119P | 2013-11-15 | 2013-11-15 | |
| US201361905101P | 2013-11-15 | 2013-11-15 | |
| US201361905128P | 2013-11-15 | 2013-11-15 | |
| US201361905129P | 2013-11-15 | 2013-11-15 | |
| US201361905114P | 2013-11-15 | 2013-11-15 | |
| US201361905105P | 2013-11-15 | 2013-11-15 | |
| US201361905243P | 2013-11-17 | 2013-11-17 | |
| US201361905247P | 2013-11-17 | 2013-11-17 | |
| US14/231,917 US20150095846A1 (en) | 2013-09-30 | 2014-04-01 | Pan and selection gesture detection |
Publications (1)
| Publication Number | Publication Date |
|---|---|
| US20150095846A1 true US20150095846A1 (en) | 2015-04-02 |
Family
ID=52741177
Family Applications (11)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/231,862 Active 2035-01-24 US9792354B2 (en) | 2013-09-30 | 2014-04-01 | Context aware user interface parts |
| US14/231,912 Abandoned US20150095849A1 (en) | 2013-09-30 | 2014-04-01 | Dialogs positioned with action visualization |
| US14/231,891 Expired - Fee Related US9483549B2 (en) | 2013-09-30 | 2014-04-01 | Persisting state at scale across browser sessions |
| US14/231,883 Expired - Fee Related US9672276B2 (en) | 2013-09-30 | 2014-04-01 | Multi-act creation user interface element |
| US14/231,917 Abandoned US20150095846A1 (en) | 2013-09-30 | 2014-04-01 | Pan and selection gesture detection |
| US14/231,873 Abandoned US20150095812A1 (en) | 2013-09-30 | 2014-04-01 | Extensible and context-aware commanding infrastructure |
| US14/231,880 Abandoned US20150095365A1 (en) | 2013-09-30 | 2014-04-01 | Query building using schema |
| US14/231,846 Abandoned US20150095842A1 (en) | 2013-09-30 | 2014-04-01 | Extendable blade sequence along pannable canvas direction |
| US14/231,869 Expired - Fee Related US9754018B2 (en) | 2013-09-30 | 2014-04-01 | Rendering interpreter for visualizing data provided from restricted environment container |
| US14/231,905 Expired - Fee Related US9727636B2 (en) | 2013-09-30 | 2014-04-01 | Generating excutable code from complaint and non-compliant controls |
| US14/231,897 Expired - Fee Related US9805114B2 (en) | 2013-09-30 | 2014-04-01 | Composable selection model through reusable component |
Family Applications Before (4)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/231,862 Active 2035-01-24 US9792354B2 (en) | 2013-09-30 | 2014-04-01 | Context aware user interface parts |
| US14/231,912 Abandoned US20150095849A1 (en) | 2013-09-30 | 2014-04-01 | Dialogs positioned with action visualization |
| US14/231,891 Expired - Fee Related US9483549B2 (en) | 2013-09-30 | 2014-04-01 | Persisting state at scale across browser sessions |
| US14/231,883 Expired - Fee Related US9672276B2 (en) | 2013-09-30 | 2014-04-01 | Multi-act creation user interface element |
Family Applications After (6)
| Application Number | Title | Priority Date | Filing Date |
|---|---|---|---|
| US14/231,873 Abandoned US20150095812A1 (en) | 2013-09-30 | 2014-04-01 | Extensible and context-aware commanding infrastructure |
| US14/231,880 Abandoned US20150095365A1 (en) | 2013-09-30 | 2014-04-01 | Query building using schema |
| US14/231,846 Abandoned US20150095842A1 (en) | 2013-09-30 | 2014-04-01 | Extendable blade sequence along pannable canvas direction |
| US14/231,869 Expired - Fee Related US9754018B2 (en) | 2013-09-30 | 2014-04-01 | Rendering interpreter for visualizing data provided from restricted environment container |
| US14/231,905 Expired - Fee Related US9727636B2 (en) | 2013-09-30 | 2014-04-01 | Generating excutable code from complaint and non-compliant controls |
| US14/231,897 Expired - Fee Related US9805114B2 (en) | 2013-09-30 | 2014-04-01 | Composable selection model through reusable component |
Country Status (17)
| Country | Link |
|---|---|
| US (11) | US9792354B2 (en) |
| EP (6) | EP3053031A1 (en) |
| JP (2) | JP6465870B2 (en) |
| KR (3) | KR102186865B1 (en) |
| CN (6) | CN105593813B (en) |
| AU (2) | AU2014324618A1 (en) |
| BR (1) | BR112016004551A8 (en) |
| CA (2) | CA2922725A1 (en) |
| CL (1) | CL2016000729A1 (en) |
| HK (1) | HK1222731A1 (en) |
| IL (1) | IL244368A0 (en) |
| MX (2) | MX2016004113A (en) |
| PH (1) | PH12016500256A1 (en) |
| RU (2) | RU2686822C2 (en) |
| SG (2) | SG10201802632SA (en) |
| TW (4) | TW201528106A (en) |
| WO (7) | WO2015048206A1 (en) |
Cited By (2)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20170220198A1 (en) * | 2016-02-03 | 2017-08-03 | Seiko Epson Corporation | Electronic apparatus and control program of electronic apparatus |
| US10289297B2 (en) * | 2016-08-26 | 2019-05-14 | Google Llc | Animating an image to indicate that the image is pannable |
Families Citing this family (36)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| USD750112S1 (en) * | 2013-01-04 | 2016-02-23 | Samsung Electronics Co., Ltd. | Portable electronic device with graphical user interface |
| US9792354B2 (en) * | 2013-09-30 | 2017-10-17 | Microsoft Technology Licensing, Llc | Context aware user interface parts |
| USD745877S1 (en) * | 2013-10-17 | 2015-12-22 | Microsoft Corporation | Display screen with graphical user interface |
| US10176218B2 (en) * | 2014-11-07 | 2019-01-08 | Sap Se | OData custom query composer |
| US10990258B2 (en) * | 2015-08-04 | 2021-04-27 | Google Llc | Interactively presenting a visible portion of a rendering surface on a user device |
| US10452750B2 (en) | 2015-08-04 | 2019-10-22 | Google Llc | Systems and methods for interactively presenting a visible portion of a rendering surface on a user device |
| US9871911B1 (en) * | 2016-09-30 | 2018-01-16 | Microsoft Technology Licensing, Llc | Visualizations for interactions with external computing logic |
| KR102605332B1 (en) * | 2016-11-02 | 2023-11-23 | 주식회사 넥슨코리아 | Device and method to provide content |
| US10796088B2 (en) * | 2017-04-21 | 2020-10-06 | International Business Machines Corporation | Specifying a conversational computer agent and its outcome with a grammar |
| WO2018208047A1 (en) | 2017-05-09 | 2018-11-15 | Samsung Electronics Co., Ltd. | Method and system for managing and displaying application |
| US10827319B2 (en) * | 2017-06-02 | 2020-11-03 | Apple Inc. | Messaging system interacting with dynamic extension app |
| CN110019717B (en) * | 2017-12-15 | 2021-06-29 | 上海智臻智能网络科技股份有限公司 | Device for modifying multi-turn question-answering system |
| US11341422B2 (en) | 2017-12-15 | 2022-05-24 | SHANGHAI XIAOl ROBOT TECHNOLOGY CO., LTD. | Multi-round questioning and answering methods, methods for generating a multi-round questioning and answering system, and methods for modifying the system |
| CN110019718B (en) * | 2017-12-15 | 2021-04-09 | 上海智臻智能网络科技股份有限公司 | Method for modifying multi-turn question-answering system, terminal equipment and storage medium |
| US11379252B1 (en) * | 2018-01-31 | 2022-07-05 | Parallels International Gmbh | System and method for providing layouts for a remote desktop session |
| US11659003B2 (en) | 2018-08-30 | 2023-05-23 | International Business Machines Corporation | Safe shell container facilitating inspection of a virtual container |
| US11048871B2 (en) * | 2018-09-18 | 2021-06-29 | Tableau Software, Inc. | Analyzing natural language expressions in a data visualization user interface |
| US10902045B2 (en) * | 2018-09-18 | 2021-01-26 | Tableau Software, Inc. | Natural language interface for building data visualizations, including cascading edits to filter expressions |
| CN109542563B (en) * | 2018-11-09 | 2022-06-07 | 优信数享(北京)信息技术有限公司 | Multi-state integrated android page management method, device and system |
| US11385766B2 (en) | 2019-01-07 | 2022-07-12 | AppEsteem Corporation | Technologies for indicating deceptive and trustworthy resources |
| EP3764210B1 (en) | 2019-07-08 | 2025-11-19 | dSPACE GmbH | Display of display areas on a desktop |
| US11089050B1 (en) * | 2019-08-26 | 2021-08-10 | Ca, Inc. | Isolating an iframe of a webpage |
| US11042558B1 (en) | 2019-09-06 | 2021-06-22 | Tableau Software, Inc. | Determining ranges for vague modifiers in natural language commands |
| US11474975B2 (en) | 2019-09-18 | 2022-10-18 | Microsoft Technology Licensing, Llc | Identity represented assets in a content management system |
| US11199955B2 (en) * | 2019-10-02 | 2021-12-14 | Palantir Technologies Inc. | Enhanced techniques for building user interfaces |
| CN110825766A (en) * | 2019-11-13 | 2020-02-21 | 恩亿科(北京)数据科技有限公司 | Query condition generation method and device, server and readable storage medium |
| CN110995942B (en) * | 2019-12-06 | 2021-08-06 | 科大国创软件股份有限公司 | Soft switch automatic calling method and system based on interface visualization |
| CN111177455A (en) * | 2019-12-31 | 2020-05-19 | 精英数智科技股份有限公司 | Method, device and equipment for determining cutting tooth load type of coal mining machine and storage medium |
| KR102873167B1 (en) * | 2019-12-31 | 2025-10-22 | 한국과학기술원 | Secure user interface distribution method for heterogeneous multi-device interaction |
| CN111610912B (en) * | 2020-04-24 | 2023-10-10 | 北京小米移动软件有限公司 | Application display method, application display device and storage medium |
| CN114205365B (en) * | 2020-08-31 | 2023-06-20 | 华为技术有限公司 | Application interface migration system, method and related equipment |
| EP4208200A2 (en) | 2020-09-02 | 2023-07-12 | Genmab A/S | Antibody therapy |
| US11698933B1 (en) | 2020-09-18 | 2023-07-11 | Tableau Software, LLC | Using dynamic entity search during entry of natural language commands for visual data analysis |
| US11301631B1 (en) | 2020-10-05 | 2022-04-12 | Tableau Software, LLC | Visually correlating individual terms in natural language input to respective structured phrases representing the natural language input |
| CN112732243B (en) * | 2021-01-11 | 2024-06-18 | 京东科技控股股份有限公司 | Data processing method and device for generating functional components |
| US11363050B1 (en) | 2021-03-25 | 2022-06-14 | Bank Of America Corporation | Information security system and method for incompliance detection in data transmission |
Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020015054A1 (en) * | 1996-03-12 | 2002-02-07 | Sumitaka Hibino | Graphics displaying method and apparatus |
| US6473891B1 (en) * | 2000-05-03 | 2002-10-29 | Lsi Logic Corporation | Wire routing to control skew |
| US20060048071A1 (en) * | 2004-08-30 | 2006-03-02 | Microsoft Corp. | Scrolling web pages using direct interaction |
| US20060214926A1 (en) * | 2005-03-22 | 2006-09-28 | Microsoft Corporation | Targeting in a stylus-based user interface |
| US20070024646A1 (en) * | 2005-05-23 | 2007-02-01 | Kalle Saarinen | Portable electronic apparatus and associated method |
| US20090271723A1 (en) * | 2008-04-24 | 2009-10-29 | Nintendo Co., Ltd. | Object display order changing program and apparatus |
| US20090292989A1 (en) * | 2008-05-23 | 2009-11-26 | Microsoft Corporation | Panning content utilizing a drag operation |
| US20090319939A1 (en) * | 2008-06-20 | 2009-12-24 | Microsoft Corporation | Controlled interaction with heterogeneous data |
| US20100058227A1 (en) * | 2008-08-29 | 2010-03-04 | Microsoft Corporation | Internal scroll activation and cursor adornment |
| US20100328209A1 (en) * | 2008-02-15 | 2010-12-30 | Panasonic Corporation | Input device for electronic apparatus |
| US20120051596A1 (en) * | 2010-08-31 | 2012-03-01 | Activate Systems, Inc. | Methods and apparatus for improved motioin capture |
| US20130106744A1 (en) * | 2011-10-26 | 2013-05-02 | Sony Computer Entertainment Inc. | Scroll control device, terminal device, and scroll control method |
| US20130222301A1 (en) * | 2012-02-24 | 2013-08-29 | Samsung Electronics Co., Ltd. | Method and apparatus for moving contents in terminal |
| US20140096042A1 (en) * | 2012-07-09 | 2014-04-03 | Aaron Tyler Travis | Method and system for generating and storing a collection of interactive browsers within a navigation plane |
Family Cites Families (168)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| JPS6362033A (en) * | 1986-09-02 | 1988-03-18 | Nec Corp | Display device for relative information |
| AU6707394A (en) * | 1993-04-20 | 1994-11-08 | Apple Computer, Inc. | Interactive user interface |
| US5625763A (en) | 1995-05-05 | 1997-04-29 | Apple Computer, Inc. | Method and apparatus for automatically generating focus ordering in a dialog on a computer system |
| US5845299A (en) * | 1996-07-29 | 1998-12-01 | Rae Technology Llc | Draw-based editor for web pages |
| US6049812A (en) | 1996-11-18 | 2000-04-11 | International Business Machines Corp. | Browser and plural active URL manager for network computers |
| US6128632A (en) * | 1997-03-06 | 2000-10-03 | Apple Computer, Inc. | Methods for applying rubi annotation characters over base text characters |
| US6091415A (en) * | 1997-05-02 | 2000-07-18 | Inventec Corporation | System and method for displaying multiple dialog boxes in a window display |
| US5886694A (en) * | 1997-07-14 | 1999-03-23 | Microsoft Corporation | Method for automatically laying out controls in a dialog window |
| US5995101A (en) * | 1997-10-29 | 1999-11-30 | Adobe Systems Incorporated | Multi-level tool tip |
| US6236400B1 (en) * | 1998-04-02 | 2001-05-22 | Sun Microsystems, Inc. | Method and apparatus for controlling the display of hierarchical information |
| US6473102B1 (en) | 1998-05-11 | 2002-10-29 | Apple Computer, Inc. | Method and system for automatically resizing and repositioning windows in response to changes in display |
| US7801913B2 (en) | 1998-12-07 | 2010-09-21 | Oracle International Corporation | System and method for querying data for implicit hierarchies |
| US6460060B1 (en) | 1999-01-26 | 2002-10-01 | International Business Machines Corporation | Method and system for searching web browser history |
| JP2000331020A (en) * | 1999-05-21 | 2000-11-30 | Nippon Telegr & Teleph Corp <Ntt> | Information reference method, information reference device, and storage medium storing information reference program |
| US6701513B1 (en) | 2000-01-14 | 2004-03-02 | Measurement Computing Corporation | Program-development environment for use in generating application programs |
| US7243335B1 (en) | 2000-02-17 | 2007-07-10 | Microsoft Corporation | Method and system for reducing coding complexity by providing intelligent manipulable defaults |
| US6681383B1 (en) | 2000-04-04 | 2004-01-20 | Sosy, Inc. | Automatic software production system |
| US7062475B1 (en) | 2000-05-30 | 2006-06-13 | Alberti Anemometer Llc | Personalized multi-service computer environment |
| US6750887B1 (en) | 2000-06-02 | 2004-06-15 | Sun Microsystems, Inc. | Graphical user interface layout manager |
| US7171455B1 (en) | 2000-08-22 | 2007-01-30 | International Business Machines Corporation | Object oriented based, business class methodology for generating quasi-static web pages at periodic intervals |
| US6919890B2 (en) | 2000-09-28 | 2005-07-19 | Curl Corporation | Grid and table layout using elastics |
| US6640655B1 (en) * | 2000-10-03 | 2003-11-04 | Varco I/P, Inc. | Self tracking sensor suspension mechanism |
| US6950198B1 (en) | 2000-10-18 | 2005-09-27 | Eastman Kodak Company | Effective transfer of images from a user to a service provider |
| US7370040B1 (en) | 2000-11-21 | 2008-05-06 | Microsoft Corporation | Searching with adaptively configurable user interface and extensible query language |
| WO2002046878A2 (en) * | 2000-12-06 | 2002-06-13 | American Express Travel Related Services Company, Inc. | Layout generator system and method |
| US6760128B2 (en) | 2000-12-06 | 2004-07-06 | Eastman Kodak Company | Providing a payment schedule for utilizing stored images using a designated date |
| JP2002182812A (en) * | 2000-12-14 | 2002-06-28 | Smg Kk | Site map display system |
| US7233998B2 (en) | 2001-03-22 | 2007-06-19 | Sony Computer Entertainment Inc. | Computer architecture and software cells for broadband networks |
| US7203678B1 (en) | 2001-03-27 | 2007-04-10 | Bea Systems, Inc. | Reconfigurable query generation system for web browsers |
| US20020147963A1 (en) | 2001-04-09 | 2002-10-10 | Lee Rusty Shawn | Method and apparatus for generating machine control instructions |
| US20020180811A1 (en) | 2001-05-31 | 2002-12-05 | Chu Sing Yun | Systems, methods, and articles of manufacture for providing a user interface with selection and scrolling |
| US20030011638A1 (en) | 2001-07-10 | 2003-01-16 | Sun-Woo Chung | Pop-up menu system |
| US6950993B2 (en) * | 2001-08-02 | 2005-09-27 | Microsoft Corporation | System and method for automatic and dynamic layout of resizable dialog type windows |
| US6944829B2 (en) | 2001-09-25 | 2005-09-13 | Wind River Systems, Inc. | Configurable user-interface component management system |
| US7480864B2 (en) | 2001-10-12 | 2009-01-20 | Canon Kabushiki Kaisha | Zoom editor |
| US7620908B2 (en) | 2001-12-28 | 2009-11-17 | Sap Ag | Managing a user interface |
| US20050066037A1 (en) * | 2002-04-10 | 2005-03-24 | Yu Song | Browser session mobility system for multi-platform applications |
| CA2385224C (en) | 2002-05-07 | 2012-10-02 | Corel Corporation | Dockable drop-down dialogs |
| US7065707B2 (en) | 2002-06-24 | 2006-06-20 | Microsoft Corporation | Segmenting and indexing web pages using function-based object models |
| US7293024B2 (en) | 2002-11-14 | 2007-11-06 | Seisint, Inc. | Method for sorting and distributing data among a plurality of nodes |
| US20080177994A1 (en) | 2003-01-12 | 2008-07-24 | Yaron Mayer | System and method for improving the efficiency, comfort, and/or reliability in Operating Systems, such as for example Windows |
| US7000184B2 (en) | 2003-01-24 | 2006-02-14 | The Cobalt Group, Inc. | Remote web site editing in a standard web browser without external software |
| US20040165009A1 (en) * | 2003-02-20 | 2004-08-26 | International Business Machines Corporation | Expansion of interactive user interface components |
| US7823077B2 (en) | 2003-03-24 | 2010-10-26 | Microsoft Corporation | System and method for user modification of metadata in a shell browser |
| US7769794B2 (en) | 2003-03-24 | 2010-08-03 | Microsoft Corporation | User interface for a file system shell |
| US7720616B2 (en) | 2003-05-07 | 2010-05-18 | Sureprep, Llc | Multi-stage, multi-user engagement submission and tracking process |
| US7417644B2 (en) | 2003-05-12 | 2008-08-26 | Microsoft Corporation | Dynamic pluggable user interface layout |
| US7669140B2 (en) | 2003-08-21 | 2010-02-23 | Microsoft Corporation | System and method for providing rich minimized applications |
| US8230366B2 (en) | 2003-10-23 | 2012-07-24 | Apple Inc. | Dynamically changing cursor for user interface |
| US8037420B2 (en) | 2003-12-04 | 2011-10-11 | International Business Machines Corporation | Maintaining browser navigation relationships and for choosing a browser window for new documents |
| US7711742B2 (en) | 2003-12-11 | 2010-05-04 | International Business Machines Corporation | Intelligent data query builder |
| US20080109785A1 (en) | 2004-01-16 | 2008-05-08 | Bailey Bendrix L | Graphical Program Having Graphical and/or Textual Specification of Event Handler Procedures for Program Objects |
| GB2411331A (en) * | 2004-02-19 | 2005-08-24 | Trigenix Ltd | Rendering user interface using actor attributes |
| US7577938B2 (en) * | 2004-02-20 | 2009-08-18 | Microsoft Corporation | Data association |
| US7536672B1 (en) | 2004-03-05 | 2009-05-19 | Adobe Systems Incorporated | Management of user interaction history with software applications |
| US7694233B1 (en) * | 2004-04-30 | 2010-04-06 | Apple Inc. | User interface presentation of information in reconfigured or overlapping containers |
| CN100343802C (en) * | 2004-05-10 | 2007-10-17 | 华为技术有限公司 | Method and system for unifying users'interface |
| US8453065B2 (en) | 2004-06-25 | 2013-05-28 | Apple Inc. | Preview and installation of user interface elements in a display environment |
| US8046712B2 (en) | 2004-06-29 | 2011-10-25 | Acd Systems International Inc. | Management of multiple window panels with a graphical user interface |
| US8117542B2 (en) * | 2004-08-16 | 2012-02-14 | Microsoft Corporation | User interface for displaying selectable software functionality controls that are contextually relevant to a selected object |
| US20090043765A1 (en) | 2004-08-20 | 2009-02-12 | Rhoderick John Kennedy Pugh | Server authentication |
| US7720867B2 (en) | 2004-09-08 | 2010-05-18 | Oracle International Corporation | Natural language query construction using purpose-driven template |
| US8819569B2 (en) | 2005-02-18 | 2014-08-26 | Zumobi, Inc | Single-handed approach for navigation of application tiles using panning and zooming |
| US20060224951A1 (en) * | 2005-03-30 | 2006-10-05 | Yahoo! Inc. | Multiple window browser interface and system and method of generating multiple window browser interface |
| US20060236264A1 (en) * | 2005-04-18 | 2006-10-19 | Microsoft Corporation | Automatic window resize behavior and optimizations |
| US8195646B2 (en) * | 2005-04-22 | 2012-06-05 | Microsoft Corporation | Systems, methods, and user interfaces for storing, searching, navigating, and retrieving electronic information |
| US7721225B2 (en) | 2005-05-03 | 2010-05-18 | Novell, Inc. | System and method for creating and presenting modal dialog boxes in server-side component web applications |
| US7730418B2 (en) | 2005-05-04 | 2010-06-01 | Workman Nydegger | Size to content windows for computer graphics |
| US20060282771A1 (en) * | 2005-06-10 | 2006-12-14 | Tad Vinci | Verifying document compliance to a subsidiary standard |
| US20070033522A1 (en) | 2005-08-02 | 2007-02-08 | Lin Frank L | System and method for dynamic resizing of web-based GUIs |
| US7933632B2 (en) | 2005-09-16 | 2011-04-26 | Microsoft Corporation | Tile space user interface for mobile devices |
| US8543824B2 (en) | 2005-10-27 | 2013-09-24 | Apple Inc. | Safe distribution and use of content |
| US7954064B2 (en) * | 2005-10-27 | 2011-05-31 | Apple Inc. | Multiple dashboards |
| US8434021B2 (en) * | 2005-11-30 | 2013-04-30 | Microsoft Corporation | Centralized user interface for displaying contextually driven business content and business related functionality |
| US7836303B2 (en) | 2005-12-09 | 2010-11-16 | University Of Washington | Web browser operating system |
| US8898203B2 (en) | 2005-12-27 | 2014-11-25 | International Business Machines Corporation | Generating a separable query design object and database schema through visual view editing |
| JP4635894B2 (en) | 2006-02-13 | 2011-02-23 | ソニー株式会社 | Information processing apparatus and method, and program |
| JP4415961B2 (en) | 2006-03-15 | 2010-02-17 | ブラザー工業株式会社 | Removable media device and data control program |
| US20070233854A1 (en) | 2006-03-31 | 2007-10-04 | Microsoft Corporation | Management status summaries |
| US20070234195A1 (en) | 2006-04-03 | 2007-10-04 | National Instruments Corporation | Simultaneous update of a plurality of user interface elements displayed in a web browser |
| US7685519B1 (en) * | 2006-07-18 | 2010-03-23 | Intuit Inc. | Process and apparatus for providing a customizable content tooltip |
| US20080018665A1 (en) * | 2006-07-24 | 2008-01-24 | Jay Behr | System and method for visualizing drawing style layer combinations |
| US8869027B2 (en) | 2006-08-04 | 2014-10-21 | Apple Inc. | Management and generation of dashboards |
| US20080065974A1 (en) | 2006-09-08 | 2008-03-13 | Tom Campbell | Template-based electronic presence management |
| US7890957B2 (en) | 2006-09-08 | 2011-02-15 | Easyonme, Inc. | Remote management of an electronic presence |
| US20080109714A1 (en) | 2006-11-03 | 2008-05-08 | Sap Ag | Capturing screen information |
| US8082539B1 (en) | 2006-12-11 | 2011-12-20 | Parallels Holdings, Ltd. | System and method for managing web-based forms and dynamic content of website |
| JP5031353B2 (en) | 2006-12-15 | 2012-09-19 | キヤノン株式会社 | Display device, control method, and program |
| CN101004685A (en) * | 2007-01-08 | 2007-07-25 | 叶炜 | Method for realizing graphical user interface |
| US9032329B2 (en) | 2007-03-23 | 2015-05-12 | Siemens Product Lifecycle Management Software Inc. | System and method for dialog position management |
| US8321847B1 (en) | 2007-05-17 | 2012-11-27 | The Mathworks, Inc. | Dynamic function wizard |
| US20080306933A1 (en) * | 2007-06-08 | 2008-12-11 | Microsoft Corporation | Display of search-engine results and list |
| US10019570B2 (en) | 2007-06-14 | 2018-07-10 | Microsoft Technology Licensing, Llc | Protection and communication abstractions for web browsers |
| US8065628B2 (en) | 2007-06-25 | 2011-11-22 | Microsoft Corporation | Dynamic user interface for previewing live content |
| KR20090000507A (en) * | 2007-06-28 | 2009-01-07 | 삼성전자주식회사 | Information display method and device |
| US8762880B2 (en) * | 2007-06-29 | 2014-06-24 | Microsoft Corporation | Exposing non-authoring features through document status information in an out-space user interface |
| US7873746B2 (en) | 2007-07-27 | 2011-01-18 | Lagavulin Limited | User interface for a portable, image-processing transmitter |
| US9009181B2 (en) | 2007-08-23 | 2015-04-14 | International Business Machines Corporation | Accessing objects in a service registry and repository |
| US8126840B2 (en) | 2007-10-22 | 2012-02-28 | Noria Corporation | Lubrication program management system and methods |
| US8046353B2 (en) | 2007-11-02 | 2011-10-25 | Citrix Online Llc | Method and apparatus for searching a hierarchical database and an unstructured database with a single search query |
| CN101499004A (en) * | 2008-01-31 | 2009-08-05 | 株式会社日立制作所 | System and method for connecting virtual machine and user interface |
| US20090254822A1 (en) | 2008-04-04 | 2009-10-08 | International Business Machines Corporation | Hi-efficiency wizard framework system and method |
| US8219385B2 (en) | 2008-04-08 | 2012-07-10 | Incentive Targeting, Inc. | Computer-implemented method and system for conducting a search of electronically stored information |
| US20100005053A1 (en) | 2008-07-04 | 2010-01-07 | Estes Philip F | Method for enabling discrete back/forward actions within a dynamic web application |
| US8345014B2 (en) * | 2008-07-12 | 2013-01-01 | Lester F. Ludwig | Control of the operating system on a computing device via finger angle using a high dimensional touchpad (HDTP) touch user interface |
| US8402381B2 (en) * | 2008-09-23 | 2013-03-19 | International Business Machines Corporation | Automatically arranging widgets of a model within a canvas using iterative region based widget relative adjustments |
| US8095412B1 (en) | 2008-11-03 | 2012-01-10 | Intuit Inc. | Method and system for evaluating expansion of a business |
| KR20100049474A (en) * | 2008-11-03 | 2010-05-12 | 삼성전자주식회사 | A method for remote user interface session migration to other device |
| WO2010063016A1 (en) * | 2008-11-26 | 2010-06-03 | Lila Aps (Ahead.) | Dynamic level of detail |
| US7962547B2 (en) | 2009-01-08 | 2011-06-14 | International Business Machines Corporation | Method for server-side logging of client browser state through markup language |
| US20100229115A1 (en) | 2009-03-05 | 2010-09-09 | Microsoft Corporation | Zoomable user interface data generation |
| US8806371B2 (en) * | 2009-03-26 | 2014-08-12 | Apple Inc. | Interface navigation tools |
| US8819570B2 (en) * | 2009-03-27 | 2014-08-26 | Zumobi, Inc | Systems, methods, and computer program products displaying interactive elements on a canvas |
| US20100251143A1 (en) | 2009-03-27 | 2010-09-30 | The Ransom Group, Inc. | Method, system and computer program for creating and editing a website |
| US8819597B2 (en) | 2009-04-10 | 2014-08-26 | Google Inc. | Glyph entry on computing device |
| US9213541B2 (en) | 2009-04-17 | 2015-12-15 | ArtinSoft Corporation, S.A. | Creation, generation, distribution and application of self-contained modifications to source code |
| US20100287530A1 (en) | 2009-05-05 | 2010-11-11 | Borland Software Corporation | Requirements definition using interactive prototyping |
| US8269737B2 (en) * | 2009-08-20 | 2012-09-18 | Hewlett-Packard Development Company, L.P. | Method and apparatus for interpreting input movement on a computing device interface as a one- or two-dimensional input |
| EP2467800B1 (en) | 2009-11-13 | 2021-06-30 | Irdeto B.V. | System and method to protect java bytecode code against static and dynamic attacks within hostile execution environments |
| US8782562B2 (en) * | 2009-12-02 | 2014-07-15 | Dell Products L.P. | Identifying content via items of a navigation system |
| US8407598B2 (en) | 2009-12-09 | 2013-03-26 | Ralph Lee Burton | Dynamic web control generation facilitator |
| JP5523090B2 (en) * | 2009-12-25 | 2014-06-18 | キヤノン株式会社 | INPUT DEVICE, CONTROL METHOD FOR INPUT DEVICE, PROGRAM, AND STORAGE MEDIUM |
| US8533667B2 (en) | 2009-12-30 | 2013-09-10 | International Business Machines Corporation | Call wizard for information management system (IMS) applications |
| CN101763218A (en) * | 2010-01-06 | 2010-06-30 | 广东欧珀移动通信有限公司 | Input method of handheld device |
| WO2011085387A2 (en) | 2010-01-11 | 2011-07-14 | Everspeech, Inc. | Integrated data processing and transcription service |
| EP2548200A4 (en) | 2010-03-19 | 2014-01-22 | Siemens Healthcare Diagnostics | SYSTEM AND METHOD FOR MODEL WINDOW MODEL WINDOWS |
| US8316323B2 (en) | 2010-03-26 | 2012-11-20 | Microsoft Corporation | Breadcrumb navigation through heirarchical structures |
| US8631350B2 (en) * | 2010-04-23 | 2014-01-14 | Blackberry Limited | Graphical context short menu |
| US20120089914A1 (en) * | 2010-04-27 | 2012-04-12 | Surfwax Inc. | User interfaces for navigating structured content |
| US20110271184A1 (en) * | 2010-04-28 | 2011-11-03 | Microsoft Corporation | Client application and web page integration |
| US9160756B2 (en) * | 2010-05-19 | 2015-10-13 | International Business Machines Corporation | Method and apparatus for protecting markup language document against cross-site scripting attack |
| US9110586B2 (en) * | 2010-06-03 | 2015-08-18 | Panasonic Intellectual Property Corporation Of America | Scrolling apparatus, scrolling method, non-transitory computer readable recording medium and intergrated circuit |
| CN102270125A (en) * | 2010-06-04 | 2011-12-07 | 中兴通讯股份有限公司 | Device and method for developing Web application |
| US20110314415A1 (en) * | 2010-06-21 | 2011-12-22 | George Fitzmaurice | Method and System for Providing Custom Tooltip Messages |
| US8706854B2 (en) | 2010-06-30 | 2014-04-22 | Raytheon Company | System and method for organizing, managing and running enterprise-wide scans |
| US8544027B2 (en) | 2010-07-30 | 2013-09-24 | Sap Ag | Logical data model abstraction in a physically distributed environment |
| JP2012069065A (en) * | 2010-09-27 | 2012-04-05 | Nintendo Co Ltd | Information processing program, and information processing device and method |
| US8612366B2 (en) | 2010-09-29 | 2013-12-17 | Moresteam.Com Llc | Systems and methods for performing design of experiments |
| US8990199B1 (en) | 2010-09-30 | 2015-03-24 | Amazon Technologies, Inc. | Content search with category-aware visual similarity |
| US20120124555A1 (en) * | 2010-11-11 | 2012-05-17 | Codekko Software, Inc. | Optimization of Compiled Control Objects |
| US9069577B2 (en) * | 2010-11-23 | 2015-06-30 | Apple Inc. | Grouping and browsing open windows |
| CN102023749A (en) * | 2010-12-02 | 2011-04-20 | 广东宝莱特医用科技股份有限公司 | Area dragging treating method of list type control on touch screen interface of medical equipment |
| US8612874B2 (en) * | 2010-12-23 | 2013-12-17 | Microsoft Corporation | Presenting an application change through a tile |
| US20120191502A1 (en) * | 2011-01-20 | 2012-07-26 | John Nicholas Gross | System & Method For Analyzing & Predicting Behavior Of An Organization & Personnel |
| JP2012168790A (en) | 2011-02-15 | 2012-09-06 | Brother Ind Ltd | Display program and display device |
| US9384183B2 (en) * | 2011-03-31 | 2016-07-05 | Infosys Limited | Method and system for reporting web standard non-compliance of web pages |
| US9152616B2 (en) | 2011-04-28 | 2015-10-06 | Flipboard, Inc. | Template-based page layout for web content |
| US9753699B2 (en) | 2011-06-16 | 2017-09-05 | Microsoft Technology Licensing, Llc | Live browser tooling in an integrated development environment |
| US8566100B2 (en) | 2011-06-21 | 2013-10-22 | Verna Ip Holdings, Llc | Automated method and system for obtaining user-selected real-time information on a mobile communication device |
| US8799862B2 (en) | 2011-06-24 | 2014-08-05 | Alcatel Lucent | Application testing using sandboxes |
| CN102253841B (en) * | 2011-08-09 | 2014-07-23 | 东莞兆田数码科技有限公司 | Small-sized graphic user interface system |
| US20130080913A1 (en) * | 2011-09-22 | 2013-03-28 | Microsoft Corporation | Multi-column notebook interaction |
| US8836654B2 (en) | 2011-10-04 | 2014-09-16 | Qualcomm Incorporated | Application window position and size control in (multi-fold) multi-display devices |
| KR101888457B1 (en) | 2011-11-16 | 2018-08-16 | 삼성전자주식회사 | Apparatus having a touch screen processing plurality of apllications and method for controlling thereof |
| US8799780B2 (en) | 2011-11-28 | 2014-08-05 | International Business Machines Corporation | Installation wizard with multidimensional views |
| US8799988B2 (en) | 2012-01-25 | 2014-08-05 | Microsoft Corporation | Document communication runtime interfaces |
| US20150058709A1 (en) | 2012-01-26 | 2015-02-26 | Michael Edward Zaletel | Method of creating a media composition and apparatus therefore |
| DE102013202782A1 (en) | 2012-02-20 | 2013-08-22 | Wixpress Ltd | Server-based web page design system that integrates dynamic layout and dynamic content |
| US9389872B2 (en) * | 2012-03-16 | 2016-07-12 | Vmware, Inc. | Software wizard implementation framework |
| EP2665042A1 (en) | 2012-05-14 | 2013-11-20 | Crytek GmbH | Visual processing based on interactive rendering |
| US9043722B1 (en) * | 2012-06-19 | 2015-05-26 | Surfwax, Inc. | User interfaces for displaying relationships between cells in a grid |
| US9195477B1 (en) | 2012-10-09 | 2015-11-24 | Sencha, Inc. | Device profiles, deep linking, and browser history support for web applications |
| US9244971B1 (en) | 2013-03-07 | 2016-01-26 | Amazon Technologies, Inc. | Data retrieval from heterogeneous storage systems |
| US9158518B2 (en) * | 2013-03-11 | 2015-10-13 | Blackberry Limited | Collaborative application development environment using a connected device |
| WO2014157908A1 (en) * | 2013-03-27 | 2014-10-02 | Samsung Electronics Co., Ltd. | Device and method for displaying execution result of application |
| US10410003B2 (en) | 2013-06-07 | 2019-09-10 | Apple Inc. | Multiple containers assigned to an application |
| US9792354B2 (en) * | 2013-09-30 | 2017-10-17 | Microsoft Technology Licensing, Llc | Context aware user interface parts |
| US9875116B2 (en) | 2013-11-26 | 2018-01-23 | Cellco Partnership | Sharing of a user input interface of an application session of one application between two or more applications |
-
2014
- 2014-04-01 US US14/231,862 patent/US9792354B2/en active Active
- 2014-04-01 US US14/231,912 patent/US20150095849A1/en not_active Abandoned
- 2014-04-01 US US14/231,891 patent/US9483549B2/en not_active Expired - Fee Related
- 2014-04-01 US US14/231,883 patent/US9672276B2/en not_active Expired - Fee Related
- 2014-04-01 US US14/231,917 patent/US20150095846A1/en not_active Abandoned
- 2014-04-01 US US14/231,873 patent/US20150095812A1/en not_active Abandoned
- 2014-04-01 US US14/231,880 patent/US20150095365A1/en not_active Abandoned
- 2014-04-01 US US14/231,846 patent/US20150095842A1/en not_active Abandoned
- 2014-04-01 US US14/231,869 patent/US9754018B2/en not_active Expired - Fee Related
- 2014-04-01 US US14/231,905 patent/US9727636B2/en not_active Expired - Fee Related
- 2014-04-01 US US14/231,897 patent/US9805114B2/en not_active Expired - Fee Related
- 2014-08-28 TW TW103129730A patent/TW201528106A/en unknown
- 2014-09-24 TW TW103133045A patent/TW201516834A/en unknown
- 2014-09-24 TW TW103133046A patent/TW201528108A/en unknown
- 2014-09-25 WO PCT/US2014/057324 patent/WO2015048206A1/en not_active Ceased
- 2014-09-25 WO PCT/US2014/057321 patent/WO2015048203A1/en not_active Ceased
- 2014-09-25 CN CN201480054050.6A patent/CN105593813B/en active Active
- 2014-09-25 WO PCT/US2014/057322 patent/WO2015048204A1/en not_active Ceased
- 2014-09-25 KR KR1020167008327A patent/KR102186865B1/en not_active Expired - Fee Related
- 2014-09-25 EP EP14825480.8A patent/EP3053031A1/en not_active Ceased
- 2014-09-25 EP EP14783733.0A patent/EP3053027A1/en not_active Ceased
- 2014-09-25 CN CN201480053811.6A patent/CN105683907B/en active Active
- 2014-09-25 CN CN201480054151.3A patent/CN105683909B/en active Active
- 2014-09-25 WO PCT/US2014/057323 patent/WO2015048205A1/en not_active Ceased
- 2014-09-25 EP EP14783734.8A patent/EP3053028B1/en active Active
- 2014-09-25 TW TW103133249A patent/TW201528103A/en unknown
- 2014-09-29 JP JP2016518686A patent/JP6465870B2/en not_active Expired - Fee Related
- 2014-09-29 CA CA2922725A patent/CA2922725A1/en not_active Abandoned
- 2014-09-29 RU RU2016111610A patent/RU2686822C2/en not_active IP Right Cessation
- 2014-09-29 MX MX2016004113A patent/MX2016004113A/en unknown
- 2014-09-29 CN CN201480054049.3A patent/CN105593812A/en active Pending
- 2014-09-29 EP EP14783748.8A patent/EP3053029A1/en not_active Ceased
- 2014-09-29 JP JP2016518687A patent/JP6446038B2/en not_active Expired - Fee Related
- 2014-09-29 CN CN201480053988.6A patent/CN105683908B/en not_active Expired - Fee Related
- 2014-09-29 CN CN201480053821.XA patent/CN105659199A/en active Pending
- 2014-09-29 EP EP14786398.9A patent/EP3053030A1/en not_active Withdrawn
- 2014-09-29 WO PCT/US2014/057938 patent/WO2015048600A1/en not_active Ceased
- 2014-09-29 WO PCT/US2014/057940 patent/WO2015048602A1/en not_active Ceased
- 2014-09-29 AU AU2014324618A patent/AU2014324618A1/en not_active Abandoned
- 2014-09-29 RU RU2016111604A patent/RU2679540C2/en not_active IP Right Cessation
- 2014-09-29 EP EP14790838.8A patent/EP3053017A1/en not_active Withdrawn
- 2014-09-29 SG SG10201802632SA patent/SG10201802632SA/en unknown
- 2014-09-29 BR BR112016004551A patent/BR112016004551A8/en not_active Application Discontinuation
- 2014-09-29 KR KR1020167008368A patent/KR20160064115A/en not_active Withdrawn
- 2014-09-29 KR KR1020167008352A patent/KR20160063340A/en not_active Withdrawn
- 2014-09-29 MX MX2016003946A patent/MX2016003946A/en unknown
- 2014-09-29 CA CA2922985A patent/CA2922985A1/en not_active Abandoned
- 2014-09-29 HK HK16110904.6A patent/HK1222731A1/en unknown
- 2014-09-29 SG SG11201601888UA patent/SG11201601888UA/en unknown
- 2014-09-29 WO PCT/US2014/057939 patent/WO2015048601A1/en not_active Ceased
- 2014-09-29 AU AU2014324620A patent/AU2014324620A1/en not_active Abandoned
-
2016
- 2016-02-05 PH PH12016500256A patent/PH12016500256A1/en unknown
- 2016-03-01 IL IL244368A patent/IL244368A0/en unknown
- 2016-03-30 CL CL2016000729A patent/CL2016000729A1/en unknown
Patent Citations (14)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20020015054A1 (en) * | 1996-03-12 | 2002-02-07 | Sumitaka Hibino | Graphics displaying method and apparatus |
| US6473891B1 (en) * | 2000-05-03 | 2002-10-29 | Lsi Logic Corporation | Wire routing to control skew |
| US20060048071A1 (en) * | 2004-08-30 | 2006-03-02 | Microsoft Corp. | Scrolling web pages using direct interaction |
| US20060214926A1 (en) * | 2005-03-22 | 2006-09-28 | Microsoft Corporation | Targeting in a stylus-based user interface |
| US20070024646A1 (en) * | 2005-05-23 | 2007-02-01 | Kalle Saarinen | Portable electronic apparatus and associated method |
| US20100328209A1 (en) * | 2008-02-15 | 2010-12-30 | Panasonic Corporation | Input device for electronic apparatus |
| US20090271723A1 (en) * | 2008-04-24 | 2009-10-29 | Nintendo Co., Ltd. | Object display order changing program and apparatus |
| US20090292989A1 (en) * | 2008-05-23 | 2009-11-26 | Microsoft Corporation | Panning content utilizing a drag operation |
| US20090319939A1 (en) * | 2008-06-20 | 2009-12-24 | Microsoft Corporation | Controlled interaction with heterogeneous data |
| US20100058227A1 (en) * | 2008-08-29 | 2010-03-04 | Microsoft Corporation | Internal scroll activation and cursor adornment |
| US20120051596A1 (en) * | 2010-08-31 | 2012-03-01 | Activate Systems, Inc. | Methods and apparatus for improved motioin capture |
| US20130106744A1 (en) * | 2011-10-26 | 2013-05-02 | Sony Computer Entertainment Inc. | Scroll control device, terminal device, and scroll control method |
| US20130222301A1 (en) * | 2012-02-24 | 2013-08-29 | Samsung Electronics Co., Ltd. | Method and apparatus for moving contents in terminal |
| US20140096042A1 (en) * | 2012-07-09 | 2014-04-03 | Aaron Tyler Travis | Method and system for generating and storing a collection of interactive browsers within a navigation plane |
Cited By (4)
| Publication number | Priority date | Publication date | Assignee | Title |
|---|---|---|---|---|
| US20170220198A1 (en) * | 2016-02-03 | 2017-08-03 | Seiko Epson Corporation | Electronic apparatus and control program of electronic apparatus |
| US10838550B2 (en) * | 2016-02-03 | 2020-11-17 | Seiko Epson Corporation | Changing specification of operation based on start position |
| US10289297B2 (en) * | 2016-08-26 | 2019-05-14 | Google Llc | Animating an image to indicate that the image is pannable |
| US10884601B2 (en) * | 2016-08-26 | 2021-01-05 | Google Llc | Animating an image to indicate that the image is pannable |
Also Published As
Similar Documents
| Publication | Publication Date | Title |
|---|---|---|
| US20150095846A1 (en) | Pan and selection gesture detection | |
| KR102123089B1 (en) | Automatically expanding panes | |
| US9189131B2 (en) | Context menus | |
| JP5970086B2 (en) | Touch screen hover input processing | |
| US8732613B2 (en) | Dynamic user interface for navigating among GUI elements | |
| US9851896B2 (en) | Edge swiping gesture for home navigation | |
| EP2282259B1 (en) | User interface method used in web browsing, electronic device for performing the same and computer readable recording medium thereof | |
| RU2540826C2 (en) | Internal scroll activation and cursor adornment | |
| US20150324067A1 (en) | Vehicle infotainment gateway - multi-application interface | |
| US20140098053A1 (en) | Electronic device, method for viewing desktop thereof, and computer-readable medium | |
| JP2004152169A (en) | Window switching device and window switching program | |
| US20140344735A1 (en) | Methods, apparatuses and computer program products for managing different visual variants of objects via user interfaces | |
| EP3025222A1 (en) | Scrollable smart menu | |
| US20140351749A1 (en) | Methods, apparatuses and computer program products for merging areas in views of user interfaces | |
| US20100333016A1 (en) | Scrollbar | |
| KR20130119708A (en) | Portable appratus comprising touch screens for browsing information displyed on screen of external appratus and method for browsing information thereof | |
| CN110865748A (en) | Menu presenting method and device, electronic equipment and computer readable storage medium | |
| EP2851772B1 (en) | Display device, display control method and non-temporary computer-readable medium | |
| JP2014106625A (en) | Portable terminal, control method of portable terminal, program and recording medium | |
| US11132104B1 (en) | Managing user interface items in a visual user interface (VUI) | |
| US20160103573A1 (en) | Scalable and tabbed user interface | |
| US20210191605A1 (en) | Method for displaying web browser and terminal device using the same | |
| WO2014132882A1 (en) | Terminal device, information display method and recording medium | |
| KR102031104B1 (en) | Web browser display apparatus and web browser display method | |
| JP2021060667A (en) | Information processing device, information processing method and program |
Legal Events
| Date | Code | Title | Description |
|---|---|---|---|
| AS | Assignment |
Owner name: MICROSOFT CORPORATION, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:DANTON, STEPHEN MICHAEL;OWENS, KRISTOFER JOHN;STERLING, JONAH BUSH;AND OTHERS;SIGNING DATES FROM 20140630 TO 20140806;REEL/FRAME:033490/0912 |
|
| AS | Assignment |
Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:034747/0417 Effective date: 20141014 Owner name: MICROSOFT TECHNOLOGY LICENSING, LLC, WASHINGTON Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MICROSOFT CORPORATION;REEL/FRAME:039025/0454 Effective date: 20141014 |
|
| STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |