WO2002033578A2 - Dynamically displaying current status of tasks - Google Patents

Dynamically displaying current status of tasks Download PDF

Info

Publication number
WO2002033578A2
WO2002033578A2 PCT/US2001/031987 US0131987W WO0233578A2 WO 2002033578 A2 WO2002033578 A2 WO 2002033578A2 US 0131987 W US0131987 W US 0131987W WO 0233578 A2 WO0233578 A2 WO 0233578A2
Authority
WO
WIPO (PCT)
Prior art keywords
list
user
current
recited
task
Prior art date
Application number
PCT/US2001/031987
Other languages
French (fr)
Other versions
WO2002033578A8 (en
Inventor
Kenneth H. Abbott, Iii
Dan Newell
James O. Robarts
Original Assignee
Tangis Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tangis Corporation filed Critical Tangis Corporation
Priority to AU2002211699A priority Critical patent/AU2002211699A1/en
Publication of WO2002033578A2 publication Critical patent/WO2002033578A2/en
Publication of WO2002033578A8 publication Critical patent/WO2002033578A8/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06QINFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
    • G06Q10/00Administration; Management
    • G06Q10/10Office automation; Time management
    • G06Q10/109Time management, e.g. calendars, reminders, meetings or time accounting

Definitions

  • the present invention is directed to graphical user interfaces and more particularly to dynamically displaying the current status of tasks.
  • computers As computers become increasingly powerful and commonplace, they are being used for an increasingly broad variety of tasks. For example, in addition to traditional activities such as running word processing and database applications, computers are increasingly becoming an integral part of users' daily lives. Programs to schedule activities, generate reminders, and provide rapid communication capabilities are becoming increasingly popular. Moreover, computers are increasingly present during virtually all of a person's daily activities. For example, hand-held computer organizers (e.g., PDAs) are increasingly common, and communication devices such as portable phones are increasingly incorporating computer capabilities. More recently, the field of wearable computers (e.g., with eyeglass displays) has begun to expand, creating a further presence of computers in people's daily lives. Computers often progress through a particular series of steps when allowing a user to accomplish a particular task.
  • a user desires to enter a new name and address to an electronic address book
  • the computer progresses through a series of steps prompting the user to enter the desired information (e.g., name, street address, city, state, zip code, phone number, etc.).
  • desired information e.g., name, street address, city, state, zip code, phone number, etc.
  • large displays e.g., typical desktop computers
  • UI informative and useable user interface
  • the nature of many new computing devices with small displays are examples of many new computing devices with small displays
  • a list of items corresponding to tasks that are to be performed are displayed.
  • the tasks may be performed by a user (e.g., data entered by the user, words spoken by the user, actions taken by the user, and so forth) or alternatively by a computer (e.g., the steps followed in carrying out a programmed task).
  • a user e.g., data entered by the user, words spoken by the user, actions taken by the user, and so forth
  • a computer e.g., the steps followed in carrying out a programmed task.
  • At least a portion of the list is displayed at any given time along with an indication of which task is the next task to be performed.
  • the current status of his or her progression through the corresponding items on the list is dynamically updated so as to readily inform the user (or someone else) as to what the current task is that needs to be performed, as well as what tasks have already been performed and/or what tasks remain to be performed.
  • only a subset of the list of items is displayed at any given time.
  • the list is scrolled through as the tasks are performed so that different items are displayed as part of the subset as tasks are performed.
  • multiple lists of tasks to be performed by multiple individuals are displayed on a display of the user.
  • an indication of such completion is forwarded to the user's computer, which updates the display to indicate the next task in the list to be displayed. The user is thus able to monitor the progress of the multiple individuals (or computing devices) in carrying out their respective tasks.
  • Fig. 1 illustrates an exemplary computing device such as may be used in accordance with certain embodiments of the invention.
  • Fig. 2 illustrates an exemplary user interface display in accordance with certain embodiments of the invention.
  • Fig. 3 illustrates an exemplary display of an item list and current location marker such as may be used in accordance with certain embodiments of the invention.
  • Figs. 4A and 4B illustrates different ways in which the prompt in a sequence can be changed.
  • Fig. 5 is a flowchart illustrating an exemplary process for displaying the current status of tasks in accordance with certain embodiments of the invention.
  • Figs. 6 and 7 illustrate alternative displays of the item list and current location identifiers with reference to a sequence of tasks to be completed in order to record a new inspection (e.g., a building inspection).
  • a new inspection e.g., a building inspection
  • Fig. 8 illustrates an exemplary distributed environment in which the status of tasks being performed by multiple users can be monitored.
  • Fig. 9 illustrates an exemplary group of lists that may be displayed for the distributed environment of Fig. 8. DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • a list of items or prompts that is to be traversed by a user in a particular order is displayed to the user (e.g., a set of tasks the user is to perform in a particular sequence as part of his or her job, a set of words to be spoken, a list of questions or fields to be answered, and so forth). At least a portion of the list is displayed at any given time along with an indication of which item in the list is the next item that the user needs to handle (e.g., the next task to perform, the next word to speak, the next question to answer, and so forth).
  • the current status of his or her progression through the prompts on the list is dynamically updated so as to readily inform the user as to what the current task is that needs to be performed, as well as what tasks have already been performed and/or what tasks remain to be performed.
  • Fig. 1 illustrates an exemplary computing device 100 such as may be used in accordance with certain embodiments of the invention.
  • Computing device 100 represents a wide variety of computing devices, such as wearable computers, personal digital assistants (PDAs), handheld or pocket computers, telephones (e.g., cell phones), laptop computers, gaming consoles or portable gaming devices, desktop computers, Internet appliances, etc.
  • PDAs personal digital assistants
  • telephones e.g., cell phones
  • laptop computers e.g., cell phones
  • gaming consoles or portable gaming devices e.g., desktop computers, Internet appliances, etc.
  • the dynamic displaying of current status of tasks described herein is particularly useful if computing device 100 has a small display, any size display may be used with the invention.
  • Computing device 100 includes a central processing unit (CPU) 102, memory 104, a storage device 106, one or more input controllers 108, and one or more output controllers 1 10 (alternatively, a single controller may be used for both input and output) coupled together via a bus 1 12.
  • Bus 1 12 represents one or more conventional computer buses, including a processor bus, system bus, accelerated graphics port (AGP), universal serial bus (USB), peripheral component interconnect bus (PCI), etc.
  • Memory 104 may be implemented using volatile and/or non-volatile memory, such as random access memory (RAM), read only memory (ROM), Flash memory, electronically erasable programmable read only memory (EEP OM), disk, and so forth.
  • Storage device 106 is typically implemented using non-volatile "permanent" memory, such as ROM, EEPROM, magnetic or optical diskette, memory cards, and the like.
  • Input controller(s) 108 are coupled to receive inputs from one or more input devices 1 14.
  • Input devices 114 include any of a variety of conventional input devices, such as a microphone, voice recognition devices, traditional qwerty keyboards, chording keyboards, half qwerty keyboards, dual forearm keyboards, chest mounted keyboards, handwriting recognition and digital ink devices, a mouse, a track pad, a digital stylus, a finger or glove device to capture user movement, pupil tracking devices, a gyropoint, a trackball, a voice grid device, digital cameras (still and motion), and so forth.
  • Output controller(s) 1 10 are coupled to output data to one or more output devices 1 16.
  • Output devices 116 include any of a variety of conventional output devices, such as a display device (e.g., a hand-held flat panel display, an eyeglass- mounted display that allows the user to view the real world surroundings while simultaneously overlaying or otherwise presenting information to the user in an unobtrusive manner), a speaker, an olfactory output device, tactile output devices, and so forth.
  • One or more application programs 118 are stored in memory 104 and executed by CPU 102. When executed, application programs 1 18 generate data that may be output to the user via one or more of the output devices 1 16 and also receive data that may be input by the user via one or more of the input devices 1 14. For discussion purposes, one particular application program is illustrated with a user interface (UI) component 120 that is designed to present information to the user including dynamically displaying the current status of tasks as discussed in more detail below.
  • UI user interface
  • the invention may be implemented in hardware or a combination of hardware, software, and/or firmware.
  • ASICs application specific integrated circuits
  • Fig. 2 illustrates an exemplary user interface display in accordance with certain embodiments of the invention.
  • User interface display 150 can be, for example, the display generated by user interface 120 of Fig. 1.
  • UI display 150 includes an item or prompt list portion 152, a user choices portion 154, and an applet window portion 156. Additional labels or prompts 158 may also be included (e.g., a title for the task being handled, the current time, the amount of time left to finish the task, etc.).
  • List portion 152 displays a list that prompts the user of tasks that are to be handled by the user in a particular order.
  • the manner in which an item or task is handled by the user is dependent on the nature of the list, as discussed in more detail below.
  • User choices portion 154 displays the options for the user to select from based on the next item or task in the list that needs to be handled by the user. For example, assume that the list in portion 152 is a list prompting the user regarding what information needs to be gathered in order for the user to set up a meeting with a potential customer.
  • the list of prompts in list portion 152 could be a list of tasks the user must perform — that is, a list of information that needs to be collected (e.g., the customer's name, the location of the meeting, the time of the meeting, and so forth). If we further assume that the current task that needs to be handled by the user is entry of the location of the meeting, user choices portion 154 could display the various permissible inputs for the location of the meeting (e.g., at the user's main office, at a remote office, at the customer's facility, and so forth).
  • the item list may be a list of prompts for the information to be verbally input by the user in each step, with user choices portion 154 displaying a list of which words can be spoken in each step.
  • Applet window portion 156 displays additional information clarifying or amplifying the choices in user choices portion 154 (or the current item or task in item list portion 152). Following the previous example, if the current task that needs to be handled by the user is entry of the location of the meeting, applet window portion 156 could display additional descriptive information for one or more of the permissible inputs for the location of the meeting (e.g., a street address, a distance from the user's home, a map flagging the locations of the various meeting locations, and so forth).
  • additional descriptive information for one or more of the permissible inputs for the location of the meeting e.g., a street address, a distance from the user's home, a map flagging the locations of the various meeting locations, and so forth.
  • the list displayed in list portion 152 is a list of items that is to be traversed by a user in a particular order. This can be a list of task prompts regarding tasks that the user is to perform, a list of tasks prompts regarding tasks to be performed by another user or computer, and so forth.
  • any of a wide variety of lists can be displayed, such as a set of tasks the user is to perform in a particular sequence as part of his or her job (this can be used, for example, to assist in training users to do their jobs), a set of tasks the user is to perform in a particular sequence in order to assemble or install a product he or she has purchased, a set of words to be spoken (e.g., queues as to what voice inputs the user is to make in order to carry out a task), a list of questions or fields to be answered, and so forth.
  • the list of items may be a list of tasks or steps to be performed by a computer or computer program.
  • Such a list can be used, for example, by a user to track the process of the computer or program in carrying out the particular sequence of steps.
  • multiple lists of items may be displayed (e.g., a multi-tiered item list). Situations can arise in which the list of items or prompts is too large to be displayed in its entirety. In such situations, only a portion of the list is displayed (e.g., centered on the item or prompt for the next task to be performed). This subset of the steps to be performed is then scrolled as tasks are completed, resulting in a dynamic list display that changes when a task is completed.
  • the user By displaying the list of prompts (or at least a portion thereof), the user is able to readily identify the status of the set of tasks being performed (in other words, the user is also able to obtain a feel for where he or she is (or where the user or computer being monitored is) in progressing through the sequence of tasks).
  • the user is able to quickly identify one or more previous tasks (if any) in the sequence, as well as one or more future tasks (if any) in the sequence.
  • Such information is particularly helpful in reorienting the user to the sequence of tasks if his or her attention has been diverted away from the sequence. For example, the user's attention may be diverted away from the sequence to answer questions from another employee.
  • Item lists may be a set of predetermined items, such as a particular set of steps to be followed to assemble a machine or a set of words to be uttered to carry- out a task for a speech-recognizing computer.
  • item lists may be dynamic, changing based on the user's current location, current activity, past behavior, etc.
  • computer 100 of Fig. 1 may detect where the user is currently located (e.g., in his or her office, in the assembly plant, which assembly plant, etc.), and provide the appropriate instructions to perform a particular task based on that current location.
  • Fig. 3 illustrates an exemplary display of an item list and current location marker such as may be used in accordance with certain embodiments of the invention.
  • the sequence of items on the list is a set of prompts regarding information that needs to be supplied by the user in order to schedule a meeting.
  • this list includes the following information: who the meeting is with (who), the date and time for the meeting (when), the duration of the meeting (how long), the location of the meeting (where), an indication of any materials to bring to the material (bring), and an indication of anyone else that should be notified of the meeting (cc).
  • Fig. 3 illustrates an example item list displayed in list portion 152 of Fig. 2.
  • the item list 170 is displayed, including the following prompts: "who?", “when?”, “how long?", “where?", and "bring?”.
  • the prompts in list 170 provide a quick identification to the user of what information he or she needs to input for each task in the sequence of tasks for scheduling a meeting. Due to the limited display area, list 170 does not include the prompts for each step in the sequence, but rather scrolls through the prompts as discussed in more detail below.
  • a current location marker 172 is also illustrated in Fig. 3 to identify to the user what the current step is in the sequence.
  • the first step in the sequence is to identify who the meeting is with (who), which is identified by current location marker 172 being situated above the prompt "who?".
  • location marker 172 is a circle or ball.
  • other types of presentation changes may be made to alter the appearance of a prompt (or area surrounding a prompt) in order to distinguish the current step from other steps in the sequence.
  • the text for the prompt may be altered (e.g., a different color, a different font, a different size, a different position on screen (e.g., slightly higher or lower than other prompts in the list), and so forth), the display around the prompt may be altered (e.g., the prompt may be inverted so that it appears white on a black background rather than the more traditional black on a white background, the prompt may be highlighted, the prompt may be encircled by a border, and so forth), etc.
  • Those skilled in the art can easily determine a variety of alternate methods for marking the current step.
  • One additional presentation change that can be made to distinguish the current step from other steps in the sequence is to change the prompt itself.
  • the prompt could be replaced with another prompt, or another prompt could be superimposed on the prompt for the current step.
  • the user may have a set of individuals that he or she typically meets with, and these may be superimposed on the "who?" prompt when it is the current step.
  • Figs. 4A — 4B illustrates different ways in which the prompt in a sequence can be changed.
  • Fig. 4A illustrates an example item list with the prompt for the current step in the sequence being superimposed with various input options.
  • a list 190 is illustrated and the current step is to input who the meeting is to be with (the "who?" prompt).
  • a set of common people that the user schedules meetings with (Jane, David, Lisa, and Richard) are superimposed on the "who?" prompt.
  • the appearance of the underlying prompt "who?" may be changed (e.g., shadowed out, different color, etc.) in order for overlying input options to be more easily viewed. It is to be appreciated that the exact location of the superimposed set of input options can vary (e.g., the characters of one or more input options may overlap the prompt, or be separated from the prompt).
  • Fig. 4B illustrates an example item list with the prompt for the current step in the sequence being replaced by the set of input options.
  • a list 192 is illustrated and the current step is to input who the meeting is to be with (the "who?" prompt). However, as illustrated, the "who?" prompt is replaced with a set of common people that the user schedules meetings with (Jane, David, Lisa, and Richard).
  • the user is thus given an indication of both the current step in the sequence as well as common responses to that step.
  • the type of information that is superimposed on or replaces the prompt can vary based on the current step. For example, when the "when?" prompt is the current step it may have superimposed thereon the times that the user is available for the current day (or current week, and so forth).
  • list 170 is changed to list 174 in which the prompt "who?" is replaced with the name "Bob Smith” and the current location marker 172 is changed to indicate the next prompt ("when?") is the current task that needs to be handled by the user.
  • list 174 is changed to list 176 in which the prompt "when?” is replaced with the date and time of the meeting, and the current location marker 172 is changed to indicate the next prompt ("how long?") is the current task that needs to be handled by the user.
  • the current location marker 172 "bounces" along the list from item to item, making the user readily aware of what the current task is that he or she should be performing (that is, which data he or she should be inputting in the present example).
  • list 176 is changed to list 178.
  • the user interface now scrolls the list so that the leftmost item is no longer shown but a new item is added at the right.
  • the identification of "Bob Smith” is no longer shown, but a prompt for who else should be notified of the meeting ("cc?") is now shown.
  • list 178 is changed to list 180 and current location marker 172 is changed to indicate the next prompt (“bring?") is the current task that needs to be handled by the user.
  • current location marker 172 may not be moved in response to an input but the list may be scrolled.
  • the item list provides a series of prompts identifying what tasks (if any) in the sequence have already been performed and what tasks (if any) remain to be performed. For those tasks that have already been performed, an indication is made in the list as to what action was taken by the user for those tasks (e.g., what information was entered by the user in the illustrated example).
  • the prompts in the list need not be changed when the user enters the data (e.g., "who?" need not be replaced by "Bob Smith").
  • the data input by the user can alternatively be displayed elsewhere (e.g., in applet window portion 156).
  • the user is able to select and zoom in on the "where?" prompt and have additional information about that task displayed (e.g., the possible locations for the meeting).
  • the user is able to "backtrack” through the list (e.g., by moving a cursor to the desired item and selecting it, or using a back arrow key or icon, or changing the current location marker (e.g., dragging and dropping the location marker to the desired item), etc.) and see this additional information for tasks already completed.
  • the "backtracking" may be for navigational rather than informational purposes. Moving back through the list (whether by manipulation of the location marker or in some other manner) may also be used to accomplish other types of operations, such as defining a macro or annotation.
  • the speed of handling of the sequence of the items by the user can potentially be increased. For example, the user can see the prompt for the next one or more items in the list and begin thinking about how he or she is going to handle that particular item even before the computing device is finished processing the input for the item he or she just handled.
  • multiple location markers are displayed along with the item list - one marker identifying the current item to be handled by the user and another marker identifying the current item being processed by the computing device.
  • Situations can arise where the user can input data quicker than it can be processed by the computing device. For example, the user may be able to talk at a faster rate than the computing device is able to analyze the speech.
  • the use of two such markers can allow the user to identify if the computing device is hung up on or having difficulty processing a particular input (e.g., identify a particular word spoken by the user, misrecognition of the input, improper parsing, etc.), the user can identify this situation and go back to the task the computing device is having difficulty processing and re-enter the speech.
  • identify a particular word spoken by the user e.g., identify a particular word spoken by the user, misrecognition of the input, improper parsing, etc.
  • Fig. 5 is a flowchart illustrating an exemplary process for displaying the current status of tasks in accordance with certain embodiments of the invention.
  • the process of Fig. 5 is carried out by the user interface of a computing device (e.g., interface 120 of Fig. 1), and may be performed in software.
  • a computing device e.g., interface 120 of Fig. 1
  • Fig. 5 is discussed with reference to a location marker, it is to be appreciated that any of the presentation changes discussed above an be used to identify items in the list.
  • an item list is displayed (act 200), which is a sequence of items or prompts for the user to follow.
  • a current location marker is also displayed to identify the first item in the list (act 202), and input corresponding to the first item in the list is received (act 204).
  • this input can vary depending on the sequence of tasks itself (e.g., it may be data input by a user, an indication from another computer program that the task has been accomplished, etc.).
  • a check is then made as to whether the end of the list has been reached (at 206). If the end of the list has been reached then the process stops (act 208), waiting for the next sequence of tasks to begin or for the user to backtrack to a previously completed task.
  • the item list is scrolled by one item (or alternatively more items) in the appropriate direction (act 212).
  • the amount that the item list is scrolled can vary (e.g., based on the sizes of the different items in the list).
  • the appropriate direction for scrolling can vary based on the activity being performed by the user and the layout of the list (e.g., in the example of Fig. 3, the scrolling is from right to left when progressing forward through the list, and left to right when backtracking through the list). Regardless of whether the ordered item list is scrolled, after act 210 or 212 the current location marker is moved as necessary to identify the next item in the list that is to be handled by the user (act 214).
  • movement of the current location marker may not be necessary due to the scrolling performed (e.g., as illustrated with reference to lists 176 and 178 in Fig. 3).
  • user input is received corresponding to the identified next item in the list (act 216). The process then returns to determine whether the end of the list has been reached (act 206).
  • the item list and current location identifier or marker can be displayed in a wide variety of different manners.
  • Figs. 6 and 7 illustrate alternative displays of the item list and current location identifiers with reference to a sequence of tasks to be completed in order to record a new inspection (e.g., a building inspection).
  • a new inspection e.g., a building inspection.
  • the item list portion 242 includes a list of tasks that are to be handled by the user, each of which is information to be entered by the user. Once entered, the information is displayed in applet window portion 244.
  • a current location marker 246 advances down the list in portion 242 to identify the current information that the user needs to input (the customer's state in the illustrated display). Additional information is displayed at the top of display 240, including a prompt 248 identifying a type of information being entered by the user (inspection information).
  • a multi-tiered item list is displayed including list portion 262 and list portion 264.
  • list portion 262 prompts for the overall process of recording a new inspection are listed, including selecting a new inspection option and then entering inspection information.
  • Two current location markers 266 and 268 are illustrated, each providing a visual indication of where in the overall process the current user is (inspection info in the illustrated display).
  • a prompt 270 provides a further identification to the user of where he or she is in the overall process.
  • List portion 264 includes prompts for the process of entering inspection information, with a current location marker 272 providing a visual indication of where in the inspection information entry process the user currently is (customer state in the illustrated display).
  • the dynamic displaying of the current status of tasks of the present invention can further be used to track the status of tasks being performed by multiple users.
  • information indicating the status of tasks being performed by multiple users is communicated back to the computing devices of one or more other users, who in turn can view the status information of multiple users on a single display.
  • Fig. 8 illustrates an exemplary distributed environment in which the status of tasks being performed by multiple users can be monitored.
  • multiple users Jamie, John, Max, and Carol each have a wearable computer with an eyeglass display 300, 302, 304, and 306, respectively.
  • An item list is displayed on the eyeglass display for each of these users, with a current location marker to identify to the respective users where they are in the task sequences they are performing.
  • Information regarding their current location is also communicated to another computing device of their supervisor Jane, who is also wearing an eyeglass display 308.
  • the information communicated to Jane's computer can be simply an identification of the current location (e.g., Jane's computer may already be programmed with all of the tasks in the list), or alternatively the entire (or at least a portion of) the item list.
  • the information for one or more of the users Jamie, John, Max, and Carol can then be displayed on display 308, allowing Jane to keep track of the status of each of the users Jamie, John, Max, and Carol in performing their tasks. This allows Jane, as the supervisor, to see if people are proceeding through their tasks too quickly or too slowly (e.g., a user may be having difficulty and need assistance), to know when the individual users will be finished with their tasks, etc.
  • Fig. 9 illustrates an exemplary group of lists that may be displayed on eyeglass display 308 of Fig. 8. Assume that each of the users John, Jamie, Max, and Carol are each performing a machine assembly process involving the following tasks: inventory the necessary parts, assemble an intake, lubricate a core part of the machine, install the assembled intake, verify that the batteries are fully charged, and then run a diagnostic program. The tasks in the machine assembly process are illustrated in a portion 310 of display 308 in an abbreviated form.
  • the tasks illustrated in portion 310 may not be abbreviated, or may be represented in some other manner (e.g., as icons).
  • a separate item list is displayed on display 308 for each of the users along with a corresponding current location marker in the shape of a ball or circle.
  • the viewer of display 308 can readily identify that John is at the "assemble intake” step, Jamie and Max are both at the “install intake” step, and Carol is at the “verify charge” step.
  • the supervisor viewing display 308 can quickly and easily determine, based on the item list and current location markers, that each of Jamie, Max, and Carol is proceeding normally through the assembly process, but that John is hung up on the "assemble intake” step, so the supervisor can check with John to see if he is experiencing difficulties with this step.

Abstract

The current status of a list of tasks to be performed is dynamically displayed. The tasks may be performed by a user (e.g., data entered by the user, words spoken by the user, actions taken by the user, and so forth) or alternatively by a computer (e.g., the steps it follows in carrying out a programmed task). At least a portion of the list is displayed at any given time along with an indication of which task is the next task to be performed. As the tasks are completed, the current status of the progression through the items on the list is dynamically updated so as to readily inform the user (or someone else) as to what the current task is that needs to be performed, as well as what tasks have already been performed and/or what tasks remain to be performed.

Description

DYNAMICALLY DISPLAYING CURRENT STATUS OF TASKS
RELATED APPLICATIONS
A claim of priority is made to U.S. Provisional Application No. 60/240,685, filed October 16, 2000, entitled "Method for Dynamically Displaying the Current Status of Tasks".
TECHNICAL FIELD
The present invention is directed to graphical user interfaces and more particularly to dynamically displaying the current status of tasks.
BACKGROUND OF THE INVENTION
As computers become increasingly powerful and commonplace, they are being used for an increasingly broad variety of tasks. For example, in addition to traditional activities such as running word processing and database applications, computers are increasingly becoming an integral part of users' daily lives. Programs to schedule activities, generate reminders, and provide rapid communication capabilities are becoming increasingly popular. Moreover, computers are increasingly present during virtually all of a person's daily activities. For example, hand-held computer organizers (e.g., PDAs) are increasingly common, and communication devices such as portable phones are increasingly incorporating computer capabilities. More recently, the field of wearable computers (e.g., with eyeglass displays) has begun to expand, creating a further presence of computers in people's daily lives. Computers often progress through a particular series of steps when allowing a user to accomplish a particular task. For example, if a user desires to enter a new name and address to an electronic address book, the computer progresses through a series of steps prompting the user to enter the desired information (e.g., name, street address, city, state, zip code, phone number, etc.). On computers with large displays (e.g., typical desktop computers), sufficient area exists on the display to provide an informative and useable user interface (UI) that allows the user to enter the necessary data for the series of steps. However, problems exist when attempting to guide the user through the particular series of steps on smaller displays. Without the large display area, there is frequently insufficient room to provide the prompts in the same informative and useable manner. Additionally, the nature of many new computing devices with small displays
(e.g., PDAs and wearable computers) is that the computing devices are transported with the user. However, traditional computer programs are not typically designed to efficiently present information to users in a wide variety of environments. For example, most computer programs are designed with a prototypical user being seated at a stationary computer with a large display device, and with the user devoting full attention to the display. In that environment, the computer program can be designed with the assumption that the user's attention is predominately on the display device. However, many new computing devices with small displays can be used when the user's attention is more likely to be diverted to some other task (e.g., driving, using machinery, walking, etc.). Many traditional computer programs, designed with large display devices in mind, frequently do not allow the user to quickly and easily reorient him-or her-self to the task being carried out by the computer. For example, if the user is performing a task by following a series of steps on a wearable computer, looks away from the display to focus his or her attention on crossing a busy intersection, and then returns to the task, it would be desirable for the user to be able to quickly and easily reorient him- or her-self to the task (in other words, readily know what steps he or she has accomplished so far and what the next step to be performed is).
Accordingly, there is a need for new techniques to display the current status of tasks to a user.
SUMMARY OF THE INVENTION
Dynamically displaying current status of tasks is described herein. According to one aspect, a list of items corresponding to tasks that are to be performed are displayed. The tasks may be performed by a user (e.g., data entered by the user, words spoken by the user, actions taken by the user, and so forth) or alternatively by a computer (e.g., the steps followed in carrying out a programmed task). At least a portion of the list is displayed at any given time along with an indication of which task is the next task to be performed. As the user progresses through the set of tasks, the current status of his or her progression through the corresponding items on the list is dynamically updated so as to readily inform the user (or someone else) as to what the current task is that needs to be performed, as well as what tasks have already been performed and/or what tasks remain to be performed.
According to another aspect, only a subset of the list of items is displayed at any given time. The list is scrolled through as the tasks are performed so that different items are displayed as part of the subset as tasks are performed.
According to another aspect, multiple lists of tasks to be performed by multiple individuals (or computing devices) are displayed on a display of the user. As the multiple individuals (or computing devices) finish the tasks in their respective lists, an indication of such completion is forwarded to the user's computer, which updates the display to indicate the next task in the list to be displayed. The user is thus able to monitor the progress of the multiple individuals (or computing devices) in carrying out their respective tasks.
BRIEF DESCRIPTION OF THE DRAWINGS
The present invention is illustrated by way of example and not limitation in the figures of the accompanying drawings. The same numbers are used throughout the figures to reference like components and/or features.
Fig. 1 illustrates an exemplary computing device such as may be used in accordance with certain embodiments of the invention. Fig. 2 illustrates an exemplary user interface display in accordance with certain embodiments of the invention.
Fig. 3 illustrates an exemplary display of an item list and current location marker such as may be used in accordance with certain embodiments of the invention. Figs. 4A and 4B illustrates different ways in which the prompt in a sequence can be changed.
Fig. 5 is a flowchart illustrating an exemplary process for displaying the current status of tasks in accordance with certain embodiments of the invention.
Figs. 6 and 7 illustrate alternative displays of the item list and current location identifiers with reference to a sequence of tasks to be completed in order to record a new inspection (e.g., a building inspection).
Fig. 8 illustrates an exemplary distributed environment in which the status of tasks being performed by multiple users can be monitored.
Fig. 9 illustrates an exemplary group of lists that may be displayed for the distributed environment of Fig. 8. DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
Dynamically displaying the current status of tasks is described herein. A list of items or prompts that is to be traversed by a user in a particular order is displayed to the user (e.g., a set of tasks the user is to perform in a particular sequence as part of his or her job, a set of words to be spoken, a list of questions or fields to be answered, and so forth). At least a portion of the list is displayed at any given time along with an indication of which item in the list is the next item that the user needs to handle (e.g., the next task to perform, the next word to speak, the next question to answer, and so forth). As the user progresses through the list of tasks, the current status of his or her progression through the prompts on the list is dynamically updated so as to readily inform the user as to what the current task is that needs to be performed, as well as what tasks have already been performed and/or what tasks remain to be performed.
Fig. 1 illustrates an exemplary computing device 100 such as may be used in accordance with certain embodiments of the invention. Computing device 100 represents a wide variety of computing devices, such as wearable computers, personal digital assistants (PDAs), handheld or pocket computers, telephones (e.g., cell phones), laptop computers, gaming consoles or portable gaming devices, desktop computers, Internet appliances, etc. Although the dynamic displaying of current status of tasks described herein is particularly useful if computing device 100 has a small display, any size display may be used with the invention.
Computing device 100 includes a central processing unit (CPU) 102, memory 104, a storage device 106, one or more input controllers 108, and one or more output controllers 1 10 (alternatively, a single controller may be used for both input and output) coupled together via a bus 1 12. Bus 1 12 represents one or more conventional computer buses, including a processor bus, system bus, accelerated graphics port (AGP), universal serial bus (USB), peripheral component interconnect bus (PCI), etc.
Memory 104 may be implemented using volatile and/or non-volatile memory, such as random access memory (RAM), read only memory (ROM), Flash memory, electronically erasable programmable read only memory (EEP OM), disk, and so forth. Storage device 106 is typically implemented using non-volatile "permanent" memory, such as ROM, EEPROM, magnetic or optical diskette, memory cards, and the like.
Input controller(s) 108 are coupled to receive inputs from one or more input devices 1 14. Input devices 114 include any of a variety of conventional input devices, such as a microphone, voice recognition devices, traditional qwerty keyboards, chording keyboards, half qwerty keyboards, dual forearm keyboards, chest mounted keyboards, handwriting recognition and digital ink devices, a mouse, a track pad, a digital stylus, a finger or glove device to capture user movement, pupil tracking devices, a gyropoint, a trackball, a voice grid device, digital cameras (still and motion), and so forth.
Output controller(s) 1 10 are coupled to output data to one or more output devices 1 16. Output devices 116 include any of a variety of conventional output devices, such as a display device (e.g., a hand-held flat panel display, an eyeglass- mounted display that allows the user to view the real world surroundings while simultaneously overlaying or otherwise presenting information to the user in an unobtrusive manner), a speaker, an olfactory output device, tactile output devices, and so forth.
One or more application programs 118 are stored in memory 104 and executed by CPU 102. When executed, application programs 1 18 generate data that may be output to the user via one or more of the output devices 1 16 and also receive data that may be input by the user via one or more of the input devices 1 14. For discussion purposes, one particular application program is illustrated with a user interface (UI) component 120 that is designed to present information to the user including dynamically displaying the current status of tasks as discussed in more detail below.
Although discussed herein primarily with reference to software components and modules, the invention may be implemented in hardware or a combination of hardware, software, and/or firmware. For example, one or more application specific integrated circuits (ASICs) could be designed or programmed to carry out the invention.
Fig. 2 illustrates an exemplary user interface display in accordance with certain embodiments of the invention. User interface display 150 can be, for example, the display generated by user interface 120 of Fig. 1. UI display 150 includes an item or prompt list portion 152, a user choices portion 154, and an applet window portion 156. Additional labels or prompts 158 may also be included (e.g., a title for the task being handled, the current time, the amount of time left to finish the task, etc.). List portion 152 displays a list that prompts the user of tasks that are to be handled by the user in a particular order. An indication is also made to the user within list portion 152 of where the user currently is in that list (that is, what the next item or task is that needs to be handled by the user), and also identifies items or tasks (if any) that have already been handled by the user as well as future items or tasks (if any) that need to be handled by the user. The manner in which an item or task is handled by the user is dependent on the nature of the list, as discussed in more detail below. User choices portion 154 displays the options for the user to select from based on the next item or task in the list that needs to be handled by the user. For example, assume that the list in portion 152 is a list prompting the user regarding what information needs to be gathered in order for the user to set up a meeting with a potential customer. The list of prompts in list portion 152 could be a list of tasks the user must perform — that is, a list of information that needs to be collected (e.g., the customer's name, the location of the meeting, the time of the meeting, and so forth). If we further assume that the current task that needs to be handled by the user is entry of the location of the meeting, user choices portion 154 could display the various permissible inputs for the location of the meeting (e.g., at the user's main office, at a remote office, at the customer's facility, and so forth). By way of another example, the item list may be a list of prompts for the information to be verbally input by the user in each step, with user choices portion 154 displaying a list of which words can be spoken in each step.
Applet window portion 156 displays additional information clarifying or amplifying the choices in user choices portion 154 (or the current item or task in item list portion 152). Following the previous example, if the current task that needs to be handled by the user is entry of the location of the meeting, applet window portion 156 could display additional descriptive information for one or more of the permissible inputs for the location of the meeting (e.g., a street address, a distance from the user's home, a map flagging the locations of the various meeting locations, and so forth).
The list displayed in list portion 152 is a list of items that is to be traversed by a user in a particular order. This can be a list of task prompts regarding tasks that the user is to perform, a list of tasks prompts regarding tasks to be performed by another user or computer, and so forth. Any of a wide variety of lists can be displayed, such as a set of tasks the user is to perform in a particular sequence as part of his or her job (this can be used, for example, to assist in training users to do their jobs), a set of tasks the user is to perform in a particular sequence in order to assemble or install a product he or she has purchased, a set of words to be spoken (e.g., queues as to what voice inputs the user is to make in order to carry out a task), a list of questions or fields to be answered, and so forth. Alternatively, the list of items may be a list of tasks or steps to be performed by a computer or computer program. Such a list can be used, for example, by a user to track the process of the computer or program in carrying out the particular sequence of steps. Additionally, depending on the nature of the sequence of tasks being performed, multiple lists of items may be displayed (e.g., a multi-tiered item list). Situations can arise in which the list of items or prompts is too large to be displayed in its entirety. In such situations, only a portion of the list is displayed (e.g., centered on the item or prompt for the next task to be performed). This subset of the steps to be performed is then scrolled as tasks are completed, resulting in a dynamic list display that changes when a task is completed. By displaying the list of prompts (or at least a portion thereof), the user is able to readily identify the status of the set of tasks being performed (in other words, the user is also able to obtain a feel for where he or she is (or where the user or computer being monitored is) in progressing through the sequence of tasks). The user is able to quickly identify one or more previous tasks (if any) in the sequence, as well as one or more future tasks (if any) in the sequence. Such information is particularly helpful in reorienting the user to the sequence of tasks if his or her attention has been diverted away from the sequence. For example, the user's attention may be diverted away from the sequence to answer questions from another employee. After answering the question, the user can look back at display 150 and quickly reorient him- or her-self into the sequence of tasks being performed. Item lists may be a set of predetermined items, such as a particular set of steps to be followed to assemble a machine or a set of words to be uttered to carry- out a task for a speech-recognizing computer. Alternatively, item lists may be dynamic, changing based on the user's current location, current activity, past behavior, etc. For example, computer 100 of Fig. 1 may detect where the user is currently located (e.g., in his or her office, in the assembly plant, which assembly plant, etc.), and provide the appropriate instructions to perform a particular task based on that current location. Additional information regarding detecting the user's current context (e.g., current location, current activity, etc.) can be found in a co-pending U.S. Patent Application Serial No. 09/216,193, entitled "Method and System For Controlling Presentation of Information To a User Based On The User's Condition", which was filed December 18, 1998, and is commonly assigned to Tangis Corporation. This application is hereby incorporated by reference.
Fig. 3 illustrates an exemplary display of an item list and current location marker such as may be used in accordance with certain embodiments of the invention. Assume that the sequence of items on the list is a set of prompts regarding information that needs to be supplied by the user in order to schedule a meeting. In the illustrated example, this list includes the following information: who the meeting is with (who), the date and time for the meeting (when), the duration of the meeting (how long), the location of the meeting (where), an indication of any materials to bring to the material (bring), and an indication of anyone else that should be notified of the meeting (cc).
Fig. 3 illustrates an example item list displayed in list portion 152 of Fig. 2. Initially, the item list 170 is displayed, including the following prompts: "who?", "when?", "how long?", "where?", and "bring?". The prompts in list 170 provide a quick identification to the user of what information he or she needs to input for each task in the sequence of tasks for scheduling a meeting. Due to the limited display area, list 170 does not include the prompts for each step in the sequence, but rather scrolls through the prompts as discussed in more detail below. A current location marker 172 is also illustrated in Fig. 3 to identify to the user what the current step is in the sequence. Assuming the meeting scheduling process has just begun, the first step in the sequence is to identify who the meeting is with (who), which is identified by current location marker 172 being situated above the prompt "who?". In the illustrated example, location marker 172 is a circle or ball. Alternatively, other types of presentation changes may be made to alter the appearance of a prompt (or area surrounding a prompt) in order to distinguish the current step from other steps in the sequence. For example, different shapes other than a circle or ball may be used for a location marker, the text for the prompt may be altered (e.g., a different color, a different font, a different size, a different position on screen (e.g., slightly higher or lower than other prompts in the list), and so forth), the display around the prompt may be altered (e.g., the prompt may be inverted so that it appears white on a black background rather than the more traditional black on a white background, the prompt may be highlighted, the prompt may be encircled by a border, and so forth), etc. Those skilled in the art can easily determine a variety of alternate methods for marking the current step. One additional presentation change that can be made to distinguish the current step from other steps in the sequence is to change the prompt itself. The prompt could be replaced with another prompt, or another prompt could be superimposed on the prompt for the current step. For example, the user may have a set of individuals that he or she typically meets with, and these may be superimposed on the "who?" prompt when it is the current step. Figs. 4A — 4B illustrates different ways in which the prompt in a sequence can be changed. Fig. 4A illustrates an example item list with the prompt for the current step in the sequence being superimposed with various input options. A list 190 is illustrated and the current step is to input who the meeting is to be with (the "who?" prompt). As illustrated, a set of common people that the user schedules meetings with (Jane, David, Lisa, and Richard) are superimposed on the "who?" prompt. The appearance of the underlying prompt "who?" may be changed (e.g., shadowed out, different color, etc.) in order for overlying input options to be more easily viewed. It is to be appreciated that the exact location of the superimposed set of input options can vary (e.g., the characters of one or more input options may overlap the prompt, or be separated from the prompt).
Fig. 4B illustrates an example item list with the prompt for the current step in the sequence being replaced by the set of input options. A list 192 is illustrated and the current step is to input who the meeting is to be with (the "who?" prompt). However, as illustrated, the "who?" prompt is replaced with a set of common people that the user schedules meetings with (Jane, David, Lisa, and Richard).
The user is thus given an indication of both the current step in the sequence as well as common responses to that step. The type of information that is superimposed on or replaces the prompt can vary based on the current step. For example, when the "when?" prompt is the current step it may have superimposed thereon the times that the user is available for the current day (or current week, and so forth).
Returning to Fig. 3, once the user enters the information identifying who the meeting is with (assume for purposes of this example the meeting is with Bob Smith), list 170 is changed to list 174 in which the prompt "who?" is replaced with the name "Bob Smith" and the current location marker 172 is changed to indicate the next prompt ("when?") is the current task that needs to be handled by the user. Assuming the user inputs that the meeting is to occur at 10am on October 31 , list 174 is changed to list 176 in which the prompt "when?" is replaced with the date and time of the meeting, and the current location marker 172 is changed to indicate the next prompt ("how long?") is the current task that needs to be handled by the user. Thus, as can be seen from lists 172, 174, and 176, the current location marker 172 "bounces" along the list from item to item, making the user readily aware of what the current task is that he or she should be performing (that is, which data he or she should be inputting in the present example).
Once the user inputs the duration of the meeting, list 176 is changed to list 178. Given the limited display area, the user interface now scrolls the list so that the leftmost item is no longer shown but a new item is added at the right. Thus, the identification of "Bob Smith" is no longer shown, but a prompt for who else should be notified of the meeting ("cc?") is now shown. Once the user enters the location for the meeting ("home office"), list 178 is changed to list 180 and current location marker 172 is changed to indicate the next prompt ("bring?") is the current task that needs to be handled by the user. Thus, as can be seen with lists 176, 178, and 180, current location marker 172 may not be moved in response to an input but the list may be scrolled.
Thus, as can be seen in Fig. 3, the item list provides a series of prompts identifying what tasks (if any) in the sequence have already been performed and what tasks (if any) remain to be performed. For those tasks that have already been performed, an indication is made in the list as to what action was taken by the user for those tasks (e.g., what information was entered by the user in the illustrated example). Thus, the user can readily orient him- or her-self to the sequence of steps, even if his or her attention is diverted from the display for a period of time. Alternatively, the prompts in the list need not be changed when the user enters the data (e.g., "who?" need not be replaced by "Bob Smith"). The data input by the user can alternatively be displayed elsewhere (e.g., in applet window portion 156).
One advantage of the item lists described herein is that the lists present the multiple steps or items in a concise manner - these steps or items can also be referred to as idioms. When these idioms are presented together in a sequence, the provide more information to the user than when presented in independent form. For example, the idiom "bring?" by itself does not present as much information to the user as the entire sequence of idioms "who?", "when?", "how long?", "where?", and "bring?". The use of item lists as described herein also allows an individual to "zoom" in on (and thus gain more information about) a particular task. For example, with reference to Fig. 3, the user is able to select and zoom in on the "where?" prompt and have additional information about that task displayed (e.g., the possible locations for the meeting). The user is able to "backtrack" through the list (e.g., by moving a cursor to the desired item and selecting it, or using a back arrow key or icon, or changing the current location marker (e.g., dragging and dropping the location marker to the desired item), etc.) and see this additional information for tasks already completed. Alternatively, the "backtracking" may be for navigational rather than informational purposes. Moving back through the list (whether by manipulation of the location marker or in some other manner) may also be used to accomplish other types of operations, such as defining a macro or annotation.
Additionally, by displaying the prompts for future items, the speed of handling of the sequence of the items by the user can potentially be increased. For example, the user can see the prompt for the next one or more items in the list and begin thinking about how he or she is going to handle that particular item even before the computing device is finished processing the input for the item he or she just handled.
According to another embodiment, multiple location markers are displayed along with the item list - one marker identifying the current item to be handled by the user and another marker identifying the current item being processed by the computing device. Situations can arise where the user can input data quicker than it can be processed by the computing device. For example, the user may be able to talk at a faster rate than the computing device is able to analyze the speech.
The use of two such markers can allow the user to identify if the computing device is hung up on or having difficulty processing a particular input (e.g., identify a particular word spoken by the user, misrecognition of the input, improper parsing, etc.), the user can identify this situation and go back to the task the computing device is having difficulty processing and re-enter the speech.
Fig. 5 is a flowchart illustrating an exemplary process for displaying the current status of tasks in accordance with certain embodiments of the invention. The process of Fig. 5 is carried out by the user interface of a computing device (e.g., interface 120 of Fig. 1), and may be performed in software. Although Fig. 5 is discussed with reference to a location marker, it is to be appreciated that any of the presentation changes discussed above an be used to identify items in the list. Initially, an item list is displayed (act 200), which is a sequence of items or prompts for the user to follow. A current location marker is also displayed to identify the first item in the list (act 202), and input corresponding to the first item in the list is received (act 204). The nature of this input can vary depending on the sequence of tasks itself (e.g., it may be data input by a user, an indication from another computer program that the task has been accomplished, etc.). A check is then made as to whether the end of the list has been reached (at 206). If the end of the list has been reached then the process stops (act 208), waiting for the next sequence of tasks to begin or for the user to backtrack to a previously completed task.
However, if the end of the list has not been reached, then a check is made as to whether scrolling of the list is needed (act 210). Whether scrolling of the list is needed can be based on a variety of different factors. For example, the user interface may attempt to make sure that there are always at least a threshold number of prompts before and/or after the current location marker, the user interface may attempt to make sure that the current task remains as close to the center of the item list as is possible but that no portions of the item list be left empty, etc. These factors can optionally be user-configurable preferences, allowing the user to adjust the display to his or her particular likes and/or dislikes (e.g., the user may prefer to see more future tasks than previous tasks).
If scrolling is needed, then the item list is scrolled by one item (or alternatively more items) in the appropriate direction (act 212). The amount that the item list is scrolled can vary (e.g., based on the sizes of the different items in the list). The appropriate direction for scrolling can vary based on the activity being performed by the user and the layout of the list (e.g., in the example of Fig. 3, the scrolling is from right to left when progressing forward through the list, and left to right when backtracking through the list). Regardless of whether the ordered item list is scrolled, after act 210 or 212 the current location marker is moved as necessary to identify the next item in the list that is to be handled by the user (act 214). In some situations, movement of the current location marker may not be necessary due to the scrolling performed (e.g., as illustrated with reference to lists 176 and 178 in Fig. 3). At some point after the current location marker is moved (if necessary), user input is received corresponding to the identified next item in the list (act 216). The process then returns to determine whether the end of the list has been reached (act 206).
The item list and current location identifier or marker can be displayed in a wide variety of different manners. Figs. 6 and 7 illustrate alternative displays of the item list and current location identifiers with reference to a sequence of tasks to be completed in order to record a new inspection (e.g., a building inspection). In the exemplary display 240 of Fig. 6, an item list portion 242 and an applet window portion 244 are illustrated. The item list portion 242 includes a list of tasks that are to be handled by the user, each of which is information to be entered by the user. Once entered, the information is displayed in applet window portion 244. A current location marker 246 advances down the list in portion 242 to identify the current information that the user needs to input (the customer's state in the illustrated display). Additional information is displayed at the top of display 240, including a prompt 248 identifying a type of information being entered by the user (inspection information).
In the exemplary display 260 of Fig. 7, a multi-tiered item list is displayed including list portion 262 and list portion 264. In list portion 262, prompts for the overall process of recording a new inspection are listed, including selecting a new inspection option and then entering inspection information. Two current location markers 266 and 268 are illustrated, each providing a visual indication of where in the overall process the current user is (inspection info in the illustrated display). A prompt 270 provides a further identification to the user of where he or she is in the overall process. List portion 264 includes prompts for the process of entering inspection information, with a current location marker 272 providing a visual indication of where in the inspection information entry process the user currently is (customer state in the illustrated display). In addition to tracking the status of tasks being performed by a single user, the dynamic displaying of the current status of tasks of the present invention can further be used to track the status of tasks being performed by multiple users. In this situation, information indicating the status of tasks being performed by multiple users is communicated back to the computing devices of one or more other users, who in turn can view the status information of multiple users on a single display.
Fig. 8 illustrates an exemplary distributed environment in which the status of tasks being performed by multiple users can be monitored. In the illustrated example, multiple users Jamie, John, Max, and Carol each have a wearable computer with an eyeglass display 300, 302, 304, and 306, respectively. An item list is displayed on the eyeglass display for each of these users, with a current location marker to identify to the respective users where they are in the task sequences they are performing. Information regarding their current location is also communicated to another computing device of their supervisor Jane, who is also wearing an eyeglass display 308. The information communicated to Jane's computer can be simply an identification of the current location (e.g., Jane's computer may already be programmed with all of the tasks in the list), or alternatively the entire (or at least a portion of) the item list. The information for one or more of the users Jamie, John, Max, and Carol can then be displayed on display 308, allowing Jane to keep track of the status of each of the users Jamie, John, Max, and Carol in performing their tasks. This allows Jane, as the supervisor, to see if people are proceeding through their tasks too quickly or too slowly (e.g., a user may be having difficulty and need assistance), to know when the individual users will be finished with their tasks, etc. If a multi-tiered item list is being used, then the supervisor can also zoom in on the particular step of a user and get additional information regarding where the user is stuck. Fig. 9 illustrates an exemplary group of lists that may be displayed on eyeglass display 308 of Fig. 8. Assume that each of the users John, Jamie, Max, and Carol are each performing a machine assembly process involving the following tasks: inventory the necessary parts, assemble an intake, lubricate a core part of the machine, install the assembled intake, verify that the batteries are fully charged, and then run a diagnostic program. The tasks in the machine assembly process are illustrated in a portion 310 of display 308 in an abbreviated form. Alternatively, the tasks illustrated in portion 310 may not be abbreviated, or may be represented in some other manner (e.g., as icons). A separate item list is displayed on display 308 for each of the users along with a corresponding current location marker in the shape of a ball or circle. Thus, as illustrated in Fig. 9, the viewer of display 308 can readily identify that John is at the "assemble intake" step, Jamie and Max are both at the "install intake" step, and Carol is at the "verify charge" step. Thus, the supervisor viewing display 308 can quickly and easily determine, based on the item list and current location markers, that each of Jamie, Max, and Carol is proceeding normally through the assembly process, but that John is hung up on the "assemble intake" step, so the supervisor can check with John to see if he is experiencing difficulties with this step.
CONCLUSION
Although the description above uses language that is specific to structural features and/or methodological acts, it is to be understood that the invention defined in the appended claims is not limited to the specific features or acts described.
Rather, the specific features and acts are disclosed as exemplary forms of implementing the invention.

Claims

1. One or more computer-readable media storing a computer program that, when executed by one or more processors, causes the one or more processors to: display a subset of a plurality of steps in an order to be performed by a user; altering an appearance of a current step in the subset of steps that needs to be performed by the user to distinguish the current step from other steps in the subset; allowing the user to input data corresponding to the current step; and scrolling, in response to user input of data corresponding to the current step, the plurality of steps so that a new subset of the plurality of steps is presented to the user.
2. One or more computer-readable media as recited in claim 1, wherein the computer program further causes the one or more processors to: alter, in response to user input of data corresponding to the current step, the appearance of another step as necessary to identify the new current step in the subset of steps that needs to be performed by the user.
3. One or more computer-readable media as recited in claim 1 , wherein altering the appearance of the current step comprises marking the current location with a ball.
4. One or more computer-readable media as recited in claim 1 , wherein altering the appearance of the current step comprises displaying the current step differently than other steps in the subset.
5. One or more computer-readable media as recited in claim 1, wherein altering the appearance of the current step comprises replacing the current step with a set of one or more input options for the current step.
6. One or more computer- readable media as recited in claim 1, wherein altering the appearance of the current step comprises superimposing, on the current step, a set of one or more input options for the current step.
7. One or more computer-readable media as recited in claim 1, wherein the computer program further causes the one or more processors to: replace, in the subset, the display of the current step with a display of the input data.
8. One or more computer-readable media as recited in claim 1, wherein the computer program further causes the one or more processors to: display a current processing marker that identifies which step in the subset of steps is currently being processed by the one or more processors.
9. One or more computer-readable media as recited in claim 1, wherein the one or more computer-readable media comprise a computer memory of a wearable computer.
10. A method comprising: displaying a list of items to be handled by a user in a particular order; identifying one item in the list of items that is the current item; receiving a user input corresponding to the current item; and updating, in response to receiving the user input, the identification of the one item that is the current item to indicate the next item in the list of items as the current item.
11. A method as recited in claim 10, wherein displaying the list of items comprises displaying at least one item corresponding to a task that has already been performed and at least one item corresponding to a task that still needs to be performed by the user.
12. A method as recited in claim 10, wherein displaying the list of items comprises displaying, after the user input is received, the user input in place of the corresponding item.
13. A method as recited in claim 10, wherein displaying the list of items comprises displaying only a subset of the list of items at any given time.
14. A method as recited in claim 13, further comprising scrolling through the list of items to display different subsets as items in the list are handled by the user.
15. A method as recited in claim 10, further comprising displaying a current processing marker identifying an item in the list of items corresponding to a current user input being processed.
16. A method as recited in claim 10, wherein the list of items comprises a list of tasks to be completed by the user, and wherein handling of an item by the user comprises the user completing the task.
17. A method as recited in claim 16, wherein the list of tasks comprises a list of prompts corresponding to data to be entered into the computer by the user.
18. A method as recited in claim 10, wherein the list of items comprises a list of prompts of words to be spoken by the user, and wherein handling of an item by the user comprises speaking one or more words corresponding to the prompt.
19. One or more computer- readable memories containing a computer program that is executable by a processor to perform the method recited in claim 10.
20. A method comprising: displaying an identification of a plurality of users; and for each of the plurality of users, displaying a list of tasks to be performed by the user, identifying one task in the list of tasks that is the current task that needs to be performed by the user, and updating, in response to completion of the task by the user, the identification of the one task that is the current task that needs to be performed by the user to be the next task in the list of tasks.
21. A method as recited in claim 20, wherein displaying the list of tasks comprises displaying only a subset of the list of tasks to be performed by the user at any given time.
22. A method as recited in claim 21, further comprising scrolling through the list of tasks to display different subsets as tasks in the list are completed by the user.
23. A method as recited in claim 20, wherein the list of tasks comprises a list of actions to be taken by the user.
24. A method as recited in claim 20, wherein identifying one task that is the current task comprises displaying a geometric shape as a current location marker identifying the one task.
25. A method as recited in claim 20, wherein identifying one task that is the current task comprises displaying the one task differently than the other tasks in the list of tasks.
26. A method as recited in claim 20, further comprising: receiving, for each of the plurality of users, an indication from each user's computer of the current task for that user.
27. One or more computer-readable memories containing a computer program that is executable by a processor to perform the method recited in claim 20.
28. A graphical user interface comprising: a list portion identifying a list of a plurality of items to be handled by a user; a user choices portion identifying information corresponding to a current item in the list; and a current location marker that identifies one item of the list that is the current item to be handled by the user, wherein the current location marker is automatically updated to identify the next item in the list after the current item in the list has been handled by the user.
29. A graphical user interface as recited in claim 28, further comprising an applet window portion identifying information clarifying the information identified in the user choices portion.
30. A graphical user interface as recited in claim 29, wherein the user choices portion identifies information that is to be entered into a computer by the user, and wherein the applet window portion identifies information that has already been entered into the computer by the user.
31. A graphical user interface as recited in claim 28, wherein the list of a plurality of items comprises a list of words to be spoken by the user.
32. A graphical user interface as recited in claim 28, wherein the list of a plurality of items comprises a list of prompts of words to be spoken by the user, and wherein the user choices portion identifies, for each prompt, one or more words that can be spoken by the user to properly handle the prompt.
33. A graphical user interface as recited in claim 28, wherein the list portion further identifies information that has been entered by the user in handling previous items in the list.
34. A graphical user interface as recited in claim 28 implemented on a wearable computer.
35. A system comprising: a display device; a user interface component, coupled to the display device, causing a user interface to be displayed on the display device; wherein the user interface includes a list portion in which a list of a plurality of items to be handled by a user are displayed; wherein the user interface further includes a current location marker identifying one of the items in the list as the current item that needs to be handled by the user; and wherein the user interface component further automatically updates the current location marker to identify a new item in the list in response to the user handling the current item in the list.
36. A system as recited in claim 35, wherein the user interface component further replaces, after the user has handled the current item, a user input in place of the current item.
37. A system as recited in claim 35, wherein the user interface includes only a subset of the list of the plurality of items at any given time.
38. A system as recited in claim 37, wherein the user interface component further scrolls through the list of items to display different subsets as items in the list are handled by the user.
39. A system as recited in claim 35, wherein the user interface component further displays, as part of the user interface, a current processing marker identifying an item in the list that is currently being processed by the system.
40. A system as recited in claim 35, wherein the list of a plurality of items comprises a list of a plurality of tasks to be completed by the user, and wherein handling of an item by the user comprises the user completing the task.
41. A system as recited in claim 40, wherein the list of tasks comprises a list of prompts corresponding to data to be entered into the system by the user.
42. A system as recited in claim 40, wherein the user interface component is implemented in software.
43. A method comprising: displaying a list of tasks to be performed; identifying one task in the list of tasks that is the current task needing to be performed; receiving an input corresponding to the current task; and updating, in response to receiving the input, the identification of the one task that is the current task to indicate that the next task in the list of tasks is the current task needing to be performed.
44. A method as recited in claim 43, wherein the displaying comprises displaying a list of tasks to be perfoπned by a user.
45. A method as recited in claim 43, wherein the identifying comprises superimposing, on the display of the current task in the list, a set of one or more input options corresponding to the task.
46. A method as recited in claim 45, wherein the receiving comprises receiving, as the input corresponding to the current task, one of the input options from the set of one or more input options.
47. A method as recited in claim 43, wherein the receiving comprises receiving a user input.
48. A method as recited in claim 43, wherein the receiving comprises receiving an input from a computer component, wherein the input from the computer component indicates that the current task is completed.
49. A method as recited in claim 48, wherein the computer component comprises a processor executing a software program.
50. A method as recited in claim 48, wherein the computer component comprises a hardware component configured to carry out the current task.
51. A method as recited in claim 48, wherein the computer component comprises a remote computer.
52. A method as recited in claim 43, wherein displaying the list of tasks comprises displaying only a subset of the list of tasks at any given time.
53. A method as recited in claim 52, further comprising scrolling through the list of tasks to display different subsets as tasks in the list are performed by the user.
54. A method as recited in claim 43, further comprising displaying a current processing marker identifying a task in the list of tasks coπ-esponding to a current input being processed by a computer performing the method.
55. One or more computer-readable memories containing a computer program that is executable by a processor to perform the method recited in claim 43.
56. A graphical user interface comprising: a task list portion identifying a list of a plurality of tasks to be performed by a user; and an indication in the task list portion of a current task to be performed, wherein the indication is changed, in response to the current task being performed, to indicate a next task in the list as the current task to be performed.
57. A graphical user interface as recited in claim 56, further comprising a user choices portion identifying information corresponding to the current task on the list to be performed.
58. A graphical user interface as recited in claim 56, further comprising: a second task list portion identifying a list of a plurality of tasks to be performed by another user; and an indication in the second task list portion of a current task to be performed by the other user, wherein the indication is changed, in response to the current task being performed by the other user, to indicate a next task in the list of tasks to be performed by the other use as the current task to be performed.
59. A system comprising: means for displaying a list of items to be handled by a user in a particular order; and means for identifying one item in the list of items that is the current item, for receiving a user input corresponding to the current item, and for updating, in response to receiving the user input, the identification of the one item that is the current item to indicate the next item in the list of items as the current item.
PCT/US2001/031987 2000-10-16 2001-10-15 Dynamically displaying current status of tasks WO2002033578A2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
AU2002211699A AU2002211699A1 (en) 2000-10-16 2001-10-15 Dynamically displaying current status of tasks

Applications Claiming Priority (4)

Application Number Priority Date Filing Date Title
US24068500P 2000-10-16 2000-10-16
US60/240,685 2000-10-16
US09/879,829 2001-06-11
US09/879,829 US20020054130A1 (en) 2000-10-16 2001-06-11 Dynamically displaying current status of tasks

Publications (2)

Publication Number Publication Date
WO2002033578A2 true WO2002033578A2 (en) 2002-04-25
WO2002033578A8 WO2002033578A8 (en) 2002-07-11

Family

ID=26933618

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2001/031987 WO2002033578A2 (en) 2000-10-16 2001-10-15 Dynamically displaying current status of tasks

Country Status (3)

Country Link
US (2) US20020054130A1 (en)
AU (1) AU2002211699A1 (en)
WO (1) WO2002033578A2 (en)

Families Citing this family (266)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6513046B1 (en) 1999-12-15 2003-01-28 Tangis Corporation Storing and recalling information to augment human memories
US7225229B1 (en) 1998-12-18 2007-05-29 Tangis Corporation Automated pushing of computer user's context data to clients
US6920616B1 (en) 1998-12-18 2005-07-19 Tangis Corporation Interface for exchanging context data
US8225214B2 (en) 1998-12-18 2012-07-17 Microsoft Corporation Supplying enhanced computer user's context data
US7779015B2 (en) 1998-12-18 2010-08-17 Microsoft Corporation Logging and analyzing context attributes
US7231439B1 (en) 2000-04-02 2007-06-12 Tangis Corporation Dynamically swapping modules for determining a computer user's context
US8181113B2 (en) 1998-12-18 2012-05-15 Microsoft Corporation Mediating conflicts in computer users context data
US6842877B2 (en) 1998-12-18 2005-01-11 Tangis Corporation Contextual responses based on automated learning techniques
US7046263B1 (en) 1998-12-18 2006-05-16 Tangis Corporation Requesting computer user's context data
US9183306B2 (en) 1998-12-18 2015-11-10 Microsoft Technology Licensing, Llc Automated selection of appropriate information based on a computer user's context
US6791580B1 (en) 1998-12-18 2004-09-14 Tangis Corporation Supplying notifications related to supply and consumption of user context data
US6801223B1 (en) 1998-12-18 2004-10-05 Tangis Corporation Managing interactions between computer users' context models
US6999955B1 (en) 1999-04-20 2006-02-14 Microsoft Corporation Systems and methods for estimating and integrating measures of human cognitive load into the behavior of computational applications and services
US6434527B1 (en) * 1999-05-17 2002-08-13 Microsoft Corporation Signalling and controlling the status of an automatic speech recognition system for use in handsfree conversational dialogue
US6931384B1 (en) 1999-06-04 2005-08-16 Microsoft Corporation System and method providing utility-based decision making about clarification dialog given communicative uncertainty
US7103806B1 (en) * 1999-06-04 2006-09-05 Microsoft Corporation System for performing context-sensitive decisions about ideal communication modalities considering information about channel reliability
US7389351B2 (en) * 2001-03-15 2008-06-17 Microsoft Corporation System and method for identifying and establishing preferred modalities or channels for communications based on participants' preferences and contexts
US6714967B1 (en) 1999-07-30 2004-03-30 Microsoft Corporation Integration of a computer-based message priority system with mobile electronic devices
US7194681B1 (en) * 1999-07-30 2007-03-20 Microsoft Corporation Method for automatically assigning priorities to documents and messages
US6622160B1 (en) * 1999-07-30 2003-09-16 Microsoft Corporation Methods for routing items for communications based on a measure of criticality
US6847924B1 (en) * 2000-06-19 2005-01-25 Ncr Corporation Method and system for aggregating data distribution models
US7243130B2 (en) * 2000-03-16 2007-07-10 Microsoft Corporation Notification platform architecture
US7565403B2 (en) * 2000-03-16 2009-07-21 Microsoft Corporation Use of a bulk-email filter within a system for classifying messages for urgency or importance
US8024415B2 (en) * 2001-03-16 2011-09-20 Microsoft Corporation Priorities generation and management
US7444383B2 (en) * 2000-06-17 2008-10-28 Microsoft Corporation Bounded-deferral policies for guiding the timing of alerting, interaction and communications using local sensory information
US8701027B2 (en) 2000-03-16 2014-04-15 Microsoft Corporation Scope user interface for displaying the priorities and properties of multiple informational items
US7634528B2 (en) * 2000-03-16 2009-12-15 Microsoft Corporation Harnessing information about the timing of a user's client-server interactions to enhance messaging and collaboration services
US7743340B2 (en) * 2000-03-16 2010-06-22 Microsoft Corporation Positioning and rendering notification heralds based on user's focus of attention and activity
US7464153B1 (en) 2000-04-02 2008-12-09 Microsoft Corporation Generating and supplying user context data
WO2001075676A2 (en) 2000-04-02 2001-10-11 Tangis Corporation Soliciting information based on a computer user's context
US6938024B1 (en) 2000-05-04 2005-08-30 Microsoft Corporation Transmitting information given constrained resources
US8086672B2 (en) * 2000-06-17 2011-12-27 Microsoft Corporation When-free messaging
WO2002033541A2 (en) * 2000-10-16 2002-04-25 Tangis Corporation Dynamically determining appropriate computer interfaces
US20020054130A1 (en) 2000-10-16 2002-05-09 Abbott Kenneth H. Dynamically displaying current status of tasks
US7844666B2 (en) 2000-12-12 2010-11-30 Microsoft Corporation Controls and displays for acquiring preferences, inspecting behavior, and guiding the learning and decision policies of an adaptive communications prioritization and routing system
US6745193B1 (en) 2001-01-25 2004-06-01 Microsoft Corporation System and method for defining, refining, and personalizing communications policies in a notification platform
US6901398B1 (en) 2001-02-12 2005-05-31 Microsoft Corporation System and method for constructing and personalizing a universal information classifier
SE518484C2 (en) * 2001-02-27 2002-10-15 Peder Holmbom Apparatus and method for disinfecting water for medical or dental purposes
US7251696B1 (en) 2001-03-15 2007-07-31 Microsoft Corporation System and methods enabling a mix of human and automated initiatives in the control of communication policies
US7330895B1 (en) 2001-03-15 2008-02-12 Microsoft Corporation Representation, decision models, and user interface for encoding managing preferences, and performing automated decision making about the timing and modalities of interpersonal communications
US7512940B2 (en) * 2001-03-29 2009-03-31 Microsoft Corporation Methods and apparatus for downloading and/or distributing information and/or software resources based on expected utility
US6947935B1 (en) * 2001-04-04 2005-09-20 Microsoft Corporation Training, inference and user interface for guiding the caching of media content on local stores
US7757250B1 (en) 2001-04-04 2010-07-13 Microsoft Corporation Time-centric training, inference and user interface for personalized media program guides
US7039642B1 (en) * 2001-05-04 2006-05-02 Microsoft Corporation Decision-theoretic methods for identifying relevant substructures of a hierarchical file structure to enhance the efficiency of document access, browsing, and storage
US7107254B1 (en) 2001-05-07 2006-09-12 Microsoft Corporation Probablistic models and methods for combining multiple content classifiers
US7089226B1 (en) 2001-06-28 2006-08-08 Microsoft Corporation System, representation, and method providing multilevel information retrieval with clarification dialog
US7043506B1 (en) 2001-06-28 2006-05-09 Microsoft Corporation Utility-based archiving
US7233933B2 (en) 2001-06-28 2007-06-19 Microsoft Corporation Methods and architecture for cross-device activity monitoring, reasoning, and visualization for providing status and forecasts of a users' presence and availability
US7409423B2 (en) * 2001-06-28 2008-08-05 Horvitz Eric J Methods for and applications of learning and inferring the periods of time until people are available or unavailable for different forms of communication, collaboration, and information access
US7689521B2 (en) * 2001-06-28 2010-03-30 Microsoft Corporation Continuous time bayesian network models for predicting users' presence, activities, and component usage
US7493369B2 (en) * 2001-06-28 2009-02-17 Microsoft Corporation Composable presence and availability services
US7519529B1 (en) 2001-06-29 2009-04-14 Microsoft Corporation System and methods for inferring informational goals and preferred level of detail of results in response to questions posed to an automated information-retrieval or question-answering service
US7409335B1 (en) 2001-06-29 2008-08-05 Microsoft Corporation Inferring informational goals and preferred level of detail of answers based on application being employed by the user
US7644144B1 (en) 2001-12-21 2010-01-05 Microsoft Corporation Methods, tools, and interfaces for the dynamic assignment of people to groups to enable enhanced communication and collaboration
US7203909B1 (en) 2002-04-04 2007-04-10 Microsoft Corporation System and methods for constructing personalized context-sensitive portal pages or views by analyzing patterns of users' information access activities
US20030212761A1 (en) * 2002-05-10 2003-11-13 Microsoft Corporation Process kernel
US20030217098A1 (en) 2002-05-15 2003-11-20 Microsoft Corporation Method and system for supporting the communication of presence information regarding one or more telephony devices
US7203635B2 (en) 2002-06-27 2007-04-10 Microsoft Corporation Layered models for context awareness
US7069259B2 (en) * 2002-06-28 2006-06-27 Microsoft Corporation Multi-attribute specification of preferences about people, priorities and privacy for guiding messaging and communications
US7870240B1 (en) 2002-06-28 2011-01-11 Microsoft Corporation Metadata schema for interpersonal communications management systems
US7418702B2 (en) * 2002-08-06 2008-08-26 Sheng (Ted) Tai Tsao Concurrent web based multi-task support for control management system
US20040153445A1 (en) * 2003-02-04 2004-08-05 Horvitz Eric J. Systems and methods for constructing and using models of memorability in computing and communications applications
US7536650B1 (en) 2003-02-25 2009-05-19 Robertson George G System and method that facilitates computer desktop use via scaling of displayed objects with shifts to the periphery
US8230359B2 (en) * 2003-02-25 2012-07-24 Microsoft Corporation System and method that facilitates computer desktop use via scaling of displayed objects with shifts to the periphery
US8225224B1 (en) 2003-02-25 2012-07-17 Microsoft Corporation Computer desktop use via scaling of displayed objects with shifts to the periphery
US7793233B1 (en) 2003-03-12 2010-09-07 Microsoft Corporation System and method for customizing note flags
US7774799B1 (en) 2003-03-26 2010-08-10 Microsoft Corporation System and method for linking page content with a media file and displaying the links
US7164361B2 (en) * 2003-03-31 2007-01-16 Schering-Plough Animal Health Corporation System and method for collecting, processing and managing livestock data
US7457879B2 (en) 2003-04-01 2008-11-25 Microsoft Corporation Notification platform architecture
US6992625B1 (en) * 2003-04-25 2006-01-31 Microsoft Corporation Calibration of a device location measurement system that utilizes wireless signal strengths
US7162473B2 (en) * 2003-06-26 2007-01-09 Microsoft Corporation Method and system for usage analyzer that determines user accessed sources, indexes data subsets, and associated metadata, processing implicit queries based on potential interest to users
US20040267746A1 (en) * 2003-06-26 2004-12-30 Cezary Marcjan User interface for controlling access to computer objects
US7225187B2 (en) 2003-06-26 2007-05-29 Microsoft Corporation Systems and methods for performing background queries from content and activity
US8346587B2 (en) * 2003-06-30 2013-01-01 Microsoft Corporation Models and methods for reducing visual complexity and search effort via ideal information abstraction, hiding, and sequencing
US7250907B2 (en) * 2003-06-30 2007-07-31 Microsoft Corporation System and methods for determining the location dynamics of a portable computing device
US7444598B2 (en) * 2003-06-30 2008-10-28 Microsoft Corporation Exploded views for providing rich regularized geometric transformations and interaction models on content for viewing, previewing, and interacting with documents, projects, and tasks
US20040264677A1 (en) * 2003-06-30 2004-12-30 Horvitz Eric J. Ideal transfer of call handling from automated systems to human operators based on forecasts of automation efficacy and operator load
US7738881B2 (en) * 2003-07-22 2010-06-15 Microsoft Corporation Systems for determining the approximate location of a device from ambient signals
US7202816B2 (en) * 2003-07-22 2007-04-10 Microsoft Corporation Utilization of the approximate location of a device determined from ambient signals
US7319877B2 (en) * 2003-07-22 2008-01-15 Microsoft Corporation Methods for determining the approximate location of a device from ambient signals
US7454393B2 (en) * 2003-08-06 2008-11-18 Microsoft Corporation Cost-benefit approach to automatically composing answers to questions by extracting information from large unstructured corpora
US7330112B1 (en) 2003-09-09 2008-02-12 Emigh Aaron T Location-aware services
US7818317B1 (en) 2003-09-09 2010-10-19 James Roskind Location-based tasks
US20050084082A1 (en) * 2003-10-15 2005-04-21 Microsoft Corporation Designs, interfaces, and policies for systems that enhance communication and minimize disruption by encoding preferences and situations
US7831679B2 (en) * 2003-10-15 2010-11-09 Microsoft Corporation Guiding sensing and preferences for context-sensitive services
US7774349B2 (en) 2003-12-11 2010-08-10 Microsoft Corporation Statistical models and methods to support the personalization of applications and services via consideration of preference encodings of a community of users
US20050165631A1 (en) * 2004-01-28 2005-07-28 Microsoft Corporation Time management representations and automation for allocating time to projects and meetings within an online calendaring system
US7293019B2 (en) 2004-03-02 2007-11-06 Microsoft Corporation Principles and methods for personalizing newsfeeds via an analysis of information novelty and dynamics
US7327349B2 (en) * 2004-03-02 2008-02-05 Microsoft Corporation Advanced navigation techniques for portable devices
US7908663B2 (en) 2004-04-20 2011-03-15 Microsoft Corporation Abstractions and automation for enhanced sharing and collaboration
US7664249B2 (en) * 2004-06-30 2010-02-16 Microsoft Corporation Methods and interfaces for probing and understanding behaviors of alerting and filtering systems based on models and simulation from logs
US20060012183A1 (en) * 2004-07-19 2006-01-19 David Marchiori Rail car door opener
US7788589B2 (en) * 2004-09-30 2010-08-31 Microsoft Corporation Method and system for improved electronic task flagging and management
US7712049B2 (en) 2004-09-30 2010-05-04 Microsoft Corporation Two-dimensional radial user interface for computer software applications
US20060074883A1 (en) * 2004-10-05 2006-04-06 Microsoft Corporation Systems, methods, and interfaces for providing personalized search and information access
US7610560B2 (en) 2004-11-16 2009-10-27 Microsoft Corporation Methods for automated and semiautomated composition of visual sequences, flows, and flyovers based on content and context
US7519564B2 (en) * 2004-11-16 2009-04-14 Microsoft Corporation Building and using predictive models of current and future surprises
US7698055B2 (en) * 2004-11-16 2010-04-13 Microsoft Corporation Traffic forecasting employing modeling and analysis of probabilistic interdependencies and contextual data
US7327245B2 (en) * 2004-11-22 2008-02-05 Microsoft Corporation Sensing and analysis of ambient contextual signals for discriminating between indoor and outdoor locations
AU2005321876B2 (en) * 2004-12-31 2011-07-07 Ntrepid, Llc System for protecting identity in a network environment
US7689615B2 (en) * 2005-02-25 2010-03-30 Microsoft Corporation Ranking results using multiple nested ranking
US7885817B2 (en) 2005-03-08 2011-02-08 Microsoft Corporation Easy generation and automatic training of spoken dialog systems using text-to-speech
US7734471B2 (en) * 2005-03-08 2010-06-08 Microsoft Corporation Online learning for dialog systems
US20060206333A1 (en) * 2005-03-08 2006-09-14 Microsoft Corporation Speaker-dependent dialog adaptation
US7707131B2 (en) * 2005-03-08 2010-04-27 Microsoft Corporation Thompson strategy based online reinforcement learning system for action selection
US20070011109A1 (en) * 2005-06-23 2007-01-11 Microsoft Corporation Immortal information storage and access platform
US7991607B2 (en) * 2005-06-27 2011-08-02 Microsoft Corporation Translation and capture architecture for output of conversational utterances
US7643985B2 (en) * 2005-06-27 2010-01-05 Microsoft Corporation Context-sensitive communication and translation methods for enhanced interactions and understanding among speakers of different languages
US8079079B2 (en) * 2005-06-29 2011-12-13 Microsoft Corporation Multimodal authentication
US7693817B2 (en) 2005-06-29 2010-04-06 Microsoft Corporation Sensing, storing, indexing, and retrieving data leveraging measures of user activity, attention, and interest
US7428521B2 (en) * 2005-06-29 2008-09-23 Microsoft Corporation Precomputation of context-sensitive policies for automated inquiry and action under uncertainty
US20070005363A1 (en) * 2005-06-29 2007-01-04 Microsoft Corporation Location aware multi-modal multi-lingual device
US7529683B2 (en) * 2005-06-29 2009-05-05 Microsoft Corporation Principals and methods for balancing the timeliness of communications and information delivery with the expected cost of interruption via deferral policies
US7460884B2 (en) * 2005-06-29 2008-12-02 Microsoft Corporation Data buddy
US20070004969A1 (en) * 2005-06-29 2007-01-04 Microsoft Corporation Health monitor
US7647171B2 (en) * 2005-06-29 2010-01-12 Microsoft Corporation Learning, storing, analyzing, and reasoning about the loss of location-identifying signals
US7694214B2 (en) * 2005-06-29 2010-04-06 Microsoft Corporation Multimodal note taking, annotation, and gaming
US20070005646A1 (en) * 2005-06-30 2007-01-04 Microsoft Corporation Analysis of topic dynamics of web search
US7646755B2 (en) * 2005-06-30 2010-01-12 Microsoft Corporation Seamless integration of portable computing devices and desktop computers
US7925995B2 (en) * 2005-06-30 2011-04-12 Microsoft Corporation Integration of location logs, GPS signals, and spatial resources for identifying user activities, goals, and context
US20070005754A1 (en) * 2005-06-30 2007-01-04 Microsoft Corporation Systems and methods for triaging attention for providing awareness of communications session activity
JP2007040905A (en) * 2005-08-04 2007-02-15 Hitachi High-Technologies Corp Chromatographic data processor
US20070050251A1 (en) * 2005-08-29 2007-03-01 Microsoft Corporation Monetizing a preview pane for ads
US20070050253A1 (en) * 2005-08-29 2007-03-01 Microsoft Corporation Automatically generating content for presenting in a preview pane for ADS
US20070050252A1 (en) * 2005-08-29 2007-03-01 Microsoft Corporation Preview pane for ads
US8024112B2 (en) * 2005-09-29 2011-09-20 Microsoft Corporation Methods for predicting destinations from partial trajectories employing open-and closed-world modeling methods
US7319908B2 (en) * 2005-10-28 2008-01-15 Microsoft Corporation Multi-modal device power/mode management
US7467353B2 (en) * 2005-10-28 2008-12-16 Microsoft Corporation Aggregation of multi-modal devices
US20070100704A1 (en) * 2005-10-28 2007-05-03 Microsoft Corporation Shopping assistant
US7778632B2 (en) * 2005-10-28 2010-08-17 Microsoft Corporation Multi-modal device capable of automated actions
US20070112906A1 (en) * 2005-11-15 2007-05-17 Microsoft Corporation Infrastructure for multi-modal multilingual communications devices
US20070136222A1 (en) * 2005-12-09 2007-06-14 Microsoft Corporation Question and answer architecture for reasoning and clarifying intentions, goals, and needs from contextual clues and content
US20070136068A1 (en) * 2005-12-09 2007-06-14 Microsoft Corporation Multimodal multilingual devices and applications for enhanced goal-interpretation and translation for service providers
US20070150512A1 (en) * 2005-12-15 2007-06-28 Microsoft Corporation Collaborative meeting assistant
US7797638B2 (en) * 2006-01-05 2010-09-14 Microsoft Corporation Application of metadata to documents and document objects via a software application user interface
US7747557B2 (en) * 2006-01-05 2010-06-29 Microsoft Corporation Application of metadata to documents and document objects via an operating system user interface
US7617164B2 (en) * 2006-03-17 2009-11-10 Microsoft Corporation Efficiency of training for ranking systems based on pairwise training with aggregated gradients
US20070245223A1 (en) * 2006-04-17 2007-10-18 Microsoft Corporation Synchronizing multimedia mobile notes
US20070245229A1 (en) * 2006-04-17 2007-10-18 Microsoft Corporation User experience for multimedia mobile note taking
US20070288279A1 (en) * 2006-06-07 2007-12-13 Microsoft Corporation Entering And Using Time Ranges
US7761464B2 (en) * 2006-06-19 2010-07-20 Microsoft Corporation Diversifying search results for improved search and personalization
US7610151B2 (en) 2006-06-27 2009-10-27 Microsoft Corporation Collaborative route planning for generating personalized and context-sensitive routing recommendations
US20080005223A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Reputation data for entities and data processing
US20080004990A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Virtual spot market for advertisements
US20080005068A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Context-based search, retrieval, and awareness
US20080005108A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Message mining to enhance ranking of documents for retrieval
US20080005069A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Entity-specific search model
US9396269B2 (en) * 2006-06-28 2016-07-19 Microsoft Technology Licensing, Llc Search engine that identifies and uses social networks in communications, retrieval, and electronic commerce
US8874592B2 (en) 2006-06-28 2014-10-28 Microsoft Corporation Search guided by location and context
US7739221B2 (en) * 2006-06-28 2010-06-15 Microsoft Corporation Visual and multi-dimensional search
US20080005067A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Context-based search, retrieval, and awareness
US9141704B2 (en) * 2006-06-28 2015-09-22 Microsoft Technology Licensing, Llc Data management in social networks
US20080005074A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Search over designated content
US7917514B2 (en) * 2006-06-28 2011-03-29 Microsoft Corporation Visual and multi-dimensional search
US20080005104A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Localized marketing
US8788517B2 (en) * 2006-06-28 2014-07-22 Microsoft Corporation Intelligently guiding search based on user dialog
US7984169B2 (en) * 2006-06-28 2011-07-19 Microsoft Corporation Anonymous and secure network-based interaction
US7822762B2 (en) * 2006-06-28 2010-10-26 Microsoft Corporation Entity-specific search model
US20080004948A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Auctioning for video and audio advertising
US20080005095A1 (en) * 2006-06-28 2008-01-03 Microsoft Corporation Validation of computer responses
US8626136B2 (en) * 2006-06-29 2014-01-07 Microsoft Corporation Architecture for user- and context-specific prefetching and caching of information on portable devices
US20080005047A1 (en) * 2006-06-29 2008-01-03 Microsoft Corporation Scenario-based search
US8725567B2 (en) * 2006-06-29 2014-05-13 Microsoft Corporation Targeted advertising in brick-and-mortar establishments
US7873620B2 (en) * 2006-06-29 2011-01-18 Microsoft Corporation Desktop search from mobile device
US7552862B2 (en) * 2006-06-29 2009-06-30 Microsoft Corporation User-controlled profile sharing
US20080005079A1 (en) * 2006-06-29 2008-01-03 Microsoft Corporation Scenario-based search
US20080004951A1 (en) * 2006-06-29 2008-01-03 Microsoft Corporation Web-based targeted advertising in a brick-and-mortar retail establishment using online customer information
US20080005313A1 (en) * 2006-06-29 2008-01-03 Microsoft Corporation Using offline activity to enhance online searching
US7997485B2 (en) * 2006-06-29 2011-08-16 Microsoft Corporation Content presentation based on user preferences
US8244240B2 (en) * 2006-06-29 2012-08-14 Microsoft Corporation Queries as data for revising and extending a sensor-based location service
US20080004884A1 (en) * 2006-06-29 2008-01-03 Microsoft Corporation Employment of offline behavior to display online content
US7617042B2 (en) * 2006-06-30 2009-11-10 Microsoft Corporation Computing and harnessing inferences about the timing, duration, and nature of motion and cessation of motion with applications to mobile computing and communications
US8126641B2 (en) * 2006-06-30 2012-02-28 Microsoft Corporation Route planning with contingencies
US7797267B2 (en) * 2006-06-30 2010-09-14 Microsoft Corporation Methods and architecture for learning and reasoning in support of context-sensitive reminding, informing, and service facilitation
US8112755B2 (en) * 2006-06-30 2012-02-07 Microsoft Corporation Reducing latencies in computing systems using probabilistic and/or decision-theoretic reasoning under scarce memory resources
US20080004954A1 (en) * 2006-06-30 2008-01-03 Microsoft Corporation Methods and architecture for performing client-side directed marketing with caching and local analytics for enhanced privacy and minimal disruption
US7739040B2 (en) 2006-06-30 2010-06-15 Microsoft Corporation Computation of travel routes, durations, and plans over multiple contexts
US7706964B2 (en) * 2006-06-30 2010-04-27 Microsoft Corporation Inferring road speeds for context-sensitive routing
US7696868B1 (en) 2006-08-22 2010-04-13 Emigh Aaron T Location-based alerting
US7707518B2 (en) * 2006-11-13 2010-04-27 Microsoft Corporation Linking information
US7761785B2 (en) 2006-11-13 2010-07-20 Microsoft Corporation Providing resilient links
US7711716B2 (en) * 2007-03-06 2010-05-04 Microsoft Corporation Optimizations for a background database consistency check
US20080249667A1 (en) * 2007-04-09 2008-10-09 Microsoft Corporation Learning and reasoning to enhance energy efficiency in transportation systems
US7970721B2 (en) * 2007-06-15 2011-06-28 Microsoft Corporation Learning and reasoning from web projections
US7539659B2 (en) * 2007-06-15 2009-05-26 Microsoft Corporation Multidimensional timeline browsers for broadcast media
US7979252B2 (en) * 2007-06-21 2011-07-12 Microsoft Corporation Selective sampling of user state based on expected utility
US20080320087A1 (en) * 2007-06-22 2008-12-25 Microsoft Corporation Swarm sensing and actuating
US20080319660A1 (en) * 2007-06-25 2008-12-25 Microsoft Corporation Landmark-based routing
US7912637B2 (en) * 2007-06-25 2011-03-22 Microsoft Corporation Landmark-based routing
US20080319658A1 (en) * 2007-06-25 2008-12-25 Microsoft Corporation Landmark-based routing
US7991718B2 (en) * 2007-06-28 2011-08-02 Microsoft Corporation Method and apparatus for generating an inference about a destination of a trip using a combination of open-world modeling and closed world modeling
US7696866B2 (en) * 2007-06-28 2010-04-13 Microsoft Corporation Learning and reasoning about the context-sensitive reliability of sensors
US8244660B2 (en) 2007-06-28 2012-08-14 Microsoft Corporation Open-world modeling
US8631419B2 (en) * 2007-06-29 2014-01-14 Microsoft Corporation System and methods for disruption detection, management, and recovery
US7948400B2 (en) * 2007-06-29 2011-05-24 Microsoft Corporation Predictive models of road reliability for traffic sensor configuration and routing
US7673088B2 (en) * 2007-06-29 2010-03-02 Microsoft Corporation Multi-tasking interference model
US8254393B2 (en) * 2007-06-29 2012-08-28 Microsoft Corporation Harnessing predictive models of durations of channel availability for enhanced opportunistic allocation of radio spectrum
US8095889B2 (en) * 2008-05-12 2012-01-10 Honeywell International Inc. Heuristic and intuitive user interface for access control systems
US8516001B2 (en) * 2008-06-24 2013-08-20 Microsoft Corporation Context platform
US9846049B2 (en) * 2008-07-09 2017-12-19 Microsoft Technology Licensing, Llc Route prediction
US9984340B2 (en) * 2008-08-26 2018-05-29 The Boeing Company Methods and systems for deploying a single continuous improvement approach across an enterprise
US20100083150A1 (en) * 2008-09-30 2010-04-01 Nokia Corporation User interface, device and method for providing a use case based interface
US20100088143A1 (en) * 2008-10-07 2010-04-08 Microsoft Corporation Calendar event scheduling
FR2940482B1 (en) * 2008-12-19 2011-01-14 Thales Sa DEVICE FOR MANAGING STEERING TASKS CARRIED OUT BY A CREW OF AN AIRCRAFT
FR2940480B1 (en) * 2008-12-23 2011-03-25 Thales Sa DEVICE FOR RECONFIGURING A TASK TREATMENT CONTEXT
US8957790B2 (en) * 2009-01-06 2015-02-17 The Boeing Company System and method for cruise monitoring and alerting
US8489545B2 (en) * 2009-03-18 2013-07-16 Norman Ritchie System and method for creating and maintaining pre-task planning documents
US8346800B2 (en) * 2009-04-02 2013-01-01 Microsoft Corporation Content-based information retrieval
US8661030B2 (en) 2009-04-09 2014-02-25 Microsoft Corporation Re-ranking top search results
US20100268563A1 (en) * 2009-04-16 2010-10-21 Mitel Networks Corporation Proximity scripting
US20100299669A1 (en) * 2009-05-20 2010-11-25 Microsoft Corporation Generation of a Comparison Task List of Task Items
US8700405B2 (en) * 2010-02-16 2014-04-15 Honeywell International Inc Audio system and method for coordinating tasks
US8386929B2 (en) 2010-06-22 2013-02-26 Microsoft Corporation Personal assistant for task utilization
US9009592B2 (en) 2010-06-22 2015-04-14 Microsoft Technology Licensing, Llc Population of lists and tasks from captured voice and audio content
US8381088B2 (en) 2010-06-22 2013-02-19 Microsoft Corporation Flagging, capturing and generating task list items
US8375320B2 (en) * 2010-06-22 2013-02-12 Microsoft Corporation Context-based task generation
US8355954B1 (en) * 2010-06-24 2013-01-15 Amazon Technologies, Inc. Generating and updating recommendations for merchants
US9305263B2 (en) 2010-06-30 2016-04-05 Microsoft Technology Licensing, Llc Combining human and machine intelligence to solve tasks with crowd sourcing
JP5323103B2 (en) * 2010-09-03 2013-10-23 三菱電機株式会社 Graphical user interface device
US8565783B2 (en) 2010-11-24 2013-10-22 Microsoft Corporation Path progression matching for indoor positioning systems
US9589254B2 (en) 2010-12-08 2017-03-07 Microsoft Technology Licensing, Llc Using e-mail message characteristics for prioritization
US9134137B2 (en) 2010-12-17 2015-09-15 Microsoft Technology Licensing, Llc Mobile search based on predicted location
US8781452B2 (en) 2011-02-21 2014-07-15 Motorola Mobility Llc Wireless devices and methods of operating wireless devices based on the presence of another person
US9163952B2 (en) 2011-04-15 2015-10-20 Microsoft Technology Licensing, Llc Suggestive mapping
US8981995B2 (en) 2011-06-03 2015-03-17 Microsoft Technology Licensing, Llc. Low accuracy positional data by detecting improbable samples
US9470529B2 (en) 2011-07-14 2016-10-18 Microsoft Technology Licensing, Llc Activating and deactivating sensors for dead reckoning
US9464903B2 (en) 2011-07-14 2016-10-11 Microsoft Technology Licensing, Llc Crowd sourcing based on dead reckoning
US8538686B2 (en) 2011-09-09 2013-09-17 Microsoft Corporation Transport-dependent prediction of destinations
US10192176B2 (en) 2011-10-11 2019-01-29 Microsoft Technology Licensing, Llc Motivation of task completion and personalization of tasks and lists
US10184798B2 (en) 2011-10-28 2019-01-22 Microsoft Technology Licensing, Llc Multi-stage dead reckoning for crowd sourcing
US9429657B2 (en) 2011-12-14 2016-08-30 Microsoft Technology Licensing, Llc Power efficient activation of a device movement sensor module
US8775337B2 (en) 2011-12-19 2014-07-08 Microsoft Corporation Virtual sensor development
US9230501B1 (en) 2012-01-06 2016-01-05 Google Inc. Device control utilizing optical flow
US8952869B1 (en) 2012-01-06 2015-02-10 Google Inc. Determining correlated movements associated with movements caused by driving a vehicle
US8907867B2 (en) 2012-03-21 2014-12-09 Google Inc. Don and doff sensing using capacitive sensors
US10469916B1 (en) 2012-03-23 2019-11-05 Google Llc Providing media content to a wearable device
US10152193B1 (en) * 2012-07-12 2018-12-11 Eharmony, Inc. Systems and methods for online coaching in a matching service
US9817125B2 (en) 2012-09-07 2017-11-14 Microsoft Technology Licensing, Llc Estimating and predicting structures proximate to a mobile device
US9400682B2 (en) 2012-12-06 2016-07-26 Hewlett Packard Enterprise Development Lp Ranking and scheduling of monitoring tasks
WO2014134196A1 (en) * 2013-02-26 2014-09-04 Eastern Virginia Medical School Augmented shared situational awareness system
US9558220B2 (en) 2013-03-04 2017-01-31 Fisher-Rosemount Systems, Inc. Big data in process control systems
US10649424B2 (en) 2013-03-04 2020-05-12 Fisher-Rosemount Systems, Inc. Distributed industrial performance monitoring and analytics
US10866952B2 (en) 2013-03-04 2020-12-15 Fisher-Rosemount Systems, Inc. Source-independent queries in distributed industrial system
US10282676B2 (en) 2014-10-06 2019-05-07 Fisher-Rosemount Systems, Inc. Automatic signal processing-based learning in a process plant
US10649449B2 (en) 2013-03-04 2020-05-12 Fisher-Rosemount Systems, Inc. Distributed industrial performance monitoring and analytics
US10678225B2 (en) 2013-03-04 2020-06-09 Fisher-Rosemount Systems, Inc. Data analytic services for distributed industrial performance monitoring
US10386827B2 (en) 2013-03-04 2019-08-20 Fisher-Rosemount Systems, Inc. Distributed industrial performance monitoring and analytics platform
US10223327B2 (en) 2013-03-14 2019-03-05 Fisher-Rosemount Systems, Inc. Collecting and delivering data to a big data machine in a process control system
US10909137B2 (en) 2014-10-06 2021-02-02 Fisher-Rosemount Systems, Inc. Streaming data for analytics in process control systems
US9665088B2 (en) 2014-01-31 2017-05-30 Fisher-Rosemount Systems, Inc. Managing big data in process control systems
US9805163B1 (en) 2013-03-13 2017-10-31 Wellframe, Inc. Apparatus and method for improving compliance with a therapeutic regimen
US10255573B2 (en) * 2013-03-15 2019-04-09 WebINTENSIVE Software, LLC Systems, methods, and media for presenting interactive checklists
US10691281B2 (en) 2013-03-15 2020-06-23 Fisher-Rosemount Systems, Inc. Method and apparatus for controlling a process plant with location aware mobile control devices
US10296668B2 (en) 2013-03-15 2019-05-21 Fisher-Rosemount Systems, Inc. Data modeling studio
US9336502B2 (en) * 2013-04-30 2016-05-10 Oracle International Corporation Showing relationships between tasks in a Gantt chart
US20150153826A1 (en) * 2013-12-01 2015-06-04 Apx Labs, Llc Systems and methods for providing a virtual menu
US8856948B1 (en) * 2013-12-23 2014-10-07 Google Inc. Displaying private information on personal devices
US8811951B1 (en) 2014-01-07 2014-08-19 Google Inc. Managing display of private information
US20160063450A1 (en) * 2014-08-28 2016-03-03 Google Inc. Systems and Methods for Task Countdowns for Specified Tasks
US10168691B2 (en) 2014-10-06 2019-01-01 Fisher-Rosemount Systems, Inc. Data pipeline for process control system analytics
JP5913694B1 (en) * 2015-07-03 2016-04-27 株式会社リクルートホールディングス Order management system and order management program
US9842365B2 (en) * 2015-09-15 2017-12-12 Google Inc. Guided purchasing via smartphone
US10503483B2 (en) 2016-02-12 2019-12-10 Fisher-Rosemount Systems, Inc. Rule builder in a process control network
US10635461B2 (en) * 2017-02-03 2020-04-28 Dassault Systemes Solidworks Corporation Computer-aided interactive learning
US10311249B2 (en) 2017-03-31 2019-06-04 Google Llc Selectively obscuring private information based on contextual information
JP7042573B2 (en) * 2017-08-18 2022-03-28 三菱重工業株式会社 Evaluation system, evaluation method and program

Family Cites Families (313)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US556012A (en) * 1896-03-10 Excavator
US3973251A (en) 1975-04-25 1976-08-03 Stephans J Larry Timing device for exercise program with signalling means
US4283712A (en) 1976-12-22 1981-08-11 Goody Roy W Portable exercise totalizer
US4569026A (en) * 1979-02-05 1986-02-04 Best Robert M TV Movies that talk back
US4458331A (en) * 1981-10-09 1984-07-03 International Business Machines Corporation Interactive display terminal with alternating data processing and text processing sessions with text processing status line operable during data processing session
US6041365A (en) 1985-10-29 2000-03-21 Kleinerman; Aurel Apparatus and method for high performance remote application gateway servers
US5227614A (en) 1986-08-15 1993-07-13 Norand Corporation Core computer processor module, and peripheral shell module assembled to form a pocket size data capture unit
US4970683A (en) * 1986-08-26 1990-11-13 Heads Up Technologies, Inc. Computerized checklist with predetermined sequences of sublists which automatically returns to skipped checklists
US4815030A (en) * 1986-09-03 1989-03-21 Wang Laboratories, Inc. Multitask subscription data retrieval system
GB8719572D0 (en) * 1987-08-19 1987-09-23 Krebs M S Sigscan text retrieval system
US4916441A (en) * 1988-09-19 1990-04-10 Clinicom Incorporated Portable handheld terminal
US5201034A (en) * 1988-09-30 1993-04-06 Hitachi Ltd. Interactive intelligent interface
US4905163A (en) * 1988-10-03 1990-02-27 Minnesota Mining & Manufacturing Company Intelligent optical navigator dynamic information presentation and navigation system
US5133075A (en) 1988-12-19 1992-07-21 Hewlett-Packard Company Method of monitoring changes in attribute values of object in an object-oriented database
US5146221A (en) * 1989-01-13 1992-09-08 Stac, Inc. Data compression apparatus and method
US5349678A (en) 1991-08-21 1994-09-20 Norand Corporation Versatile RF data capture system
JP2938104B2 (en) 1989-11-08 1999-08-23 株式会社日立製作所 Shared resource management method and information processing system
US5278946A (en) * 1989-12-04 1994-01-11 Hitachi, Ltd. Method of presenting multimedia data in a desired form by comparing and replacing a user template model with analogous portions of a system
US5032083A (en) * 1989-12-08 1991-07-16 Augmentech, Inc. Computerized vocational task guidance system
US5251294A (en) 1990-02-07 1993-10-05 Abelow Daniel H Accessing, assembling, and using bodies of information
US5214757A (en) 1990-08-07 1993-05-25 Georesearch, Inc. Interactive automated mapping system
US6568595B1 (en) 1994-08-17 2003-05-27 Metrologic Instruments, Inc. System and method for carrying out electronic-commerce transactions using web documents embodying electronic-commerce enabling applets automatically launched and executed in response to reading url-encoded symbols pointing thereto
US5327529A (en) 1990-09-24 1994-07-05 Geoworks Process of designing user's interfaces for application programs
US5267147A (en) * 1990-10-19 1993-11-30 Heads Up Technologies, Inc. Portable checklist system
US5586218A (en) 1991-03-04 1996-12-17 Inference Corporation Autonomous learning and reasoning agent
US5317568A (en) 1991-04-11 1994-05-31 Galileo International Partnership Method and apparatus for managing and facilitating communications in a distributed hetergeneous network
US5630157A (en) 1991-06-13 1997-05-13 International Business Machines Corporation Computer organization for multiple and out-of-order execution of condition code testing and setting instructions
US5781913A (en) 1991-07-18 1998-07-14 Felsenstein; Lee Wearable hypermedium system
US5592664A (en) * 1991-07-29 1997-01-07 Borland International Inc. Database server system with methods for alerting clients of occurrence of database server events of interest to the clients
US5237684A (en) 1991-08-12 1993-08-17 International Business Machines Corporation Customized and versatile event monitor within event management services of a computer system
US5208449A (en) 1991-09-09 1993-05-04 Psc, Inc. Portable transaction terminal
US5454074A (en) * 1991-09-18 1995-09-26 The Boeing Company Electronic checklist system
US6418424B1 (en) 1991-12-23 2002-07-09 Steven M. Hoffberg Ergonomic man-machine interface incorporating adaptive pattern recognition based control system
WO1993014454A1 (en) 1992-01-10 1993-07-22 Foster-Miller, Inc. A sensory integrated data interface
US5481667A (en) * 1992-02-13 1996-01-02 Microsoft Corporation Method and system for instructing a user of a computer system how to perform application program tasks
US5388198A (en) * 1992-04-16 1995-02-07 Symantec Corporation Proactive presentation of automating features to a computer user
US5491651A (en) 1992-05-15 1996-02-13 Key, Idea Development Flexible wearable computer
US5285398A (en) * 1992-05-15 1994-02-08 Mobila Technology Inc. Flexible wearable computer
US5535323A (en) 1992-06-29 1996-07-09 Digital Equipment Corporation Method of and system for displaying context sensitive and application independent help information
JP2597802B2 (en) 1992-08-04 1997-04-09 インターナショナル・ビジネス・マシーンズ・コーポレイション Method for controlling an image capture device, image capture device and user interface
US5339395A (en) 1992-09-17 1994-08-16 Delco Electronics Corporation Interface circuit for interfacing a peripheral device with a microprocessor operating in either a synchronous or an asynchronous mode
US5513646A (en) 1992-11-09 1996-05-07 I Am Fine, Inc. Personal security monitoring system and method
US5601435A (en) * 1994-11-04 1997-02-11 Intercare Method and apparatus for interactively monitoring a physiological condition and for interactively providing health related information
US5879163A (en) 1996-06-24 1999-03-09 Health Hero Network, Inc. On-line health education and feedback system using motivational driver profile coding and automated content fulfillment
US5335276A (en) 1992-12-16 1994-08-02 Texas Instruments Incorporated Communication system and methods for enhanced information transfer
US5769085A (en) 1993-01-06 1998-06-23 Mitsubishi Jidosha Kogyo Kabushiki Kaisha Apparatus for detecting awareness of a vehicle driver and method thereof
US5720001A (en) 1993-04-16 1998-02-17 Compaq Computer Corporation Questionless case-based knowledge base and a method for constructing the same
JPH06332786A (en) 1993-05-25 1994-12-02 Fujitsu Ltd Data processing system with composite object
TW359781B (en) * 1993-05-25 1999-06-01 Casio Computer Co Ltd Animal image display controlling devices and method thereof
US5398021A (en) * 1993-07-19 1995-03-14 Motorola, Inc. Reliable information service message delivery system
GB9315448D0 (en) 1993-07-26 1993-09-08 Rank Xerox Ltd Recording and retrieval of information relevant to the activities of a user
KR100324989B1 (en) * 1993-11-08 2002-06-24 마츠시타 덴끼 산교 가부시키가이샤 Input display integrated information processing device
US5416730A (en) 1993-11-19 1995-05-16 Appcon Technologies, Inc. Arm mounted computer
US5806079A (en) 1993-11-19 1998-09-08 Smartpatents, Inc. System, method, and computer program product for using intelligent notes to organize, link, and manipulate disparate data objects
US5812865A (en) 1993-12-03 1998-09-22 Xerox Corporation Specifying and establishing communication data paths between particular media devices in multiple media device computing systems based on context of a user or users
US5493692A (en) * 1993-12-03 1996-02-20 Xerox Corporation Selective delivery of electronic messages in a multiple computer system based on context and environment of a user
US5555376A (en) 1993-12-03 1996-09-10 Xerox Corporation Method for granting a user request having locational and contextual attributes consistent with user policies for devices having locational attributes consistent with the user request
US5572401A (en) 1993-12-13 1996-11-05 Key Idea Development L.L.C. Wearable personal computer system having flexible battery forming casing of the system
CA2179523A1 (en) 1993-12-23 1995-06-29 David A. Boulton Method and apparatus for implementing user feedback
US5513342A (en) 1993-12-28 1996-04-30 International Business Machines Corporation Display window layout system that automatically accommodates changes in display resolution, font size and national language
US5566069A (en) 1994-03-07 1996-10-15 Monsanto Company Computer network for collecting and analyzing agronomic data
US5470233A (en) 1994-03-17 1995-11-28 Arkenstone, Inc. System and method for tracking a pedestrian
US5522026A (en) * 1994-03-18 1996-05-28 The Boeing Company System for creating a single electronic checklist in response to multiple faults
US5642129A (en) 1994-03-23 1997-06-24 Kopin Corporation Color sequential display panels
US5566337A (en) 1994-05-13 1996-10-15 Apple Computer, Inc. Method and apparatus for distributing events in an operating system
WO1995031773A1 (en) 1994-05-16 1995-11-23 Apple Computer, Inc. Switching between appearance/behavior themes in graphical user interfaces
US5646629A (en) 1994-05-16 1997-07-08 Trimble Navigation Limited Memory cartridge for a handheld electronic video game
US5704366A (en) * 1994-05-23 1998-01-06 Enact Health Management Systems System for monitoring and reporting medical measurements
WO1995035541A1 (en) 1994-06-22 1995-12-28 Molloy Bruce G A system and method for representing and retrieving knowledge in an adaptive cognitive network
US6321158B1 (en) 1994-06-24 2001-11-20 Delorme Publishing Company Integrated routing/mapping information
US6023683A (en) * 1994-08-10 2000-02-08 Fisher Scientific Company Electronic sourcing system and method
US5559520A (en) 1994-09-26 1996-09-24 Lucent Technologies Inc. Wireless information system for acquiring location related information
EP0712075A3 (en) 1994-11-08 1997-10-22 Canon Kk Information processing system which understands information and acts accordingly and method therefor
US6539336B1 (en) * 1996-12-12 2003-03-25 Phatrat Technologies, Inc. Sport monitoring system for determining airtime, speed, power absorbed and other factors such as drop distance
US6571279B1 (en) 1997-12-05 2003-05-27 Pinpoint Incorporated Location enhanced information delivery system
US5758257A (en) 1994-11-29 1998-05-26 Herz; Frederick System and method for scheduling broadcast of and access to video programs and other data using customer profiles
US6460036B1 (en) 1994-11-29 2002-10-01 Pinpoint Incorporated System and method for providing customized electronic newspapers and target advertisements
DE69531599T2 (en) 1994-12-20 2004-06-24 Sun Microsystems, Inc., Mountain View Method and device for finding and obtaining personalized information
US5659746A (en) 1994-12-30 1997-08-19 Aegis Star Corporation Method for storing and retrieving digital data transmissions
US5980096A (en) 1995-01-17 1999-11-09 Intertech Ventures, Ltd. Computer-based system, methods and graphical interface for information storage, modeling and stimulation of complex systems
US5553609A (en) 1995-02-09 1996-09-10 Visiting Nurse Service, Inc. Intelligent remote visual monitoring system for home health care service
US5959611A (en) 1995-03-06 1999-09-28 Carnegie Mellon University Portable computer system with ergonomic input device
JP4309480B2 (en) * 1995-03-07 2009-08-05 株式会社東芝 Information processing device
US5745110A (en) * 1995-03-10 1998-04-28 Microsoft Corporation Method and apparatus for arranging and displaying task schedule information in a calendar view format
US5710884A (en) * 1995-03-29 1998-01-20 Intel Corporation System for automatically updating personal profile server with updates to additional user information gathered from monitoring user's electronic consuming habits generated on computer during use
US5689708A (en) 1995-03-31 1997-11-18 Showcase Corporation Client/server computer systems having control of client-based application programs, and application-program control means therefor
US5911132A (en) 1995-04-26 1999-06-08 Lucent Technologies Inc. Method using central epidemiological database
US5832296A (en) 1995-04-26 1998-11-03 Interval Research Corp. Wearable context sensitive user interface for interacting with plurality of electronic devices of interest to the user
US5642303A (en) 1995-05-05 1997-06-24 Apple Computer, Inc. Time and location based computing
US5873070A (en) * 1995-06-07 1999-02-16 Norand Corporation Data collection system
JP2842308B2 (en) * 1995-06-30 1999-01-06 日本電気株式会社 Battery case mounting structure for electronic equipment
US6006251A (en) 1995-07-11 1999-12-21 Hitachi, Ltd. Service providing system for providing services suitable to an end user request based on characteristics of a request, attributes of a service and operating conditions of a processor
JP3729918B2 (en) * 1995-07-19 2005-12-21 株式会社東芝 Multimodal dialogue apparatus and dialogue method
US5715451A (en) * 1995-07-20 1998-02-03 Spacelabs Medical, Inc. Method and system for constructing formulae for processing medical data
US5826253A (en) 1995-07-26 1998-10-20 Borland International, Inc. Database system with methodology for notifying clients of any additions, deletions, or modifications occurring at the database server which affect validity of a range of data records cached in local memory buffers of clients
US5942986A (en) 1995-08-09 1999-08-24 Cedars-Sinai Medical Center System and method for automatic critical event notification
EP0759591B1 (en) 1995-08-18 2003-01-08 International Business Machines Corporation Event management service
CA2179337C (en) 1995-08-22 2000-11-14 Thomas F. Laporta Network-based migrating user agents for personal communication services
US5726688A (en) * 1995-09-29 1998-03-10 Ncr Corporation Predictive, adaptive computer interface
US6131067A (en) 1995-10-09 2000-10-10 Snaptrack, Inc. Client-server based remote locator device
US5701894A (en) 1995-11-09 1997-12-30 Del Mar Avionics Modular physiological computer-recorder
US6127990A (en) 1995-11-28 2000-10-03 Vega Vista, Inc. Wearable display and methods for controlling same
US6088689A (en) 1995-11-29 2000-07-11 Hynomics Corporation Multiple-agent hybrid control architecture for intelligent real-time control of distributed nonlinear processes
US5726660A (en) * 1995-12-01 1998-03-10 Purdy; Peter K. Personal data collection and reporting system
US6122960A (en) 1995-12-12 2000-09-26 Acceleron Technologies, Llc. System and method for measuring movement of objects
US5899963A (en) 1995-12-12 1999-05-04 Acceleron Technologies, Llc System and method for measuring movement of objects
US5752019A (en) 1995-12-22 1998-05-12 International Business Machines Corporation System and method for confirmationally-flexible molecular identification
EP0868701A2 (en) 1995-12-22 1998-10-07 Xerox Corporation Process support using system and user states
US5787279A (en) 1995-12-22 1998-07-28 International Business Machines Corporation System and method for conformationally-flexible molecular recognition
US5740037A (en) * 1996-01-22 1998-04-14 Hughes Aircraft Company Graphical user interface system for manportable applications
US6092725A (en) * 1997-01-24 2000-07-25 Symbol Technologies, Inc. Statistical sampling security methodology for self-scanning checkout system
JP3216514B2 (en) 1996-01-31 2001-10-09 株式会社デンソー Information service equipment
US6047327A (en) * 1996-02-16 2000-04-04 Intel Corporation System for distributing electronic information to a targeted group of users
US5910799A (en) * 1996-04-09 1999-06-08 International Business Machines Corporation Location motion sensitive user interface
US5971580A (en) 1996-04-17 1999-10-26 Raytheon Company Tactical awareness monitoring and direct response system
US5848396A (en) 1996-04-26 1998-12-08 Freedom Of Information, Inc. Method and apparatus for determining behavioral profile of a computer user
US5790974A (en) 1996-04-29 1998-08-04 Sun Microsystems, Inc. Portable calendaring device having perceptual agent managing calendar entries
US6047301A (en) 1996-05-24 2000-04-04 International Business Machines Corporation Wearable computer
US6014638A (en) * 1996-05-29 2000-01-11 America Online, Inc. System for customizing computer displays in accordance with user preferences
US5717747A (en) * 1996-05-31 1998-02-10 Lucent Technologies Inc. Arrangement for facilitating plug-and-play call features
US5945988A (en) 1996-06-06 1999-08-31 Intel Corporation Method and apparatus for automatically determining and dynamically updating user preferences in an entertainment system
US5966533A (en) 1996-06-11 1999-10-12 Excite, Inc. Method and system for dynamically synthesizing a computer program by differentially resolving atoms based on user context data
US5831594A (en) * 1996-06-25 1998-11-03 Sun Microsystems, Inc. Method and apparatus for eyetrack derived backtrack
FI102923B (en) 1996-08-08 1999-03-15 Nokia Mobile Phones Ltd Information printing system, procedure for printing information and terminals for printing information
US5832500A (en) 1996-08-09 1998-11-03 Digital Equipment Corporation Method for searching an index
US5689619A (en) 1996-08-09 1997-11-18 The United States Of America As Represented By The Secretary Of The Army Eyetracker control of heads-up displays
US5719743A (en) * 1996-08-15 1998-02-17 Xybernaut Corporation Torso worn computer which can stand alone
JP3688822B2 (en) 1996-09-03 2005-08-31 株式会社東芝 Electronic medical record system
US7040541B2 (en) 1996-09-05 2006-05-09 Symbol Technologies, Inc. Portable shopping and order fulfillment system
US6837436B2 (en) * 1996-09-05 2005-01-04 Symbol Technologies, Inc. Consumer interactive shopping system
US5924074A (en) 1996-09-27 1999-07-13 Azron Incorporated Electronic medical records system
US5974572A (en) * 1996-10-15 1999-10-26 Mercury Interactive Corporation Software system and methods for generating a load test using a server access log
US5938721A (en) 1996-10-24 1999-08-17 Trimble Navigation Limited Position based personal digital assistant
US5818446A (en) 1996-11-18 1998-10-06 International Business Machines Corporation System for changing user interfaces based on display data content
US5902347A (en) 1996-11-19 1999-05-11 American Navigation Systems, Inc. Hand-held GPS-mapping device
US6012152A (en) * 1996-11-27 2000-01-04 Telefonaktiebolaget Lm Ericsson (Publ) Software fault management system
US6198394B1 (en) 1996-12-05 2001-03-06 Stephen C. Jacobsen System for remote monitoring of personnel
US5905492A (en) 1996-12-06 1999-05-18 Microsoft Corporation Dynamically updating themes for an operating system shell
US5966126A (en) 1996-12-23 1999-10-12 Szabo; Andrew J. Graphic user interface for database system
FI106990B (en) 1996-12-31 2001-05-15 Nokia Mobile Phones Ltd A method of transmitting information to a user
US5798733A (en) 1997-01-21 1998-08-25 Northrop Grumman Corporation Interactive position guidance apparatus and method for guiding a user to reach a predetermined target position
US6098065A (en) 1997-02-13 2000-08-01 Nortel Networks Corporation Associative search engine
US5977968A (en) 1997-03-14 1999-11-02 Mindmeld Multimedia Inc. Graphical user interface to communicate attitude or emotion to a computer program
US6704785B1 (en) * 1997-03-17 2004-03-09 Vitria Technology, Inc. Event driven communication system
US5913030A (en) 1997-03-18 1999-06-15 International Business Machines Corporation Method and system for client/server communications with user information revealed as a function of willingness to reveal and whether the information is required
US5796952A (en) 1997-03-21 1998-08-18 Dot Com Development, Inc. Method and apparatus for tracking client interaction with a network resource and creating client profiles and resource database
US6289316B1 (en) 1997-03-25 2001-09-11 International Business Machines Corporation Progress notes model in a clinical information system
JPH10326289A (en) 1997-03-28 1998-12-08 Nippon Telegr & Teleph Corp <Ntt> Method for providing information and system therefor and storage medium for storing the same program
US6041331A (en) * 1997-04-01 2000-03-21 Manning And Napier Information Services, Llc Automatic extraction and graphic visualization system and method
US6272507B1 (en) 1997-04-09 2001-08-07 Xerox Corporation System for ranking search results from a collection of documents using spreading activation techniques
WO1998047284A1 (en) * 1997-04-14 1998-10-22 Thomson Consumer Electronics, Inc. System for forming program guide information for user initiation of control and communication functions
US5937160A (en) 1997-05-01 1999-08-10 Reedy Creek Technologies, Inc. Systems, methods and computer program products for updating hypertext documents via electronic mail
US6023729A (en) * 1997-05-05 2000-02-08 Mpath Interactive, Inc. Method and apparatus for match making
US6092101A (en) 1997-06-16 2000-07-18 Digital Equipment Corporation Method for filtering mail messages for a plurality of client computers connected to a mail service system
US6453334B1 (en) 1997-06-16 2002-09-17 Streamtheory, Inc. Method and apparatus to allow remotely located computer programs and/or data to be accessed on a local computer in a secure, time-limited manner, with persistent caching
US5991687A (en) 1997-07-02 1999-11-23 Case Corporation System and method for communicating information related to a geographical area
US6094625A (en) 1997-07-03 2000-07-25 Trimble Navigation Limited Augmented vision for survey work and machine control
US6108665A (en) 1997-07-03 2000-08-22 The Psychological Corporation System and method for optimizing behaviorial health care collection
US6081814A (en) 1997-07-07 2000-06-27 Novell, Inc. Document reference environment manager
US6442549B1 (en) 1997-07-25 2002-08-27 Eric Schneider Method, product, and apparatus for processing reusable information
US6111893A (en) 1997-07-31 2000-08-29 Cisco Technology, Inc. Universal protocol conversion
JP3714378B2 (en) * 1997-07-31 2005-11-09 セイコーエプソン株式会社 Device setup support device, device setup support method, and medium recording device setup support program
US6409086B1 (en) 1997-08-08 2002-06-25 Symbol Technolgies, Inc. Terminal locking system
US5974262A (en) 1997-08-15 1999-10-26 Fuller Research Corporation System for generating output based on involuntary and voluntary user input without providing output information to induce user to alter involuntary input
US6263268B1 (en) 1997-08-26 2001-07-17 Transcontech Corporation System and method for providing mobile automotive telemetry
JPH1196099A (en) * 1997-09-19 1999-04-09 Hitachi Ltd Service providing system
JP3922482B2 (en) 1997-10-14 2007-05-30 ソニー株式会社 Information processing apparatus and method
WO1999019823A2 (en) 1997-10-10 1999-04-22 Interval Research Corporation Methods and systems for providing human/computer interfaces
US6236768B1 (en) 1997-10-14 2001-05-22 Massachusetts Institute Of Technology Method and apparatus for automated, context-dependent retrieval of information
US6067084A (en) * 1997-10-29 2000-05-23 International Business Machines Corporation Configuring microphones in an audio interface
US5999943A (en) 1997-10-31 1999-12-07 Oracle Corporation Lob locators
US6061610A (en) 1997-10-31 2000-05-09 Nissan Technical Center North America, Inc. Method and apparatus for determining workload of motor vehicle driver
AU1519499A (en) 1997-11-07 1999-05-31 Via, Inc. Interactive devices and methods
US5953718A (en) 1997-11-12 1999-09-14 Oracle Corporation Research mode for a knowledge base search and retrieval system
US6134532A (en) 1997-11-14 2000-10-17 Aptex Software, Inc. System and method for optimal adaptive matching of users to most relevant entity and information in real-time
US6405264B1 (en) 1997-12-18 2002-06-11 Sun Microsystems, Inc. Marshaling and unmarshaling framework for supporting filters in a distributed object system
US6122348A (en) 1997-12-22 2000-09-19 Nortel Networks Corporation System and method for managing incoming communication events using multiple media options
US6427142B1 (en) 1998-01-06 2002-07-30 Chi Systems, Inc. Intelligent agent workbench
US5999932A (en) 1998-01-13 1999-12-07 Bright Light Technologies, Inc. System and method for filtering unsolicited electronic mail messages using data matching and heuristic processing
US6195622B1 (en) 1998-01-15 2001-02-27 Microsoft Corporation Methods and apparatus for building attribute transition probability models for use in pre-fetching resources
US6044415A (en) * 1998-02-27 2000-03-28 Intel Corporation System for transferring I/O data between an I/O device and an application program's memory in accordance with a request directly over a virtual connection
US6314384B1 (en) 1998-03-04 2001-11-06 Gerald E. Goetz Medication management apparatus
JPH11272689A (en) 1998-03-19 1999-10-08 Seiko Epson Corp Method and device for information retrieval processing and recording medium where information retrieval processing program is recorded
US6185534B1 (en) 1998-03-23 2001-02-06 Microsoft Corporation Modeling emotion and personality in a computer user interface
US6003082A (en) 1998-04-22 1999-12-14 International Business Machines Corporation Selective internet request caching and execution system
US6215405B1 (en) * 1998-04-23 2001-04-10 Digital Security Controls Ltd. Programmable temperature sensor for security system
JPH11306002A (en) 1998-04-23 1999-11-05 Fujitsu Ltd Editing device and editing method for gui environment
US6155960A (en) 1998-05-01 2000-12-05 Roberts; Donald Training dummy assembly in human form
US6105063A (en) 1998-05-05 2000-08-15 International Business Machines Corp. Client-server system for maintaining application preferences in a hierarchical data structure according to user and user group or terminal and terminal group contexts
US6188399B1 (en) * 1998-05-08 2001-02-13 Apple Computer, Inc. Multiple theme engine graphical user interface architecture
US6625135B1 (en) 1998-05-11 2003-09-23 Cargenie Mellon University Method and apparatus for incorporating environmental information for mobile communications
US6051365A (en) * 1998-05-21 2000-04-18 New England Science & Specialty Products, Inc. Waterless lithographic printing plates
US6640214B1 (en) * 1999-01-16 2003-10-28 Symbol Technologies, Inc. Portable electronic terminal and data processing system
US7010501B1 (en) * 1998-05-29 2006-03-07 Symbol Technologies, Inc. Personal shopping system
US6405159B2 (en) * 1998-06-03 2002-06-11 Sbc Technology Resources, Inc. Method for categorizing, describing and modeling types of system users
US6526035B1 (en) 1998-06-09 2003-02-25 Telefonaktiebolaget Lm Ericsson (Publ) Method and apparatus for identifying the source of a digital signal
US6256633B1 (en) 1998-06-25 2001-07-03 U.S. Philips Corporation Context-based and user-profile driven information retrieval
US6446109B2 (en) 1998-06-29 2002-09-03 Sun Microsystems, Inc. Application computing environment
US6169976B1 (en) * 1998-07-02 2001-01-02 Encommerce, Inc. Method and apparatus for regulating the use of licensed products
US6490579B1 (en) 1998-07-16 2002-12-03 Perot Systems Corporation Search engine system and method utilizing context of heterogeneous information resources
US6305007B1 (en) 1998-07-24 2001-10-16 Computer Associates Think, Inc. Object property meta model emulator for legacy data structures
US6262720B1 (en) * 1998-07-24 2001-07-17 The Boeing Company Electronic checklist system with checklist inhibiting
US6311162B1 (en) 1998-07-25 2001-10-30 Ernst F. Reichwein Interactive symptomatic recording system and methods
US6363377B1 (en) * 1998-07-30 2002-03-26 Sarnoff Corporation Search data processor
US6230111B1 (en) 1998-08-06 2001-05-08 Yamaha Hatsudoki Kabushiki Kaisha Control system for controlling object using pseudo-emotions and pseudo-personality generated in the object
US6442620B1 (en) 1998-08-17 2002-08-27 Microsoft Corporation Environment extensibility and automatic services for component applications using contexts, policies and activators
US7010603B2 (en) * 1998-08-17 2006-03-07 Openwave Systems Inc. Method and apparatus for controlling network connections based on destination locations
US7769620B1 (en) 1998-09-01 2010-08-03 Dennis Fernandez Adaptive direct transaction for networked client group
US6321279B1 (en) 1998-09-14 2001-11-20 Compaq Computer Corporation System for implementing intelligent I/O processing in a multi-processor system by redirecting I/O messages to a target central processor selected from the multi-processor system
US6507845B1 (en) * 1998-09-14 2003-01-14 International Business Machines Corporation Method and software for supporting improved awareness of and collaboration among users involved in a task
US6563430B1 (en) 1998-12-11 2003-05-13 Koninklijke Philips Electronics N.V. Remote control device with location dependent interface
US6487552B1 (en) 1998-10-05 2002-11-26 Oracle Corporation Database fine-grained access control
US6218958B1 (en) * 1998-10-08 2001-04-17 International Business Machines Corporation Integrated touch-skin notification system for wearable computing devices
US6643684B1 (en) 1998-10-08 2003-11-04 International Business Machines Corporation Sender- specified delivery customization
US6546425B1 (en) * 1998-10-09 2003-04-08 Netmotion Wireless, Inc. Method and apparatus for providing mobile and other intermittent connectivity in a computing environment
US6112246A (en) 1998-10-22 2000-08-29 Horbal; Mark T. System and method for accessing information from a remote device and providing the information to a client workstation
US6353853B1 (en) * 1998-10-26 2002-03-05 Triatek, Inc. System for management of building automation systems through an HTML client program
US6446076B1 (en) 1998-11-12 2002-09-03 Accenture Llp. Voice interactive web-based agent system responsive to a user location for prioritizing and formatting information
US6263317B1 (en) 1998-12-01 2001-07-17 Fogdog, Inc. Web sales channel conflict resolution system
US6438618B1 (en) 1998-12-16 2002-08-20 Intel Corporation Method and device for filtering events in an event notification service
US7225229B1 (en) 1998-12-18 2007-05-29 Tangis Corporation Automated pushing of computer user's context data to clients
US8181113B2 (en) * 1998-12-18 2012-05-15 Microsoft Corporation Mediating conflicts in computer users context data
US7107539B2 (en) * 1998-12-18 2006-09-12 Tangis Corporation Thematic response to a computer user's context, such as by a wearable personal computer
US6801223B1 (en) 1998-12-18 2004-10-05 Tangis Corporation Managing interactions between computer users' context models
US7137069B2 (en) 1998-12-18 2006-11-14 Tangis Corporation Thematic response to a computer user's context, such as by a wearable personal computer
US7076737B2 (en) 1998-12-18 2006-07-11 Tangis Corporation Thematic response to a computer user's context, such as by a wearable personal computer
US7055101B2 (en) 1998-12-18 2006-05-30 Tangis Corporation Thematic response to a computer user's context, such as by a wearable personal computer
US6842877B2 (en) * 1998-12-18 2005-01-11 Tangis Corporation Contextual responses based on automated learning techniques
US6791580B1 (en) 1998-12-18 2004-09-14 Tangis Corporation Supplying notifications related to supply and consumption of user context data
US7046263B1 (en) * 1998-12-18 2006-05-16 Tangis Corporation Requesting computer user's context data
US7779015B2 (en) * 1998-12-18 2010-08-17 Microsoft Corporation Logging and analyzing context attributes
US6513046B1 (en) * 1999-12-15 2003-01-28 Tangis Corporation Storing and recalling information to augment human memories
US6812937B1 (en) 1998-12-18 2004-11-02 Tangis Corporation Supplying enhanced computer user's context data
US6920616B1 (en) * 1998-12-18 2005-07-19 Tangis Corporation Interface for exchanging context data
US6466232B1 (en) 1998-12-18 2002-10-15 Tangis Corporation Method and system for controlling presentation of information to a user based on the user's condition
US6747675B1 (en) 1998-12-18 2004-06-08 Tangis Corporation Mediating conflicts in computer user's context data
US7231439B1 (en) * 2000-04-02 2007-06-12 Tangis Corporation Dynamically swapping modules for determining a computer user's context
US9183306B2 (en) * 1998-12-18 2015-11-10 Microsoft Technology Licensing, Llc Automated selection of appropriate information based on a computer user's context
US6349307B1 (en) * 1998-12-28 2002-02-19 U.S. Philips Corporation Cooperative topical servers with automatic prefiltering and routing
US6385589B1 (en) 1998-12-30 2002-05-07 Pharmacia Corporation System for monitoring and managing the health care of a patient population
US6442589B1 (en) 1999-01-14 2002-08-27 Fujitsu Limited Method and system for sorting and forwarding electronic messages and other data
US6282517B1 (en) 1999-01-14 2001-08-28 Autobytel.Com, Inc. Real time communication of purchase requests
US6430531B1 (en) 1999-02-04 2002-08-06 Soliloquy, Inc. Bilateral speech system
US6292796B1 (en) 1999-02-23 2001-09-18 Clinical Focus, Inc. Method and apparatus for improving access to literature
US6462759B1 (en) 1999-02-25 2002-10-08 International Business Machines Corporation Adaptive computer display screen window
US6317718B1 (en) 1999-02-26 2001-11-13 Accenture Properties (2) B.V. System, method and article of manufacture for location-based filtering for shopping agent in the physical world
US6294953B1 (en) 1999-02-26 2001-09-25 Axcess, Inc. High sensitivity demodulator for a radio tag and method
US6199099B1 (en) * 1999-03-05 2001-03-06 Ac Properties B.V. System, method and article of manufacture for a mobile communication network utilizing a distributed communication network
US6356905B1 (en) * 1999-03-05 2002-03-12 Accenture Llp System, method and article of manufacture for mobile communication utilizing an interface support framework
US6401085B1 (en) 1999-03-05 2002-06-04 Accenture Llp Mobile communication and computing system and method
US6353823B1 (en) 1999-03-08 2002-03-05 Intel Corporation Method and system for using associative metadata
JP4299911B2 (en) * 1999-03-24 2009-07-22 株式会社東芝 Information transfer system
US6636831B1 (en) 1999-04-09 2003-10-21 Inroad, Inc. System and process for voice-controlled information retrieval
US6507567B1 (en) * 1999-04-09 2003-01-14 Telefonaktiebolaget Lm Ericsson (Publ) Efficient handling of connections in a mobile communications network
US6289513B1 (en) 1999-06-01 2001-09-11 Isaac Bentwich Interactive application generation and text processing
US7103806B1 (en) 1999-06-04 2006-09-05 Microsoft Corporation System for performing context-sensitive decisions about ideal communication modalities considering information about channel reliability
US6484200B1 (en) 1999-06-11 2002-11-19 Sun Microsystems, Inc. Distinguished name scoping system for event filtering
US6553336B1 (en) * 1999-06-25 2003-04-22 Telemonitor, Inc. Smart remote monitoring system and method
US7000187B2 (en) * 1999-07-01 2006-02-14 Cisco Technology, Inc. Method and apparatus for software technical support and training
US6529723B1 (en) * 1999-07-06 2003-03-04 Televoke, Inc. Automated user notification system
US20010030664A1 (en) 1999-08-16 2001-10-18 Shulman Leo A. Method and apparatus for configuring icon interactivity
TW455769B (en) 1999-08-18 2001-09-21 Jian Huei Jiuan Eye-protection method and apparatus set up for monitor screen
US6885734B1 (en) * 1999-09-13 2005-04-26 Microstrategy, Incorporated System and method for the creation and automatic deployment of personalized, dynamic and interactive inbound and outbound voice services, with real-time interactive voice database queries
US6519552B1 (en) * 1999-09-15 2003-02-11 Xerox Corporation Systems and methods for a hybrid diagnostic approach of real time diagnosis of electronic systems
US6353398B1 (en) 1999-10-22 2002-03-05 Himanshu S. Amin System for dynamically pushing information to a user utilizing global positioning system
US6714977B1 (en) * 1999-10-27 2004-03-30 Netbotz, Inc. Method and system for monitoring computer networks and equipment
US6829639B1 (en) 1999-11-15 2004-12-07 Netvision, Inc. Method and system for intelligent global event notification and control within a distributed computing environment
US6704722B2 (en) * 1999-11-17 2004-03-09 Xerox Corporation Systems and methods for performing crawl searches and index searches
US6834208B2 (en) 1999-12-30 2004-12-21 Microsoft Corporation Method and apparatus for providing distributed control of a home automation and control system
US6652283B1 (en) 1999-12-30 2003-11-25 Cerego, Llc System apparatus and method for maximizing effectiveness and efficiency of learning retaining and retrieving knowledge and skills
US6546554B1 (en) * 2000-01-21 2003-04-08 Sun Microsystems, Inc. Browser-independent and automatic apparatus and method for receiving, installing and launching applications from a browser on a client computer
US6542889B1 (en) 2000-01-28 2003-04-01 International Business Machines Corporation Methods and apparatus for similarity text search based on conceptual indexing
US6868525B1 (en) * 2000-02-01 2005-03-15 Alberti Anemometer Llc Computer graphic display visualization system and method
SE0000385D0 (en) 2000-02-07 2000-02-07 Om Technology Ab A trading system
US20010045965A1 (en) 2000-02-14 2001-11-29 Julian Orbanes Method and system for receiving user input
AU2001247789A1 (en) * 2000-03-22 2001-10-03 Sidestep, Inc. Method and apparatus for dynamic information connection engine
WO2001075676A2 (en) 2000-04-02 2001-10-11 Tangis Corporation Soliciting information based on a computer user's context
US7464153B1 (en) 2000-04-02 2008-12-09 Microsoft Corporation Generating and supplying user context data
WO2001076120A2 (en) 2000-04-04 2001-10-11 Stick Networks, Inc. Personal communication device for scheduling presentation of digital content
US6327535B1 (en) 2000-04-05 2001-12-04 Microsoft Corporation Location beaconing methods and systems
US7051079B2 (en) 2000-05-16 2006-05-23 Fuji Photo Film Co., Ltd. Information intermediary apparatus, information management apparatus, and information communication system
US6712615B2 (en) * 2000-05-22 2004-03-30 Rolf John Martin High-precision cognitive performance test battery suitable for internet and non-internet use
US20020191034A1 (en) 2000-06-28 2002-12-19 Sowizral Henry A. Size conditioned visibility search system and method
US6477117B1 (en) 2000-06-30 2002-11-05 International Business Machines Corporation Alarm interface for a smart watch
US6707476B1 (en) * 2000-07-05 2004-03-16 Ge Medical Systems Information Technologies, Inc. Automatic layout selection for information monitoring system
US6738759B1 (en) 2000-07-07 2004-05-18 Infoglide Corporation, Inc. System and method for performing similarity searching using pointer optimization
EP1182541A3 (en) 2000-08-22 2005-11-30 Siemens Aktiengesellschaft System and method for combined use of different display/apparatus types with system controlled context dependant information representation
US7392486B1 (en) * 2000-09-19 2008-06-24 Honeywell International Inc. Method and apparatus for automatic display and removal of required synoptic pages in a checklist context
US6795806B1 (en) * 2000-09-20 2004-09-21 International Business Machines Corporation Method for enhancing dictation and command discrimination
US20020054130A1 (en) 2000-10-16 2002-05-09 Abbott Kenneth H. Dynamically displaying current status of tasks
US20020044152A1 (en) 2000-10-16 2002-04-18 Abbott Kenneth H. Dynamic integration of computer generated and real world images
WO2002033541A2 (en) 2000-10-16 2002-04-25 Tangis Corporation Dynamically determining appropriate computer interfaces
US6704812B2 (en) 2000-11-30 2004-03-09 International Business Machines Corporation Transparent and dynamic management of redundant physical paths to peripheral devices
US7110764B1 (en) 2000-12-21 2006-09-19 Cisco Technology, Inc. Mobility manager in a wireless IP core network
US20020082730A1 (en) * 2000-12-21 2002-06-27 Microsoft Corporation Universal media player
US6813587B2 (en) 2001-06-22 2004-11-02 Invensys Systems, Inc. Remotely monitoring/diagnosing distributed components of a supervisory process control and manufacturing information application from a central location
WO2003019325A2 (en) 2001-08-31 2003-03-06 Kent Ridge Digital Labs Time-based media navigation system
US7283992B2 (en) 2001-11-30 2007-10-16 Microsoft Corporation Media agent to suggest contextually related media content
US6766245B2 (en) 2002-03-14 2004-07-20 Microsoft Corporation Landmark-based location of users
US7395221B2 (en) 2002-05-09 2008-07-01 International Business Machines Corporation Intelligent free-time search
US7899915B2 (en) 2002-05-10 2011-03-01 Richard Reisman Method and apparatus for browsing using multiple coordinated device sets
US20040133600A1 (en) 2002-07-19 2004-07-08 Homer Gregg S. Rechargeable media distribution and play system
US7124125B2 (en) 2002-11-01 2006-10-17 Loudeye Corp. System and method for providing media samples on-line in response to media related searches on the internet
USD494584S1 (en) 2002-12-05 2004-08-17 Symbol Technologies, Inc. Mobile companion
AU2004208274B2 (en) 2003-01-28 2007-09-06 Samsung Electronics Co., Ltd. Method and system for managing media file database
US7162473B2 (en) * 2003-06-26 2007-01-09 Microsoft Corporation Method and system for usage analyzer that determines user accessed sources, indexes data subsets, and associated metadata, processing implicit queries based on potential interest to users
US20040267812A1 (en) 2003-06-26 2004-12-30 Microsoft Corporation Media platform
US7836010B2 (en) * 2003-07-30 2010-11-16 Northwestern University Method and system for assessing relevant properties of work contexts for use by information services
US20050193017A1 (en) 2004-02-19 2005-09-01 Han-Gyoo Kim Portable multimedia player/recorder that accesses data contents from and writes to networked device
US7561200B2 (en) 2004-07-26 2009-07-14 Csi Technology, Inc. Apparatus and method for automation of imaging and dynamic signal analyses
US7260453B2 (en) * 2005-01-06 2007-08-21 The Boeing Company Checklist error mitigation system
US20060259494A1 (en) 2005-05-13 2006-11-16 Microsoft Corporation System and method for simultaneous search service and email search
US7617200B2 (en) 2006-01-31 2009-11-10 Northwestern University Displaying context-sensitive ranked search results

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
No Search *
See references of EP1334438A1 *

Also Published As

Publication number Publication date
US20020054130A1 (en) 2002-05-09
AU2002211699A1 (en) 2002-04-29
US20070089067A1 (en) 2007-04-19
WO2002033578A8 (en) 2002-07-11
US7877686B2 (en) 2011-01-25

Similar Documents

Publication Publication Date Title
US7877686B2 (en) Dynamically displaying current status of tasks
KR102509975B1 (en) User Interfaces for Watches
CN109690481B (en) Method and apparatus for dynamic function row customization
CN110825299B (en) Reduced size user interface
CN108089727B (en) Handwriting keyboard for screen
CN111666354B (en) Structured advice
CN114564113A (en) Handwriting input on electronic devices
CN113795815A (en) Clock face for electronic equipment
EP1603031A2 (en) Strategies for providing just-in-time user assistance
US20110087974A1 (en) User interface controls including capturing user mood in response to a user cue
ZA200507985B (en) A windowing and controlling system thereof comprising a computer device
EP3876085A1 (en) Self-learning digital interface
CN113093955A (en) Emoticons and preset replies
JP2011081778A (en) Method and device for display-independent computerized guidance
Elguera Paez et al. Elderly users and their main challenges usability with mobile applications: a systematic review
Lu et al. Exploring spatial UI transition mechanisms with head-worn augmented reality
US20210125584A1 (en) Avatar group control concept
CN116382557A (en) Navigating a user interface using hand gestures
Lavric et al. An industry-adapted AR training method for manual assembly operations
CN110651242B (en) Apparatus, method and graphical user interface for touch input processing
Karam et al. A study on the use of semaphoric gestures to support secondary task interactions
CN117581188A (en) Interaction with a note user interface
Frauenberger et al. Pattern design in the context space: A methodological framework for auditory display design
Soares Designing Culturally Sensitive Icons for User Interfaces: An approach for the Interaction Design of smartphones in developing countries
Ramos Accessibility Guidelines Proposal for the Interaction Design of Mobile Applications: Creating a More Inclusive User Experienc

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

AK Designated states

Kind code of ref document: C1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NO NZ PL PT RO RU SD SE SG SI SK SL TJ TM TR TT TZ UA UG US UZ VN YU ZA ZW

AL Designated countries for regional patents

Kind code of ref document: C1

Designated state(s): GH GM KE LS MW MZ SD SL SZ TZ UG ZW AM AZ BY KG KZ MD RU TJ TM AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

D17 Declaration under article 17(2)a
121 Ep: the epo has been informed by wipo that ep was designated in this application
REG Reference to national code

Ref country code: DE

Ref legal event code: 8642

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: COMMUNICATION PURSUANT TO RULE 69 EPC (EPO FORM 1205A OF 120803)

122 Ep: pct application non-entry in european phase
NENP Non-entry into the national phase

Ref country code: JP

DPE2 Request for preliminary examination filed before expiration of 19th month from priority date (pct application filed from 20040101)