Search Images Maps Play YouTube News Gmail Drive More »
Sign in
Screen reader users: click this link for accessible mode. Accessible mode has the same essential features but works better with your reader.

Patents

  1. Advanced Patent Search
Publication numberUS20100064238 A1
Publication typeApplication
Application numberUS 12/618,698
Publication date11 Mar 2010
Filing date13 Nov 2009
Priority date13 Feb 2004
Also published asUS7557797, US7620915, US8816956, US9417716, US20050179650, US20050179651, US20050179652, US20050179655, US20050179663, US20100060607, US20110128224, US20110134039, US20150062014, US20170139491
Publication number12618698, 618698, US 2010/0064238 A1, US 2010/064238 A1, US 20100064238 A1, US 20100064238A1, US 2010064238 A1, US 2010064238A1, US-A1-20100064238, US-A1-2010064238, US2010/0064238A1, US2010/064238A1, US20100064238 A1, US20100064238A1, US2010064238 A1, US2010064238A1
InventorsLester Frank Ludwig
Original AssigneeLester Frank Ludwig
Export CitationBiBTeX, EndNote, RefMan
External Links: USPTO, USPTO Assignment, Espacenet
Electronic document editing employing multiple cursors
US 20100064238 A1
Abstract
Two independently adjustable and positionable cursors are employed in a visual interface for editing an electronic document. The document may comprise text symbols, text objects, and graphics objects, among others. If both cursors are located in close proximity within the document, then these cursors may be simultaneously displayed in a single window. Otherwise, two areas of the document, each comprising one of the cursors, may be simultaneously displayed using a separate window for each cursor, or alternately selected and displayed in a single window. Copy or cut operations may be made with one cursor, while paste operations may be repeated using the second cursor. Cursor locations may be left unchanged, or they may be moved between or within editing operations. The two cursors may be freely controlled by an enhanced pointing device, such as a high-dimensional touchpad providing at least four independently adjustable parameters.
Images(19)
Previous page
Next page
Claims(19)
1. A method for facilitating computer editing of an electronic document, the electronic document comprising a plurality of objects that are graphically rendered in a layout comprising a plurality of unique locations, the method comprising:
positioning a first cursor at a first screen location within the electronic document responsive to user input from a first pair of parameters produced by a touchpad, the touchpad providing at least four independently adjustable parameters;
positioning a second cursor at a second screen location within the electronic document responsive to user input from a second pair of parameters produced by the touchpad, wherein the first and second cursors are independently displayable, positionable, and operable at any of the plurality of unique locations of the electronic document;
defining a selection string via the first cursor based upon a selected contiguous arrangement of at least one object of the plurality of objects;
transferring the selection string to a location within the electronic document as determined by the second location of the second cursor; and
wherein the first and second sensors are of different types located in different regions of the user interface device.
2. The method according to claim 1, wherein the transferring comprises a copy-and paste operation.
3. The method according to claim 1, wherein the transferring comprises a cut-and paste operation.
4. The method according to claim 1, the method further comprising:
positioning the second cursor at a third location within the electronic document; and
transferring the selection string to a location within the electronic document as determined by the third location of the second cursor.
5. The method according to claim 1, wherein the first and second cursors are simultaneously displayed and positionable in a common graphical user interface (GUI) window.
6. The method according to claim 1, wherein the first and second locations are separated by a distance such that the first and second cursors are unable to be simultaneously displayed in a common GUI window, and wherein only one of the first and second cursors is displayed when the GUI window displays a portion of the electronic document that comprises a selected one of the first and second cursors.
7. The method according to claim 6, wherein the GUI window displays a portion of the electronic document that comprises a last positioned one of the first and second cursors.
8. The method according to claim 6, wherein the GUI window displays a portion of the electronic document comprising one of the first and second cursors that is associated with a last manipulated one of the first and second sensors.
9. The method according to claim 6, wherein a user interface selection event selectively causes the GUI window to display either a portion of the electronic document that comprises the first cursor, or a portion of the electronic document that comprises the second cursor.
10. The method according to claim 1, wherein the first and second locations are separated by a distance such that the first and second cursors are unable to be simultaneously displayed in a common graphical user interface (GUI) window, and wherein only one of the first and second cursors are positionable when the GUI window displays a portion of the electronic document that comprises a selected one of the first and second cursors.
11. The method according to claim 10, wherein the GUI window displays a portion of the electronic document that comprises a last positioned one of the first and second cursors.
12. The method according to claim 10, wherein the GUI window displays a portion of the electronic document comprising one of the first and second cursors that is associated with a last manipulated one of the first and second sensors.
13. The method according to claim 10, wherein a user interface selection event selectively causes the GUI window to display either a portion of the electronic document that comprises the first cursor, or a portion of the electronic document that comprises the second cursor.
14. The method according to claim 1, wherein the first cursor is displayed within a first graphical user interface (GUI) window, and the second cursor is displayed within a second graphical user interface window.
15. The method according to claim 1, wherein the first cursor is positionable within a first graphical user interface (GUI) window, and the second cursor is positionable within a separate, second graphical user interface window.
16. The method according to claim 1, wherein one or more of the plurality of objects comprise a text symbol.
17. The method according to claim 1, wherein one or more of the plurality of objects comprise a text object.
18. The method according to claim 1, wherein one or more of the plurality of objects comprise a graphics object.
19. A method for facilitating computer editing of an electronic document, the electronic document comprising a plurality of objects that are graphically rendered in a layout comprising a plurality of unique locations, the method comprising:
positioning a first cursor at a first screen location within the electronic document responsive to user input from a first pair of parameters produced by a user interface touchpad, said touchpad providing at least four independently adjustable parameters;
positioning a second cursor at a second screen location within the electronic document responsive to user input from a second pair of parameters produced by the user interface touchpad, wherein the first and second cursors are independently displayable, positionable, and operable at any of the plurality of unique locations of the electronic document; and
defining a selection string, via the first cursor, based upon a selected contiguous arrangement of at least one object of the plurality of objects, wherein the selection string is capable of being transferred to a location within the electronic document as determined by the second location of the second cursor; and
wherein the first and the second cursor are of different types located in different regions of the user interface device.
Description
    CROSS REFERENCE TO RELATED APPLICATIONS
  • [0001]
    This application is a continuation of U.S. application Ser. No. 10/779,368, filed on Feb. 13, 2004.
  • BACKGROUND OF THE INVENTION
  • [0002]
    1. Field of the Invention
  • [0003]
    This invention relates to user interface devices for use with a computer, and in particular to computer mice, trackballs, touchpads, multiple-parameter pointing and data entry devices, and user interface metaphors.
  • [0004]
    2. Description of the Related Art
  • [0005]
    User interface devices for data entry and graphical user interface pointing have been known for many years. The most common devices include the computer mouse (usually attributed to English, Engelbart, and Bennan “Display-Selection Techniques for Text Manipulation, IEEE Transactions on Human Factors in Electronics, pp. 5-15, vol. HFE-8, No. 1, March 1967), the trackball, the touchpad in both finger-operated (for example, the various finger-operated devices produced by Symantec Corp., of Springfield, Oreg.) and stylus-operated (for example, products used with desktop workstation computers—Wacom Technology Corp., of Vancouver, Wash.) versions, and display-overlay touchscreens. Other historical and exotic devices include various types of light pens and the Data Love™ (produced by VPL Research, Inc., of Redwood City, Calif.).
  • [0006]
    Most user interface devices for data entry and graphical user interface pointing commonly used with computers or with equipment providing computer-like user interfaces have two wide-range parameter adjustment capabilities that are usually assigned to the task of positioning a screen cursor within a two-dimensional display. In many cases, one, two, or three binary-valued “discrete-event” controls are provided, typically in the form of spring-loaded push-buttons.
  • [0007]
    More recently, computer mice have emerged that provide an additional “scroll” finger-wheel adjustment (for example, between two control buttons) to provide a third wide-range parameter adjustment capability (for example, various products developed by Logitech Inc., of Fremont, Calif.). A mouse of this configuration is often referred to as a “Scroll Mouse” since this third wide-range parameter is typically assigned the task of positioning a vertical scroll bar in an actively selected window. This additional finger-wheel adjustment may also operate as a spring-loaded push-button, thus providing an additional binary-valued “discrete-event” control. Typically this additional binary-valued “discrete-event” control is used to turn on and off an automatic scrolling feature which controls the rate and direction of automatic scrolling according to vertical displacement of the displayed cursor.
  • SUMMARY OF THE INVENTION
  • [0008]
    In accordance with embodiments of the invention, two independently adjustable and positionable cursors are employed in a visual interface for editing an electronic document. The document may comprise text symbols, text objects, and graphics objects, among others. If both cursors are located in close proximity within the document, then these cursors may be simultaneously displayed in a single window. Otherwise, two areas of the document, each comprising one of the cursors, may be simultaneously displayed using a separate window for each cursor, or alternately selected and displayed in a single window. Copy or cut operations may be made with one cursor, while paste operations may be repeated using the second cursor. Cursor locations may be left unchanged, or they may be moved between or within editing operations. The two cursors may be freely controlled by an enhanced pointing device, such as a mouse with an added touchpad or trackball, or sequentially selected and controlled by a conventional mouse.
  • [0009]
    In accordance with embodiments of the invention, a traditional hand-movable computer mouse is configured with an additional user interface sensor. For convenience, the term “user interface sensor” will be used herein to collectively refer to devices such as trackballs, touchpads, mouse devices, scroll-wheels, joysticks, and other such devices.
  • [0010]
    In one aspect of the invention, the addition of a user interface sensor provides alternative physical modalities for the same pair of adjustable parameters so that a user may switch between using the user interface device as a traditional hand-movable computer mouse and using the user interface device as a trackball or touchpad.
  • [0011]
    In another aspect of the invention, the addition of a user interface sensor provides alternative resolution modalities for the same pair of adjustable parameters so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one level of parameter adjustment resolution, and using the invention as a trackball or touchpad, for example, to obtain a different level of parameter adjustment resolution.
  • [0012]
    In another aspect of the invention, the addition of a user interface sensor provides alternative types of warping modalities for the same pair of adjustable parameters so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one type of parameter adjustment (for example, linear) and using the invention as a trackball or touchpad, for example, to obtain a different type of parameter adjustment (for example, logarithmic, gamma-corrected, arccosine, exponential, etc.).
  • [0013]
    In another aspect of the invention, the addition of a user interface sensor provides alternative offset modalities for the same pair of adjustable parameters so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one type of centering of parameter adjustment and using the invention as a trackball or touchpad, for example, to obtain a different centering of parameter adjustment.
  • [0014]
    In another aspect of the invention, the addition of a user interface sensor may be used to provide additional parameters that may be simultaneously controlled.
  • [0015]
    In another aspect of the invention, the addition of a user interface sensor may be used to provide additional parameters that are of a different isolated context from those assigned to a traditional hand-movable computer mouse.
  • [0016]
    In a further more detailed aspect of the invention, the addition of a touchpad may be used to provide many additional parameters that are of a different context than those of a traditional hand-movable computer mouse.
  • [0017]
    In a further more detailed aspect of the invention, the touchpad may be a null-contact touchpad adapted to measure at least one maximum spatial span of contact in a given direction.
  • [0018]
    In a yet further detailed aspect of the invention, the null-contact touchpad is adapted to measure at least one maximum spatial span of contact in a given direction at a specifiable angle.
  • [0019]
    In an additional further detailed aspect of the invention, the null-contact touchpad is adapted to measure pressure applied to the null-contact touchpad.
  • [0020]
    In a further more detailed aspect of the invention, the touchpad may comprise a pressure sensor array touchpad adapted to measure, among other things, one or more of the following: the rocking position of a contacting finger in a given direction; the rotational position of a contacting finger; the pressure of a contacting finger; and parameters relating to a plurality of contacting fingers.
  • [0021]
    In another aspect of the invention the addition of a user interface sensor may be realized via a replaceable module accepted by an adaptation of a traditional hand-movable computer mouse. In this implementation, a user may initially obtain the invention in one configuration and field-modify it to another configuration.
  • [0022]
    In another aspect of the invention, a traditional hand-movable computer mouse may be implemented as a removable module in a laptop computer or other affiliated equipment, and may include a wireless link with the laptop computer or other affiliated equipment.
  • [0023]
    In yet a further aspect of the invention, a traditional hand-movable computer mouse is implemented as a removable module in a laptop computer or other affiliated equipment, and the mouse further comprises a user interface sensor.
  • [0024]
    In another aspect of the invention, a traditional hand-movable computer mouse additionally comprises a trackball or touchpad, for example. In this aspect, the mouse comprises a wireless link to an associated computer or other affiliated equipment.
  • [0025]
    In another aspect of the invention, a visual display is provided.
  • [0026]
    In another aspect of the invention, an auditory output is provided.
  • [0027]
    In another aspect of the invention, two or more individual user interface sensors may be combined without incorporation of such sensors with a traditional hand-movable computer mouse.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • [0028]
    The above and other aspects, features and advantages of the present invention will become more apparent upon consideration of the following description of preferred embodiments taken in conjunction with the accompanying drawing figures, wherein:
  • [0029]
    FIGS. 1 a-1 i illustrate various exemplary implementations involving merging, selecting, multiplexing, and preprocessing distributed in various ways between the body of the user interface device and an associated piece of equipment;
  • [0030]
    FIGS. 2 a-2 c depict an embodiment of the invention comprising a traditional mouse fitted with a trackball, illustrating three exemplary button configurations;
  • [0031]
    FIGS. 3 a-3 c depict an embodiment of the invention comprising a traditional mouse fitted with a touchpad, illustrating three exemplary button configurations;
  • [0032]
    FIGS. 4 a-4 d depict various exemplary degrees of freedom that may be measurably assigned to a trackball for interactively controlling parameters in a user interface;
  • [0033]
    FIGS. 5 a-5 d depict various exemplary degrees of freedom that may be measurably assigned to a touchpad for interactively controlling parameters in a user interface;
  • [0034]
    FIG. 6 depicts an exemplary implementation of the invention directed towards the control of both a traditional text cursor and a dual-scrollbar in a typesetting application;
  • [0035]
    FIG. 7 depicts an exemplary implementation of the invention directed towards the active selection from a clip-art or symbol library and adjustment of positioning or other attributes of the active selection in a drawing or layout application;
  • [0036]
    FIG. 8 is a flowchart showing exemplary operations and overhead involved in selecting and adjusting a specific pair of parameters from among a larger group of adjustable parameters;
  • [0037]
    FIGS. 9 a-9 b illustrate how the exemplary operations and overhead depicted in FIG. 8 introduce excessive overhead in situations where many parameters with a larger group of adjustable parameters must be adjusted in pairs;
  • [0038]
    FIGS. 10 a-10 b illustrate one technique for adding an additional scroll-wheel to a conventional scroll-wheel mouse;
  • [0039]
    FIGS. 11 a-11 b illustrate a simple example of open adjustments being made within various levels of hierarchy of graphical object groupings;
  • [0040]
    FIG. 12 illustrates aspects of the 3D orientation of an object in 3-dimensional space, and in particular the three coordinates of position and the three angles of rotation;
  • [0041]
    FIGS. 13 a-13 b illustrate one technique for using two cursors in a text cut-and-paste operation; and
  • [0042]
    FIGS. 14 a-14 d illustrate exemplary embodiments of a mouse where the traditional mouse buttons have been replaced by trackballs or touchpads.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • [0043]
    In the following description, reference is made to the accompanying drawing figures which form a part hereof, and which show by way of illustration specific embodiments of the invention. It is to be understood by those of ordinary skill in this technological field that other embodiments may be utilized, and structural, electrical, as well as procedural changes may be made without departing from the scope of the present invention.
  • [0044]
    By way of overview, a number of different applications that take advantage of the functionality of additional, wide-range adjustment parameters will now be discussed. In one example, an additional finger-wheel adjustment device providing a third, wide-range parameter adjustment capability is typically assigned to vertical scroll bar positioning. In accordance with the invention, such a design may be supplemented with a fourth, wide-range parameter adjustment capability so that a horizontal scroll bar position control may be achieved. With the increasing popularity of the web (with many web pages wide enough to require horizontal scrolling) and publisher layout tools (typically involving pages wide enough to require horizontal scrolling), as well as the need for simultaneous interactive horizontal and vertical scrolling actions that do not disturb a screen cursor location when using “zoom” controls, a fourth wide-range parameter adjustment capability in traditional user interface devices for data entry and graphical user interface pointing is quite valuable.
  • [0045]
    There are many other potential uses for additional wide-range adjustment parameters in traditional user interface devices for data entry and graphical user interface pointing. Some opportunities have wide-range applicability, such as in providing interactive separate adjustment of the selections for “cut” or “copy” operations from the interactive adjustment of insertion location or selection for a “paste” operation. Other opportunities are more specialized but still widely applicable, such as making an active selection from a clip-art or symbol library and adjusting the position or other attributes of said active selection in a drawing or layout application. Yet other opportunities may be very specialized, such as in 3D modeling, data visualization, advanced color adjustment, lighting control, machine control, or audio and image signal processing.
  • [0046]
    There are many opportunities for adjusting the same two widely-varying parameters in more than one way. For example, one user interface modality (such as normal mouse operation) may be used for normal parameter adjustment, while a second user interface modality may be used for adjustments involving a different resolution, warping (i.e., logarithmic, gamma-corrected, arccosine, exponential, etc.), centering offset, etc. Another important case is where the same two widely-varying parameters are controlled with the same resolution, warping, offset, etc., but in a different user interface modality (e.g., a trackball or touchpad may have some advantages in certain situations over use of a traditional mouse). A more widely applicable example is that of responding to and preventing handlwrist/arm fatigue and injury. A traditional mouse fitted with an additional user interface sensor allows a user to interchangeably enter information with either the mouse body or another user interface sensor, changing which user interface modality is used (obtaining the same results with either) to relieve fatigue or pain, or prevent injury.
  • [0047]
    More specifically, the addition of a user interface sensor provides many opportunities for alternative means of adjustment of a common pair of adjustable parameters. The user may benefit from having both adjustment modalities available, changing modalities as needed or desired. For example:
      • A user may simply switch between using the invention as a traditional hand-movable computer mouse and using the invention as another kind of user interface sensor.
        • The user may benefit from having both modalities available to avoid or in response to hand fatigue.
        • The user may also benefit from having both modalities available due to the type of pointing or data entry interaction needed—depending on the case, one type of modality may perform better than another.
      • The trackball, touchpad, or other user interface sensor apparatus may be used to provide alternative resolution modalities so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one level of parameter adjustment resolution and using the invention as a user interface sensor to obtain a different level of parameter adjustment resolution.
      • The trackball or touchpad may be used to provide alternative warping modalities for the same pair of adjustable parameters so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one type of parameter adjustment (for example, linear) and using the invention as another kind of user interface sensor to obtain a different type of parameter adjustment resolution (for example, logarithmic, gamma-corrected, arccosine, exponential, etc.).
      • The user interface sensor may be used to provide alternative offset modalities for a common pair of adjustable parameters so that a user may switch between using the invention as a traditional hand-movable computer mouse to obtain one centering of parameter adjustment and using the invention as another kind of user interface sensor to obtain a different centering of parameter adjustment. These modalities can provide one or more “location bookmarks” for cursor location, each affiliated with a sub-context within an interactive application.
  • [0054]
    Further, the addition of another user interface sensor provides many opportunities for the simultaneous adjustment of additional parameters that may or may not require simultaneous interactive control. The traditional computer mouse may be used to simultaneously adjust two parameters while the additional user interface sensor may be configured to allow the fingers to simultaneously adjust at least two additional parameters. In some applications, these additional parameters may be closely related to those assigned to the traditional computer mouse. For example, the traditional computer mouse may be used to simultaneously adjust the location within a window of a text, graphic, or other object, while the additional user interface sensor allows the fingers to be used to adjust the type or attributes of the text, graphic, or other object. In other applications, these additional parameters may be of a different isolated context from those assigned to the traditional computer mouse. For example, the traditional computer mouse may be used to simultaneously adjust two parameters dealing with affairs within an active application window, while the addition of another user interface sensor allows the fingers to be used to adjust at least two additional parameters dealing with broader window system affairs such as vertical and horizontal scrollbars, window selection, window resizing, etc., or intermediate-level affairs such as zoom control, help-window navigation, clip-art selection, etc. Another application would be to provide separate adjustment of selections for “cut” or “copy” operations from the adjustment of insertion location or selection for a “paste” operation.
  • [0055]
    In instances of the invention involving the addition of a touchpad, the touchpad may be configured and/or enhanced to allow the fingers to adjust three or more additional interactive measured parameters. These additional interactive measured parameters may be assigned to control more sophisticated interactive affairs such as 3-dimensional space position, 3-dimensional space orientation, color model navigation, image or audio processing parameter settings, etc.
  • [0056]
    The additional interactive measured parameters (above the two typically associated with traditional touchpads) may be provided in a number of ways. For example, the touchpad may be a relatively low-cost null-contact touchpad that has been adapted to measure at least one maximum spatial span of contact in a given direction. The user may also control an additional parameter by varying the width between the spatial extremes of a single point of contact (i.e., how much finger flesh makes contact with the pad) or multiple points of contact (i.e., the spread between two contacting fingers). As there are two geometrically orthogonal sensing directions on a touchpad, this provides the user with a method for controlling four total parameters from a touchpad. Further, rotational transformations or other methodologies may be used to measure the angle of rotation of an oblong contact profile. The measured angle may be used as a fifth interactive parameter, and/or used to adapt the measurement of maximum spatial span of contact in an arbitrary angle. The null-contact touchpad may be further adapted to measure pressure applied to the null-contact touchpad via, for example, use of an attached pressure sensor. The pressure may be used as a sixth interactive parameter, and may or may not have rapid pressure changes recognized as ‘tap’ or ‘click’ events.
  • [0057]
    Another way to provide the additional interactive measured parameters (above the two typically associated with traditional touchpads) with a touchpad is to implement the touchpad with a pressure sensor array. Through use of operations effectively amounting to image processing, a pressure sensor array touchpad can be adapted to measure the rocking position of a contacting finger in two orthogonal directions, as well as the rotational position and average pressure of a contacting finger. Thus a pressure sensor array touchpad can be adapted to provide up to six widely variable interactive adjustable parameters from the contact of a single finger. A pressure sensor array touchpad can be further adapted to measure parameters relating to a plurality of contacting fingers.
  • [0058]
    All of these considerations and others demonstrate the potential value in providing the addition of another user interface sensor to a traditional hand-movable computer mouse. In the descriptions to follow, various implementations and exemplary applications of exemplary embodiments are considered and explained.
  • 1. Exemplary Signal Flow and Processing
  • [0059]
    The invention provides for a wide range of signal flow and processing configurations and implementations. FIGS. 1 a-1 i illustrate various exemplary implementations involving merging, selecting, multiplexing, and preprocessing distributed in various ways between the body of the user interface device and an associated piece of equipment. FIGS. 1 a-1 d concern the aggregated pair of user interface sensors in isolation, while FIGS. 1 e-1 i address arrangements where some functions of the invention are performed in the associated external equipment. It is noted that the invention further provides for any of these exemplary functionalities, as well as other functionalities, to be combined or made selectable. In any of the exemplary implementations disclosed herein, power may be supplied to these implementations by the associated external equipment or by other devices such as batteries, storage capacitors, photoelectric devices, and the like.
  • [0060]
    FIG. 1 a shows an implementation 100 a featuring two user interface sensors 101, 102, each of which may be a particular type of user interface sensor (which again may be a trackball, touchpad, mouse, or other user interface device) that can be collocated within the common physical enclosure 100 (demarcated by the dotted-line boundary). First user interface sensor 101 produces signal 108 and second user interface sensor 102 produces another signal 109, which are directed to a merge or select function 103. The merge or select function produces outgoing signal 110 which is provided to associated external equipment. Here signals 108, 109 (from first and second user interface sensors 101, 102) would typically lose their individual identities within the outgoing signal 110 and as such may be used or processed interchangeably (without individual attribution to either the first or second user interface sensor) by the associated external equipment.
  • [0061]
    Merge or select function 103 may take several forms in various implementations. For example, in one embodiment it may simply be fixed to only perform a merge operation. In another embodiment it may only provide a selection function; here the selection function may be controlled by the user using a switch or some sort of action, or the selection may be remotely controlled by external equipment. As an alternative, merge or select function 103 may instead provide a user adjustable “merge” or “select” function.
  • [0062]
    FIG. 1 b shows an implementation that is similar to that of FIG. 1 a. The primary difference is that the FIG. 1 b implementation 100 b replaces merge or select function 103 of FIG. 1 a with multiplex function 104 to produce outgoing signal 110. Here signals 108, 109 (from first and second user interface sensors 101, 102) retain their individual identities within outgoing signal 110 and as such may be used or processed separately by the associated external equipment.
  • [0063]
    FIG. 1 c shows implementation 100 c which is similar in many respects to that of FIG. 1 a. However, the FIG. 1 c embodiment utilizes preprocessor 105 applied to signal 109 to produce processed signal 109 a. The processed signal 109 a, along with signal 108, is directed to merge or select function 103, resulting in outgoing signal 110. Preprocessor 105 may thus introduce a pre-processing step (such as resolution modification, warping modification, offset modification, etc.) on signal 109 to produce a signal of distinguished value from that of signal 108.
  • [0064]
    FIG. 1 d illustrates another exemplary implementation 100 d which is similar to that of FIG. 1 c but with an additional preprocessor 106 applied to signal 108. Preprocessor 106 produces processed signal 108 a, which along with signal 109 a, is directed to merge or select function 103 to produce outgoing signal 110. Preprocessor 106 may therefore introduce a pre-processing step (such as resolution modification, warping modification, offset modification, etc.) on signal 108 to produce a signal of either equivalent or distinguished value from that of signal 109 a.
  • [0065]
    FIG. 1 e shows implementation 100 b of FIG. 1 b (which features two user interface sensors 101, 102 and a multiplex function 104 producing an outgoing signal 110) used in conjunction with subsequent functions provided by associated external equipment 150 a. These subsequent functions are shown within the functional boundary 150 of the associated external equipment 150 a.
  • [0066]
    Outgoing signal 110 from the common physical enclosure 100, is presented to demultiplexer 117 within external equipment 150 a. Demultiplexer 117 produces signal 118 corresponding to or associated with pre-multiplexed signal 108, and an additional signal 119 corresponding to or associated with the pre-multiplexed signal 109. Here, signals 118, 119 are presented to merge or select function 113 producing merged or selected signal 120. This implementation is functionally similar or equivalent to that of FIG. 1 a except that the various types of merge or selection functions 103 are provided within the associated external equipment 150 a (for example in software, perhaps within an application where it is customized for the needs of that application) rather than being provided within physical unit 100 b.
  • [0067]
    FIG. 1 f shows implementation 100 b of FIG. 1 b used in conjunction with subsequent functions provided by associated external equipment 150 b, and similar to that of FIG. 1 e, except signal 119 produced by demultiplexer 117 is directed to preprocessor 115 to produce processed signal 119 a before being sent to merge or selection function 113. This implementation is thus functionally similar or equivalent to that of FIG. 1 c except that the various types of merge or selection 103 and preprocessor 105 functions in FIG. 1 c are provided within the associated external equipment 150 b (for example in software, perhaps within an application where it is customized for the needs of that application) rather than being provided within the physical unit 100 b.
  • [0068]
    FIG. 1 g shows implementation 100 b of FIG. 1 b used in conjunction with subsequent functions provided by associated external equipment 150 c. This arrangement expands on that shown in FIG. 1 fin that signal 118 produced by demultiplexer 117 is directed to preprocessor 116 to produce processed signal 118 a. The processed signal is then sent to merge or selection function 113. This implementation is thus functionally similar or equivalent to that of FIG. 1 d except that the various types of merge or selection 103 and preprocessor 105, 106 functions in FIG. 1 d are provided within the associated external equipment 150 c (for example in software, perhaps within an application where it is—customized for the needs of that application) rather than being provided within the physical unit 100 b.
  • [0069]
    FIG. 1 h shows implementation 100 b of FIG. 1 b used in conjunction with subsequent functions provided by associated external equipment 150 d. Here again, as in the arrangement of FIG. 1 f, signal 119 produced by demultiplexer 117 is directed to preprocessor 115 to produce processed signal 119 a. In contrast to other embodiments, processed signal 119 a is not directed to merge or selection 113 function and retains its identity for use by a different destination from that of signal 118 within associated external equipment 150 d.
  • [0070]
    As a final illustrative example in this series, FIG. 1 i shows implementation 100 b of FIG. 1 b used in conjunction with subsequent functions provided by associated external equipment 150 e. Here, as in the arrangement of FIG. 1 g, signals 118, 119 produced by demultiplexer 117 are directed to preprocessors 115, 116 to produce processed signals 118 a, 119 a. In contrast to other embodiments, processed signals 118 a, 119 a are not directed to merge or selection 113 function and thus retain their identity for use by differing destinations within associated external equipment 150 e.
  • [0071]
    Having presented various exemplary signal flows and processing realizations provided for by the invention, attention is now directed to exemplary implementations utilizing specific types of user interface sensors. It is to be understood that the various sensors, techniques and methods disclosed herein may be implemented using computer software, hardware, and firmware, and combinations thereof.
  • 2. Implementations Utilizing Specific Types of Additional User Interface Sensors
  • [0072]
    In this section, a number of exemplary implementations of the invention utilizing various types of additional user interface sensors added to an original user interface sensor or device. The first three sections address cases where the original user interface sensor is a movable mouse and the additional user interface sensor is a trackball, touchpad, or other exemplary technology, including additional scroll-wheels. Then exemplary adaptations of trackballs and touchpads, each traditionally used to provide simultaneous adjustment of two interactive widely-varying parameters, are extended to provide simultaneous adjustment of as many as six interactive widely-varying parameters and other forms of control. This section continues by presenting exemplary implementations where the original user interface sensor is not a mouse, where there are a plurality of additional user interface sensors, where there is a visual display or auditory output, where the additional user interface sensor is a removable module, and where the implementation itself is a removable module.
  • [0073]
    2.1 Trackball Implementations of Additional User Interface Sensors
  • [0074]
    FIGS. 2 a-2 c illustrate a number of implementations provided for by the invention where a trackball controller is used as an additional user interface sensor apparatus incorporated into a traditional hand-movable computer mouse. In each of these implementations it is understood that the trackball may be freely operated without disturbing previous or currently-varying parameter adjustments made by the mouse.
  • [0075]
    In one implementation, a trackball controller is added to the top surface of a conventional computer mouse as depicted in FIG. 2 a. The conventional mouse buttons may be located in various places in view of the presence of the trackball and in synergy with it. In the configuration depicted in FIG. 2 a, buttons 201 and 202 are located on the surface of mouse 200; button 201 being on the left of the trackball and button 202 being on the right of the trackball.
  • [0076]
    In a second configuration depicted in FIG. 2 b, buttons 231 and 232 are now separated and located on the sides of mouse 230 as is the case with many trackball interfaces; button 231 being on the left side of mouse 230 and button 232 being on the right side of the mouse.
  • [0077]
    In a third possible configuration depicted in FIG. 2 c, elongated buttons 261 and 262 are shown located on the surface of the mouse 260; button 261 wraps around the left of the trackball and button 262 wraps around the right of the trackball. The elongated buttons 261 and 262 may be positioned so that a user can readily and rapidly move fingers from trackball 265 to buttons 261 and 262, or even operate one of these buttons with one finger while another finger contacts trackball 265.
  • [0078]
    It is noted that unlike the touchpad described below, the trackball has an effectively unconfined range of contiguous data entry.
  • [0079]
    2.2 Touchpad Implementations of Additional User Interface Sensor
  • [0080]
    FIGS. 3 a-3 c illustrate a number of exemplary implementations provided for by the invention where a touchpad controller is-used as an additional user interface sensor incorporated into a traditional hand-movable computer mouse. In each of these implementations it is understood that the touchpad may be fi-eely operated without disturbing previous or currently-varying parameter adjustments made by the mouse.
  • [0081]
    In one implementation, trackball 205 in FIG. 2 a can be replaced with touchpad 305 as shown in FIG. 3 a. Additionally, this touchpad implementation can also support the alternative button configurations of FIGS. 2 b and 2 c. By way of illustration, FIG. 3 b shows buttons 331,332 positioned on either side of the mouse body 330, while FIG. 3 c shows elongated buttons 361,362 on the surface of the mouse wrapping around either side of touchpad 365.
  • [0082]
    It is noted that unlike the trackball, the touchpad typically has a confined maximum range of data entry by contiguous operation of a finger, stylus, etc.
  • [0083]
    2.3 Other Implementations of Additional User Interface Sensors.
  • [0084]
    In accordance with other embodiments, the invention provides for still other types of additional user interface sensors. In each of these embodiments it is understood that any of these user interface sensors may be freely operated without disturbing previous or currently varying parameter adjustments made by the mouse or other associated device.
  • [0085]
    As a first example, an X-Y joystick may be used in place of the trackball or touchpad described above. The joystick may have a spring-return or may retain the last position it was placed. Similar to the touchpad and unlike the trackball, the X-Y joystick typically has a confined maximum range of travel.
  • [0086]
    As another example, two or more scrolling finger wheels may be used in place of the trackball or touchpad described above. The scrolling finger wheels may be implemented with an unconfined maximum range of travel similar to the trackball, or with a confined range of travel like the touchpad and X-Y joystick. In this embodiment, it may be advantageous to have one or more finger scroll-wheels mounted with its adjustment direction perpendicular to that of another finger scroll-wheel so that each wheel may be appropriately associated with vertical and horizontal scroll bars of a window, or other useful orthogonally-based user interface metaphor. For example, looking ahead to FIGS. 10 a and 10 b, embodiments 1000 are depicted comprising the usual components of a scroll-wheel mouse (mouse body 1001, buttons 1011 and 1012, and the usual scroll wheel 1021) complemented with an additional scroll wheel 1022 with adjustment direction perpendicular to that of finger scroll-wheel 1021.
  • [0087]
    As another example, two or more rotating control knobs may be used in place of the trackball or touchpad described above. Like the scrolling finger wheels, the rotating control knobs may be implemented with an unconfined maximum range of travel like the trackball or with a confined range of travel like the touchpad and X-Y joystick.
  • [0088]
    The invention also provides for more exotic types of user interface sensor technologies—for example, proximity detectors of various types (RF, infrared, etc.), video cameras (using any of the techniques provided in U.S. Pat. No. 6,570,078), and other emerging and unforeseen technologies—to be used as the additional user interface sensor aggregated in the same physical enclosure as the first user interface sensor.
  • [0089]
    2.4 Larger Numbers of Interactively Widely-Varying Adjustable Parameters from the Additional User Interface Sensor
  • [0090]
    In accordance with embodiments of the invention, additional user interfaces may be used to capture larger numbers (i.e., more than two) of widely-varying adjustable parameters from the additional user interface sensor. Some examples are provided here, but many others are possible as may be readily understood by one skilled in the art.
  • [0091]
    FIGS. 4 a-4 d address the case of the trackball. FIG. 4 a illustrates a freely-rotating trackball 205 and the two principle orthogonal adjustment directions 401,402 that are responsively resolved and measured in traditional trackball user interface devices. However, at least two other physical degrees of freedom may be readily exploited, and at least six total parameters can be interactively adjusted and measured. FIG. 4 b shows the application of downward pressure 403 on trackball 205. Such pressure 403 may be applied without disturbing current values established in orthogonal adjustment directions 401, 402. Further, the trackball may be implemented so that downward pressure 403 may be applied while simultaneously adjusting trackball 205 in orthogonal adjustment directions 401, 402, particularly if the signal produced by the measurement of downward pressure 403 incorporates a modest “grace” zone of non-responsiveness for light pressure values. Downward pressure impulses may alternatively be sensed and treated as discrete event “taps,” as commonly used in contemporary touchpad interfaces found in laptop computers, for example.
  • [0092]
    FIG. 4 c shows the application of “yaw” rotation 404 (i.e., rotation around the vertical axis) of trackball 205. This yaw rotation 404 may be applied without disturbing current values established in orthogonal adjustment directions 401,402 and can be readily measured and adjusted as a widely-varying parameter. Further, by grasping trackball 205 (or other operational methods), the yaw rotation 404 and traditional orthogonal adjustment directions 401,402 may be independently and simultaneously adjusted. It is also noted that in principle up to six widely-varying physical degrees of freedom can be simultaneously measured from a properly configured trackball 205 by placing the trackball 205 in a cradle that senses not only downward pressure 403 or displacement but also lateral pressure or displacement. As shown in FIG. 4 d, both forward-backward, non-rotational force 405 and left-right, non-rotational force 406 may be applied to the trackball in a manner that the values of force or displacement in each of these directions 405,406 can be independently measured. Thus, by grasping trackball 205 (or other operational methods), three rotational directions of orientation 401,402,404 and three non-rotational directions of force or displacement 403, 405,406 may be independently and simultaneously adjusted by a user and measured as six independent interactively adjustable user interface parameters. These correspond, effectively, to measurable adaptations of the six degrees of freedom of an orientable object in 3-dimensional space as found in classical mechanics and aeronautics—that is:
      • “roll” rotation (adapted to 401)
      • “pitch” rotation (adapted to 402)
      • “up-down: displacements (adapted to 403)
      • “yaw” rotation (adapted to 404)
      • “forward-backward” displacements (adapted to 405)
      • “left-right” displacements (adapted to 406).
  • [0099]
    Most trackball sensing technologies use optically based techniques for sensing the two traditional components of rotation (“roll” and “pitch”) of the trackball. The trackball itself may be configured with an optical pattern on it with spatially varying reflectivity for a range of the light spectrum. The pattern may be such that it can spatially vary light reflectively in these two traditional components of trackball rotation. Alternatively, two spatially varying reflectivity patterns, each active at different ranges of the light spectrum or light polarization, may be superimposed or integrated with the trackball.
  • [0100]
    A number of approaches may be used to obtain measurements for all three directions of rotation. In one completely optical approach, a second or third spatially varying reflectivity pattern active at, respectively, a second or third portion of the light spectrum (or light polarization if available) may be superimposed or integrated with the patterns employed for traditional “roll” and “pitch” rotation sensing, and an additional optical source and sensor is used to obtain measurement of the added varying reflectivity pattern. Depending on the pattern(s) used, sensor signals may be directly usable or may require processing of the three primitive signals measured by the sensors to obtain a clean decomposition of the measurement signals into independent “roll,” “pitch,” and “yaw” signals independently responsive to the “roll,” “pitch,” and “yaw” components of trackball rotation.
  • [0101]
    As another alternative, the trackball may include internally, or on its surface, or both, materials with spatially varying patterns of magnetic properties, capacitive properties, electromagnetic properties, ultrasonic acoustic properties, resonance phenomena of any of these properties, polarization phenomena of any of these properties, etc., individually or in combination, each of which may be active at specific ranges of polarization, frequencies, etc. These may be used together with or in place of optical measurement approaches. Again, depending on the pattern(s) used, sensor signals may be directly usable or may require processing of the three primitive signals measured by the sensors to obtain a clean decomposition of the measurement signals into independent “roll,” “pitch,” and “yaw” signals independently responsive to the “roll,” “pitch,” and “yaw” components of trackball rotation as is clear to one skilled in the art. It is also noted that the third component of rotation of the freely-rotating trackball may be interpreted or even measured as a discrete “click” event.
  • [0102]
    Similarly, a number of approaches may be used to obtain measurements for one, two, or three directions of non-rotational trackball displacement. For example, the trackball may be secured in a saddle allowing free rotation of the trackball but causing any displacement actions on the trackball to invoke displacements of the saddle. The saddle displacement may be measured with a displacement sensor which itself may comprise one or more pressure, optical, resistive, capacitive, magnetic, electromagnetic, continuous-range sensors, switches, etc. It is also noted that one or more components of displacement of the freely-rotating trackball may be interpreted or even measured as a discrete “click” event.
  • [0103]
    FIGS. 5 a-5 d turn now to the case of the touchpad. FIG. 5 a illustrates touchpad 305 and the two principle orthogonal data entry directions 501, 502 that are responsively resolved and measured in traditional touchpad user interface devices. The touchpad shown in FIGS. 5 a-5 d, which provides at least four other physical degrees of freedom, may be implemented using the techniques presented in U.S. Pat. No. 6,570,078, for example.
  • [0104]
    FIG. 5 b illustrates the use of downward pressure 503 in the context of a touchpad. In contemporary touchpad interfaces, such as those found in laptop computers for example, such downward pressure 503 is sensed and utilized as discrete event “taps.” However, downward pressure 503 may also be measured and adjusted as an independent and simultaneously interactive widely-varying parameter. Further, as illustrated in FIG. 5 c, the rotational angle 504 of a finger contacting a touchpad with rough-elliptical contact boundary can also be measured as a widely-varying parameter. In FIG. 5 d, both forward-backward 505 and left-right 506 components of the tilt of a contacting finger can additionally be measured as independent and simultaneously interactive widely-varying parameters.
  • [0105]
    The sensing of multiple fingers, the application of contact syntaxes and grammars, and other user interface control expansions of an adequately configured touchpad may also be achieved using, for example, the techniques presented in U.S. Pat. No. 6,570,078.
  • [0106]
    The invention also provides for larger numbers (i.e., more than two) of widely varying adjustable parameters from other types of user interface sensor technologies. In the case of an X-Y joystick, the joystick may be configured to rotate on its axis, pulled in and out, fitted with a knob or trackball, etc., in a measurable fashion to provide additional and simultaneous interactively adjustable parameters. In the cases of finger scroll-wheels and rotational knobs, three or more of these devices may be provided. When implementing video cameras, known techniques for the extraction of additional parameters may be used. Examples of the various types of video extraction techniques that may be used are presented in U.S. Pat. No. 6,570,078.
  • [0107]
    2.5 Non-Mouse User Interface Sensors
  • [0108]
    In one of its most abstract forms, the invention involves the incorporation of two conventional user interface sensors (such as a mouse, trackball, touchpad, joystick, etc.) into an embodiment where the user may freely use both of the user interface sensors individually or simultaneously with the same hand. As such, the invention provides for implementations that do not include a mouse as one of the user interface sensors. For example, two or more individual user interface sensors can be combined without need of a traditional hand-movable computer mouse. Such an implementation may be accomplished by implementing one of the possible user interface sensors in place of the traditional hand-movable computer mouse where taught in other descriptions of the invention. This configuration may be useful when built into a laptop computer, control console, musical instrument, and test instrument, among others.
  • [0109]
    In one exemplary implementation of a non-mouse embodiment, a trackball and touchpad may be arranged in the same physical enclosure so that the front, middle, or back of the palm may freely operate a conventional trackball while one or more selected extended or arching fingers may simultaneously or alternatively operate a touchpad. In this example, the touchpad may be a conventional touchpad providing two widely-varying simultaneously interactive parameters from a single finger contact, or the touchpad may be a more enhanced version providing as many as six widely-varying and simultaneously interactive parameters from a single finger contact. The touchpad may also be configured to accept multiple points of contact, recognize gestures, support syntax and grammatical constructions using, for example, the teachings provided by U.S. Pat. No. 6,570,078.
  • [0110]
    In another non-mouse implementation, two trackballs may be arranged in the same physical enclosure. In one possible arrangement, the two trackballs may be positioned so that they lie parallel to the length of the hand, enabling the front, middle, or back of the palm to freely operate a first trackball while one or more extended or arching fingers may simultaneously or alternatively operate the second trackball.
  • [0111]
    In another arrangement, the two trackballs may be positioned so that they lie parallel to the width of the hand, so that the fingers and/or thumb on the left side of the hand may operate a leftmost trackball while the remaining fingers and/or thumb on the right side of the hand may individually or simultaneously operate a rightmost trackball. In each of these arrangements, either or both of the trackballs may be a conventional trackball providing two widely-varying and simultaneously interactive parameters, or it may be a more enhanced trackball providing as many as six widely-varying and simultaneously interactive parameters as described earlier.
  • [0112]
    In addition to the just-described embodiments, alternative arrangements, such as the combination of a palm-operated trackball and a recessed joystick, and others, are also provided for by the invention.
  • [0113]
    2.6 Use of More than One Additional User Interface Sensor
  • [0114]
    Typically the arrangements of two non-mouse user interface sensors described above in Section 2.5 can also be applied to embodiments of the invention where a mouse user interface sensor is used. In such embodiments, a mouse user interface sensor is supplemented with at least two additional user interface sensors ergonomically arranged so that the two additional user interface sensors may be simultaneously or alternatively operated by the same hand. If these embodiments are further configured so the mouse body is readily moved with adequate precision via the back of the operating hand, then all three user interface sensors may be simultaneously or alternatively operated by the same hand in an ergonomically advantageous manner.
  • [0115]
    FIGS. 14 a-14 d illustrate some exemplary embodiments of the just-described features. FIG. 14 a illustrates a mouse where traditional mouse buttons have been replaced by trackballs 1405 a, 1405 b. These trackballs 1405 a, 1405 b may accept a downward pressure impulse and as such act as the traditional mouse buttons. However, trackballs 1405 a, 1405 b are also adjustable and each readily provides two or more additional widely—variable and simultaneously adjustable parameters in addition to the two parameters adjusted by moving the mouse body 1400.
  • [0116]
    FIG. 14 b shows a similar arrangement where traditional mouse buttons have been replaced by touchpads 1435 a, 1435 b. If desired, these touchpads may accept a downward pressure impulse and as such act as traditional mouse buttons, but are also adjustable as touchpads and as such each readily provides two or more additional widely-variable and simultaneously adjustable parameters. As described in Section 2.5, a single hand may be positioned to comfortably operate simultaneously or alternatively both trackballs or both touchpads. If these embodiments are further configured so the mouse body is readily moved with adequate precision via the back of the operating hand, then either of these embodiments readily provides six to twelve widely-variable and simultaneously adjustable parameters.
  • [0117]
    Other configurations are of course possible. For example, the configurations of FIGS. 14 a and 14 b may be blended as depicted in FIG. 14 c, or in its mirror image. As another example, FIG. 14 d illustrates a more extreme realization comprising a left-fingers/thumb trackball 1465 a, a right-fingers/thumb trackball 1465 b, a palm trackball 1465 c, and a traditional clickable scroll-wheel 1468. Yet another alternative is to replace one or more of the trackballs of the FIG. 14 d embodiment with a touchpad user interface sensor.
  • [0118]
    2.7 Incorporation of Visual Display and Auditory Output
  • [0119]
    If desired, any of the mouse and non-mouse embodiments may further include a visual display. The visual display may provide details of adjustable parameter values, operation modalities, etc. The visual display may be physically associated with a physical enclosure (such as that of a traditional computer mouse), or may be displayed on the computer screen or display of other associated equipment.
  • [0120]
    Alternatively or additionally, any of the mouse and non-mouse embodiments may further provide auditory output. The auditory output may provide details of adjustable parameter values, operation modalities, error conditions in usage of the invention, a condition relating to elapsed time or other metric of consistent use of a single usage modality, etc. The auditory output associated with the invention may be physically associated with a physical enclosure (such as that of a traditional computer mouse), or may be produced by speakers or headsets affiliated with the computer or other associated equipment.
  • [0121]
    2.8 Provisions for Field Installation or Replacement of Additional User Interface Sensor
  • [0122]
    The invention also provides for the user interface sensor to be implemented using a replaceable module accepted by an adaptation of a traditional computer mouse. In this implementation a user may initially obtain the invention in one configuration and field modify it to another configuration.
  • [0123]
    2.9 Implementation as a Module Removable from Affiliated Equipment
  • [0124]
    The invention also provides for a traditional computer mouse to be implemented as a removable module in a laptop computer or other affiliated equipment, and may include a wireless link with such devices. In particular, this removable module may further include one or more user interface sensors, with these sensors operable as a traditional trackball or touchpad when the invention is stowed in the laptop computer or other affiliated equipment in such a way that the invention's traditional hand-movable computer mouse modality is unmovable and hence unusable.
  • 3. Exemplary Applications
  • [0125]
    Departing now from the range of extreme realization and embodiments of the invention, attention is directed towards particular applications of the invention. A number of examples of various embodiments of the invention in a wide range of applications will now be presented. Many of these applications are viable with only the simplest physical embodiments of the invention (for example, those suggested by FIGS. 2 a-2 c and 3 a-3 c). In the discussion that follows, particular note is directed towards the discussion in Section 3.3 involving FIGS. 8 and 9 a-9 b. Although the discussion is motivated by a graphical layout application, the principles of the discussion in Section 3.3 involving FIGS. 8 and 9 a-9 b are very general, and the discussion illustrates the power of the invention for various applications in almost directly quantifiable terms.
  • [0126]
    3.1 Wrist/Hand/Arm-Fatigue Relief and Prevention Application
  • [0127]
    The danger and damage stemming from extensive continuous or mis-postured mouse usage to wrist, hand, and arms are sadly misfortunate and increasingly well recognized. As the present invention provides a plurality of different user interface sensors, it is well suited for use in responding to and preventing wrist/hand/arm fatigue due to overuse. In one exemplary implementation, user interface parameters can be interchangeably adjusted with either the movement of the mouse body or the use of an integrated trackball or touchpad with identical task results. Thus a user with a tiring hand can change at will the user interface sensor employed according to how the hand feels or the nature of a specific task. In addition, to prevent fatigue or injury, the user can also switch back and forth between moving the mouse body and using the trackball/touchpad either by free choice or by following auditory or visual prompting from a time or usage monitor.
  • [0128]
    3.2 Double-Scrollbar Application
  • [0129]
    Contemporary mice often feature a small rotating wheel between the buttons for use in operating the vertical scroll bar of a window without changing the position of the mouse. In one particular application of the invention, the left-right sensing capability of the trackball or touchpad may be used to add a similar capability for horizontal scroll bars of a window.
  • [0130]
    In a trackball implementation, a user can move the vertical bar 611 of FIG. 6 up by rotating trackball 205 away from him/herself, or one can move the vertical bar 621 down by rotating trackball 205 towards him/herself. By rotating trackball 205 to the left, the user can move the horizontal bar 621 left. Similarly, the user can move the horizontal bar 621 right by rotating the trackball 205 to the right.
  • [0131]
    In a touchpad implementation, a user can move scroll bar 611 up by sliding the finger away from her/himself or move scroll bar 611 down by sliding the contacting finger towards her/himself; similarly, the user can move the scroll bar 621 left by sliding the finger to the left or move the scroll bar 621 right by sliding the contacting finger to the right.
  • [0132]
    In another implementation, the vertical and horizontal scroll bars may be adjusted with a conventional scroll-wheel mouse that has been fitted with an additional scroll-wheel. FIGS. 10 a and 10 b depict exemplary embodiments 1000 of such an arrangement which comprise the usual components of a scroll-wheel mouse including mouse body 1001, buttons 1011 and 1012, and traditional scroll wheel 1021, with these components further complemented by an additional scroll wheel 1022 with adjustment direction perpendicular to that of finger scroll-wheel 1021. FIG. 10 a illustrates an arrangement where the additional scroll-wheel is located closer to the user while FIG. 10 b illustrates an arrangement where the additional scroll-wheel is located farther away from the user. In each of these arrangements, the two scroll-wheels are shown co-centered with respect to the mouse body; for simultaneous adjustment it may be advantageous to locate the additional scroll-wheel 1022 to one side or the other of the centered positions shown in FIGS. 10 a and 10 b. One approach useful for supporting both left-handed and right-handed users, which may provide additional utility, would be to provide two off-centered additional scroll-wheels, one on either side of the center line of the mouse body 1001 and conventional scroll-wheel 1021.
  • [0133]
    3.3 Traditional 2D Layout, CAD, and Graphics Applications
  • [0134]
    In most contemporary 2-dimensional layout and graphics applications, such as those commonly used for viewgraphs, page layout, electronic CAD, etc., numerous mouse operations are necessary for each of the many types of object attribute modification, etc. Typically, these mouse operations are required because the mouse only allows for the interactive adjustment of two widely-varying parameters at a time, and the user must change context several times as the parameters adjusted by the mouse are chosen, adjusted, and then replaced with another pair of parameters. The present invention is useful in many of these circumstances because it allows for more than two parameters to be adjusted at the same time.
  • [0135]
    FIG. 7 shows an example of a session involving the authoring of a viewgraph. The viewgraph authoring task showcased in this example includes the creation of a flowchart diagram (here depicting a business workflow process) and as such also illustrates related needs and attributes of a 2D CAD program involving layout of a diagram (such as a circuit, algorithm, etc.) or physical object (such as a PC board, control panel, semiconductor photolithography mask, etc.). In that this example further involves drawing, the example also illustrates the related needs and attributes of a paint-box or electronic drafting application.
  • [0136]
    In this broadly representative application, application window 700 is shown comprising menu area 700 a and drawing area 700 b. Within the drawing area, viewgraph title 701 and portion 702 of the flowchart to be drawn, comprising thus far a sequence of flowgraph objects connected by arrows, have already been entered and rendered. New flowgraph objects may be introduced in standard fashion by selecting the type of new object desired from a palette, initially putting an instance of the selected object type in a convenient place in the drawing area, adjusting the size, orientation, color, and/or other attributes, and putting into final position. Often a number of the last few steps are interactively cycled through multiple times before the newly introduced object is adequately drawn and the user directs their attention to the next task. In this example, the palette of available objects of a specific high-level task is shown as an overlapping stack of three sub-class palettes 703 a, 703 b, 703 c, each providing a selection of available objects within that sub-class. Here, for example, sub-class pallete 703 a has been selected (as indicated by the heavy line) from other available objects 705, within the sub-class palette 703 a. Upon selection, an initial highly adjustable rendering of a specific instance 714 of the selected object 704 appears in a convenient location, which may be selected by the user.
  • [0137]
    The specific instance 714, rendered in this highly adjustable initial state, is typically surrounded by graphical handles 716 which facilitate sizing, positioning within the drawing area, and often at least rotational orientation (for example, using the mouse with the ALT key simultaneously held down to interactively adjust the angle of rotation of the object should that be needed). Traditionally, the cursor controlled by the mouse 717 can be moved within object 714 to relocate it within drawing area 700 b or can, as shown in FIG. 7, be positioned atop one of the graphical handles 716 to permit the mouse to adjust the horizontal and vertical scale of object 714, i.e., adjust its size and aspect-ratio. In some application packages, the latter adjustment is permitted to collapse the object through to ‘zero’ thickness in one of the adjustment dimensions and continue through to re-render the object in mirror image, thus additionally providing a form of vertical and horizontal flip by using the size and aspect-ratio resizing.
  • [0138]
    As familiar and widely accepted as these sorts of operations are, there is considerable overhead involved in such sequences of repeated selecting and adjusting (and in some cases additional deselecting) pairs of parameters from a larger collection of parameters. To see several aspects of the power of the present invention, these operations are now examined in more detail in generalized form.
  • [0139]
    FIG. 8 is a flowchart showing tasks involved in selecting and adjusting one of a plurality of available pairs of adjustable parameters by using a user interface device permitting the adjustment of only one pair of parameters at a time. In FIG. 8, the task goal is simply to adjust a pair of selected parameters 801 from a larger group of adjustable parameters. However, since a larger number of parameters are available for adjustment than are available at one time with the user interface sensor, the specific pair of parameters must first be selected. In most known graphical user interface systems and methods, this typically involves first using the user interface device to control the movement of a cursor to a selection area of the graphical interface in a first overhead step 811 and then selecting the adjustment context (parameter pair) in a second overhead step 821.
  • [0140]
    In some situations the selected pair of parameters may be immediately adjusted in goal operation 801, but typically the cursor must then at least be moved, in a third overhead operation 812, to a location outside of the selection area affiliated with operations 811 and 821 to another location (such as a drawing or typing area) affiliated with the context (parameter pair) that has just been selected for adjustment. In some situations the selected pair of parameters may be immediately adjusted in goal operation 801, but typically the context must be activated (for example, by clicking in an open portion of a drawing area or selecting an existing object) in a fourth overhead operation 822.
  • [0141]
    After the selected pair of parameters are adjusted (for example, by sizing a rectangle, etc.) the cycle may then immediately be repeated in some variant form for another pair of parameters, but typically the parameters must be deselected (for example, by another click to set the final value) in a fifth overhead operation 823 before the cursor may be moved to the selection area in another instance of operation 811. In summary, in order to adjust one pair of parameters from a larger group of parameters, as many as five overhead operations (as many as two cursor movements 811, 812 and as many as three select/deselect clicks 821, 822, 823) are commonly required.
  • [0142]
    FIGS. 9 a-9 b show broader implications of the overhead called out in FIG. 8. FIG. 9 a depicts the sequential adjustment of pairs of parameters chosen from a larger group of pairs of parameters in a scenario suggestive of no interactive iteration. One pair of parameters is adjusted with up to five overhead operations in action 901, then a second pair of parameters is adjusted with up to five overhead operations in action 902, then a third pair of parameters is adjusted with up to five overhead operations in action 903, and so on. Here the overhead slows things down but may not be a significant encumbrance to the broader goal of actions 901, 902, 903, etc.
  • [0143]
    In contrast, FIG. 9 b depicts an interactive adjustment of pairs of parameters from a larger group of parameters in a scenario suggestive of one where interactive iteration is required, as the setting of one pair of parameters is difficult to complete without setting other parameters. Here the overhead is likely a significant encumbrance to the higher goal involving the pair-wise adjustment actions 901, 902, 903, etc. For example, consider the interactive adjustment of six parameters, one pair at a time, through pair-wise adjustment actions 901,902,903; not only are up to five operations of overhead involved for each of the pair-wise adjustment actions 901,902,903, but a considerable extra number of passes must be made through these pair-wise adjustment actions 901, 902,903 due to the fact that the adjustment of some parameters depends on or interacts heavily with the values of other parameters. The situation gets even more cumbersome should additional pair-wise adjustment operations be required. FIG. 9 b further shows the potential for one or more additional adjustment actions 950 which in principle may be iterated as well as and combined with pair-wise adjustment actions 901, 902, 903 (as suggested by fully-connected iteration paths 921, 922, 923). In contrast to such sequences or iterative graphs of pair-wise adjustment actions, the present invention readily offers, for example, four, six, eight or even higher numbers of simultaneously adjustable parameters controllable by the same hand which, when selected in a context, eliminate the many overhead operations depicted in FIGS. 8 and 9 a-9 b, and the many additional iteration steps depicted in FIG. 9 b.
  • [0144]
    Returning now to the generalized graphical layout situation described earlier and depicted in FIG. 7, the following operations are routinely performed in 2D graphics, layout, and CAD applications:
      • A. Selection of palette containing object;
      • B. Selection of object from palette;
      • C. Selection of “layer” object is to be assigned to (common in CAD, but typically not used in standard draw and paint packages);
      • D. Adjustment of object placement in drawing;
      • E. Adjustment of object sizing;
      • F. Adjustment of object rotation;
      • G. Adjustment of object line thickness(es);
      • H. Adjustment of object line color(s);
      • I. Adjustment of object fill color(s); and
      • J. Adjustment of object fill pattern(s);
  • [0155]
    Of these, operations B, D, and E are almost always utilized, operations A, G, and I are frequently utilized, and operations C, F, H, and J are rarely utilized.
  • [0156]
    Thus, in one exemplary application of the invention, it may be advantageous to group specific collections of operations that are commonly used together (this may be application specific) so that the benefits of having four or more widely-adjustable interactive parameters simultaneously available can be applied to speed the execution of basic common operations. For example:
      • Employing a four-parameter version of the invention:
        • Operation 1: Mouse for operation B
        • Operation 2: Mouse for operation D, trackball or touchpad for operation E
      • Employing a six-parameter version of the invention comprising a 4-parameter touchpad:
        • Operation 1: Mouse for operation B, touchpad finger-location for operation D, touchpad finger-tilt for operation E.
      • Employing a six-parameter version of the invention comprising a mouse and two trackballs or touchpads:
        • Operation 1: Mouse for operation B, first trackball/touchpad for operation D, second trackball/touchpad for operation E.
  • [0164]
    Other operations can be later applied in groupings and operations appropriate for the application.
  • [0165]
    As a possible alternative to the preceding example, it may be advantageous to assign a principal one of the user interface sensors to the sequential adjustment of each of such universal (or otherwise principal) operations and reserve the additional user interface sensors for rapid “in-context” interactive access to less frequently used operations. For example:
      • Employing a four-parameter version of the invention:
        • Operation 1: Mouse for operation B, trackball or touchpad for operation A and/or operation C;
        • Operation 2: Mouse for operation D, trackball or touchpad for operation E and/or operation F;
        • Operation 3: Mouse for operation G, trackball or touchpad for operation H; and
        • Operation 4: Mouse for operation I, trackball or touchpad for operation J.
      • Employing a six-parameter version of the invention comprising a 4-parameter touchpad:
        • Operation 1: Mouse for operation B, touchpad finger-location for operation A and touchpad finger-tilt for operation C;
        • Operation 2: Mouse for operation D, touchpad finger-location for operation E and touchpad finger-tilt for operation F;
        • Operation 3: Mouse for operation G, touchpad finger-location (and touchpad finger-tilt as useful) for operation H; and
        • Operation 4: Mouse for operation I, touchpad finger-location (and touchpad finger-tilt as useful) for operation J.
      • Employing a six-parameter version of the invention comprising a mouse and two trackballs or touchpads:
        • Operation 1: Mouse for operation B, first trackball/touchpad for operation A and second trackball/touchpad for operation C;
      • Operation 2: Mouse for operation D, first trackball/touchpad for operation E and second trackball/touchpad for operation F;
      • Operation 3: Mouse for operation G, first trackball;/touchpad (and second trackball/touchpad as useful) for operation H; and
      • Operation 4: Mouse for operation I, first trackball/touchpad (and second trackball/touchpad as useful) for operation J.
  • [0181]
    As another alternative example, the user may freely assign user interface sensor parameters to operations A through J (and others as may be useful) for each of a number of steps as may match the task or tasks at hand. These assignments may be stored for later retrieval and use, and may be named by the user. The stored assignments may be saved along with specific files, specific applications, or as a general template the user may apply to a number of applications. It is noted that such variable assignments may be particularly useful to users as their hands fatigue, to prevent fatigue or injury, or as an adjustment for a temporary or permanent disability.
  • [0182]
    3.4 Multi-Resolution Mouse Application
  • [0183]
    In another exemplary family of applications, one user interface sensor (for example, the mouse body) is used for course adjustment or fine adjustment of user interface parameters, while the additional user interface sensor (for example, a trackball or touchpad) is used for the remaining level of parameter adjustment resolution.
  • [0184]
    In most user-interface applications it is advantageous to have multiple scales of graphical user interface pointing and data entry. Many window systems provide an ‘acceleration’ setting which changes the pointing and data entry values on a more significant scale used for user interface changes made less frequently. Many applications further internally adjust the resolution as the corresponding visual display is “zoomed” in and out.
  • [0185]
    In many user interface applications, additional levels of resolution selection may be useful. For example, in pointing usage in text work, multiple resolutions would be advantageous in amending fine print or in making isolated changes in thumbnail overviews of 40% actual size or less. Similarly, in graphics work, fine resolution may be especially useful in making fine adjustments to figures. In the fine adjustment of figures, it may be further advantageous to employ each of the separate user interface sensors in conjunction with corresponding snap-grids of differing grid spacing, particularly if one of the grid spacings is a sub-multiple of the other. A potentially useful extension of this would be to impose locally-applicable grid spacing on individual graphic or other objects, particularly objects which have been resized and hence for which the standard snap-grid spacing is no longer useful.
  • [0186]
    In a further application, the user interface may be directed towards non-positional adjustments, such as the adjustment of a rotation angle or of the color of a graphic object; here multiple resolutions may be valuable to make careful adjustments and coarse adjustments as needed. Similarly, scroll bars for long documents may also benefit from rapid access to multiple resolution scales, for example one user interface sensor may be used to navigate within a page (using a fine-grained navigation scale) while a second user interface sensor may be used to navigate across pages (using a coarser-grained navigation scale).
  • [0187]
    3.5 Provision of Both Absolute and Relative Positioning
  • [0188]
    As discussed earlier, some types of user interface sensors, such as the touchpad and X-Y joystick for example, naturally have a limited maximum range of operation while others such as a mouse, trackball, and scroll-wheel have an essentially unlimited maximum range of operation. Although most user interface sensors are interpreted in relative terms (that is, the stimulus from the sensor is interpreted as a command to move a cursor, scroll bar, etc., incrementally in some direction relative to a current position), stimulus signals from any of these types of user interface signals may be interpreted in either a relative or absolute manner with varying degrees of naturalness or problematic qualities.
  • [0189]
    The present invention provides for one user interface sensor to be used for absolute positioning of a cursor, scroll bar, etc., or other means of parameter adjustment while another user interface sensor is used for traditional relative adjustment of such parameters. For example, a scroll bar may be adjusted in the usual fashion by a mouse body or trackball and in an absolute manner by a touchpad wherein the extreme values of the adjusted parameter correspond to the extreme positions at the edges of the touchpad. In one embodiment or application setting these two user interface sensors may control the same parameters—here it is often the result that the two sensors adjust the same parameters with different resolutions. Further, in this situation it is fairly likely that at least one of the resolution scale factors will be adjusted automatically. For example, in a document editor, as the number of pages of the document varies, the resolution of the absolute positioning sensor will correspondingly vary (so that the extremities in range of, for example, a touchpad correspond to the top of the first page and end of the last page) while the relative positioning sensor may retain the same incrementing/decrementing vertical scrolling resolution scale regardless of the number of pages.
  • [0190]
    3.6 Color-Selection Application
  • [0191]
    In color adjustment, three parameters are involved in the full interactive span of any complete color space (RGB, HSB, YUV, etc.). By adding additional parameters to the overall user interface, all three parameters can be adjusted simultaneously rather than simply two at a time. As the present invention provides at least four simultaneous interactively adjustable parameters overall, it is thus potentially useful for fully interactive color adjustment within a complete color space model. Further, should the additional user interface sensor be such that it alone provides three simultaneously interactively adjustable parameters, the first user interface sensor (for example, the mouse body) may be used as a pointer to select objects and the additional user interface sensor may be used to adjust attributes of the selected object such as its color, border color, etc.
  • [0192]
    3.7 Multi-level Graphic Object Grouping and Editing Application
  • [0193]
    In many drawing applications, lower-level graphical or other objects (such as lines, basic shapes, and text areas) may be grouped to form an aggregated object. This aggregated or “grouped” object (collectively referred to herein as an “aggregated object”) can then be moved, rotated, flipped, resized, etc. as if it were a lower-level graphic or other object. Grouping can also typically be done hierarchically and in mixed hierarchies, i.e., a plurality of lower-level graphical or other objects may first be grouped, and the resulting aggregated object may then itself be grouped with other aggregated objects and/or lower-level graphical or other objects.
  • [0194]
    Often one or more of the lower-level graphical or other objects comprising the aggregated object may need modification. In the case of text, most applications permit modifications to be made to individual text objects within an aggregated object. However, for any isolated adjustment to any other lower-level graphical or other object the aggregated object must be first disaggregated or “ungrouped” to completely free the involved lower-level graphical or other object from any grouping it had been involved in. After the modification, the grouping must be reconstructed. Often this becomes a cumbersome situation, particularly where the adjustments within the group are themselves an interactive response to other adjustments made within a drawing.
  • [0195]
    The additional number of widely-adjustable simultaneously interactive parameters made possible by the invention may be advantageously applied to this problem. For example, one user interface sensor may be used to navigate the levels of grouping and another user interface sensor may be used to perform operations on objects (lower level-or “grouped”) within that level of grouping of the overall aggregated object.
  • [0196]
    As an illustrative example, FIG. 11 a shows a portion 1100 of a larger drawing, the portion 1100 featuring box 1101, two mowed lines 1111, 1112, and grouped object 1102 (here itself comprising two triangles connected by a line). In this example it is given that grouped object 1102 is itself grouped with box 1101 to form a second grouped object, and this second grouped object is itself grouped with the two arrowed lines 1111, 1112 to form a third grouped object. The user's task is to modify FIG. 11 a so that it becomes FIG. 11 b. To do this, effectively the user must, in some order of operation:
      • Reposition grouped object 1102 from its original position in FIG. 11 a to a new position in FIG. 11 b;
      • Copy or otherwise reproduce grouped object 1102 to create an accompanying grouped object 1102 a, and position it within box 1101;
      • Introduce a vertically distributed ellipsis 1103 and position it within box 1101—typically, a vertically distributed ellipsis 1103 is either rotated text or itself a fourth grouped object created from three aligned text elements; and
      • Ensure elements 1102, 1102 a, and 1103 are in the end grouped with box 1101 to form the second grouped object, and this second grouped object is itself grouped with the two arrowed lines 1111, 1112 to form a third grouped object.
  • [0201]
    Utilizing the invention, one user interface sensor is used to select the second group level, and the second user interface sensor is used to perform insert, copy, paste, and position operations within this level of grouping without any form or type of ungrouping operation. If the vertically distributed ellipsis 1103 itself is realized as a fourth grouped object created from three aligned text elements, when it is pasted into the drawing via this modality its 1103 grouping is subordinated appropriately (i.e., structured as a peer to grouped objects 1102, 1102 a) within the second grouping level.
  • [0202]
    Although readily implemented using the novel user interface sensors described herein that make it particularly easy to simultaneously adjust a plurality of pairs of parameters, the aspects of the invention illustrated here can also be implemented with a conventional user interface sensor such as a traditional mouse, trackball, or touchpad. In this case, the conventional user interface sensor such as a traditional mouse, trackball, or touchpad must first be used to select the level of grouping and then be used to make the desired modifications within that level of grouping; to make modifications at a different level of grouping, the new level of grouping must be selected in a separate operation, thus adding overhead, as depicted in FIGS. 8 and 9 a-9 b. Although this novel ability to move and modify arbitrary graphic or other objects within groupings may be implemented in this way, having an additional number of widely-adjustable simultaneously interactive parameters—made possible by the main themes of the present invention—is clearly more efficient, as many or all of the overhead operations depicted in FIGS. 8 and 9 a-9 b can be eliminated via usage of the additional widely-adjustable simultaneously interactive parameters.
  • [0203]
    3.8 3D Graphic Object Placement and Orientation Application
  • [0204]
    CAD and drawing packages that enable 3D object placement and orientation within a 3D space typically extend the capabilities of traditional 2D layout, CAD, and graphics applications as described in Section 3.3 to serve additional geometric needs. As illustrated in FIG. 12, the placement and orientation of 3D object 1200 within a 3D space 1250 (oriented with respect to a reference point 1251) requires that one additional position dimension and two additional orientation angles be specified to complete the full collection of three position dimensions 1201, 1202, 1203 and three orientation angles 1211, 1212, 1213.
  • [0205]
    To interactively adjust these parameters pairwise (or individually with a knob box as has been done historically in some systems) involves complex repetitive passes among high overhead operations as depicted in FIGS. 9 a-9 b, for example among steps 901, 902, 903. The necessity of making many high-overhead operations, for example moving among steps 901, 902, 903, can be functionally disruptive as well as slow and inefficient. The ability to interactively freely adjust the full collection of three position dimensions 1201, 1202, 1203 and three orientation angles 1211, 1212, 1213 is thus of extremely high value.
  • [0206]
    The invention provides for a wide range of mappings between the six position and orientation parameters 1201, 1202, 1203, 1211, 1212, 1213 involved in the placement and orientation of 3D object 1200 within a 3D space, and the large numbers of widely-adjustable and simultaneously interactive parameters facilitated by various realizations of the invention. As one example, a mouse fitted with two trackballs as in FIG. 14 a may be used to specify these six parameters in various ways. One technique is to use the position of mouse body 1400 to control two of the position coordinates (for example 1202, 1203), one of the trackballs (for example 1405 a) to control the orientation angles (1212, 1213) corresponding to these two axes, and the remaining trackball (1405 b) to control the remaining axis (1201) and its corresponding orientation angle (1211). In this example, trackballs 1405 a, 1405 b are configured or used in 2-parameter modalities.
  • [0207]
    In another example, the mouse of FIG. 14 a is fitted with two trackballs, the trackballs may be configured in 3-parameter modalities with one of the trackballs used for controlling three position dimensions 1201, 1202, 1203 and the second trackball configured to correspondingly control the three orientation angles 1211, 1212, 1213. Here the position of mouse body 1400 may be used to control other aspects of drawing operations.
  • [0208]
    In another implementation, a touchpad configured for 4-parameter operation involving two parameters of finger position and two parameters of finger tilt may be combined with a trackball configured for 2-parameter operation. In this example, finger position may be used to control two of the position coordinates (for example 1202, 1203), finger tilt may be used to control the orientation angles (1212, 1213) corresponding to these two axes, and the trackball to control the remaining axis (1201) and its corresponding orientation angle (1211). If the configuration includes a mouse body, its position may be used to control other aspects of drawing operations.
  • [0209]
    In another example, a configuration like that of FIG. 14 d may use left-fingers/thumb trackball 1465 a to control a first position coordinate 1201 and its corresponding orientation angle 1211, the right-fingers/thumb trackball 1465 b to control a second position coordinate 1202 and its corresponding orientation angle 1212, and palm trackball 1465 c to control third position coordinate 1203 and its corresponding orientation angle 1213.
  • [0210]
    The invention further provides for a wide range of additional mappings and geometric metaphors between the user interface sensor geometry and the three position dimensions 1201, 1202, 1203 and three orientation angles 1211, 1212, 1213 of a 3D object.
  • [0211]
    3.9 Multiple Cursors and Cut and Paste Application
  • [0212]
    The invention additionally provides for a plurality of pairs of user interface sensor parameters to be used to control the respective positions of a corresponding plurality of individual cursors, selections, and/or insertion points. Multiple cursors and associated operations of multiple selection and insertion points can have many applications. Below, a few of these possibilities that would be apparent to one skilled in the art are showcased in a cut-and-paste editing example.
  • [0213]
    Cut, copy, and paste operations using traditional user interface devices usually involve multiple operations to switch between contexts introducing considerable overhead as depicted in FIGS. 9 a-9 b and 8. For instance, FIG. 13 a illustrates a text editing example with text display window 1300 involving the selection of a clause 1320 (highlighted in this example) with the intention of relocating it to a new position 1351. In such an operation with a traditional 2-parameter mouse/trackball/touchpad user interface, the cursor is first used to select clause 1320 and then used to select the insertion position 1351.
  • [0214]
    When writing or editing it is often the case that material needs to be fetched from elsewhere and put in the spot where one is currently writing. Here, the cursor is initially in the spot where the insertion is to occur and the user must then lose the cursor position currently set in this spot to go searching and then to select and cut or copy the material to be pasted; following this the user must then search again, perhaps taking considerable time, for said initial spot and re-establish the cursor location there. Equally often there are other situations where material must be split up and distributed in a number of far-flung places. Here, the cursor is initially in the spot where the material to be sequentially divided and relocated is originally aggregated; the user must repeatedly select the portion of the remaining aggregated material to be relocated and then lose that cursor position to go searching for the new destination insertion spot, perform the insertion, and then search again, perhaps taking considerable time, for the initial spot and re-establish the cursor location there. In both of these cases it would be advantageous if the user could “bookmark” an initial cursor location, search and perform the desired fetch or relocation operations, and readily return without search to the “bookmarked” location. Although this novel and advantageously valuable capability could be realized with a conventional mouse through context redirection operations involving the steps depicted in FIGS. 8 and 9 a-9 b, the present invention provides for a wide range of readily realized and easy-to-use implementations.
  • [0215]
    The invention may be used in a minimal configuration capable of interactively specifying at least two pairs of widely adjustable interactive parameters. Returning to the specific example associated with FIG. 13 a, one pair of parameters is used to set the location of first cursor 1301 which is used in a selection operation, while the second pair of parameters is used to set the location of second cursor 1351 which is to be used to independently set an insertion point. The user may then perform the cut and paste operation with a single mouse click, resulting in the outcome depicted in FIG. 13 b. The relocated text clause 1320 has been transferred to a position determined by the insertion cursor 1351 (here shown to the left of the cursor 1351; it could just as easily be to the right of it), and cursor 1301 used to make the selection remains in position. Either cursor 1301 or 1351 may now be moved and/or used for other cut, copy, paste, or (via the keyboard) new text insertion operations.
  • [0216]
    Although in this example the two cursor locations were close enough to be displayed in the same window 1300, the value of this application of the invention is significantly increased should the two positions be separated by many pages, many tens of pages, or even many hundreds of pages of text. Such situations may be handled by any number of approaches as is clear to one skilled in the art. In one approach involving a single display window, the area comprising the cursor whose corresponding user interface sensor was last manipulated is displayed in the single display window. In another approach involving a single display window, a click event or other user interface stimulus may be used to toggle among the areas comprising the various cursor locations. In yet another approach, at least two windows may be rendered, with one of the cursors displayed and operable within one window and a second cursor displayed and operable in a second window.
  • [0217]
    The invention also provides for these general principles to be applied to other types of objects and applications, such as spreadsheet cells (involving data, formula objects, and cell formats), graphical objects, electronic CAD diagrams (where objects may be connected with formulas, dynamic models, etc.), and others as will be apparent to one skilled in the art.
  • [0218]
    3.10 Simulation, Processing, and Analysis Applications
  • [0219]
    Simulation, processing, and analysis applications typically involve a large number of parameters which are adjusted to model, affect or investigate the resulting behaviors, end results, and/or implications. Conventional 2-parameter user interface devices such as a mouse/trackball/touchpad require the user to adjust these parameters pairwise (or individually with a knob box as has been done historically in some systems) involving complex repetitive passes among high-overhead operations as depicted in FIGS. 9 a-9 b, for example among steps 901, 902, 903. As in the case of 3D object positioning and orientation, the division among high-overhead operations, for example moving among steps 901, 902,903, can be functionally disruptive as well as slow and inefficient. The ability to interactively and freely adjust larger collections of parameters simultaneously is thus also of extremely high value.
  • [0220]
    3.11 Live Signal Processing and Lighting Applications
  • [0221]
    In artistic performance, composition, and recording applications, control of large numbers of parameters requiring simultaneous interactive adjustment is common. Conventional recording, mixing, video, and light control consoles typically have large numbers of controls with carefully designed spatial layouts to facilitate the rapid and precise adjustment of multiple parameters via knobs, sliders, pushbuttons, toggle switches, etc. The introduction of computer GUIs has added considerable value and new capabilities, including “soft” reconfigurable consoles and functional assignments, but in the bargain, typically encumber users—accustomed to rapid and precise operation of multiple parameters—with a 2-parameter mouse/trackball/touchpad having the overhead of iterative context-switching operations depicted in FIGS. 8 and 9 a-9 b. As in the case of 3D object positioning and orientation, the division among high-overhead operations, for example moving among steps 901, 902, 903, can be functionally disruptive as well as slow and inefficient. The ability to interactively freely adjust larger collections of parameters simultaneously is thus also of extremely high value.
  • [0222]
    3.12 Real-Time Machine Control and Plant Operations
  • [0223]
    Similarly, real-time machine control and plant (manufacturing, chemical, energy, etc.) operations also traditionally involve controlling a significant number of parameters requiring simultaneous interactive adjustment. Conventional real-time machine control and plant operation consoles typically have large numbers of controls with carefully designed spatial layouts to facilitate the rapid and precise adjustment of multiple parameters via knobs, sliders, pushbuttons, toggle switches, etc. The introduction of computer GUIs can add considerable value and new capabilities, including “soft” reconfigurable consoles and functional assignments, but in the bargain typically significantly encumber users—accustomed to rapid and precise operation of multiple parameters—with a 2-parameter mouse/trackball/touchpad having the overhead of iterative context-switching operations depicted in FIGS. 8 and 9 a-9 b. As in the case of 3D object positioning/orientation and artistic applications described earlier, the division among high-overhead operations, for example moving among steps 901, 902, 903, can be functionally disruptive as well as slow and inefficient. The ability to interactively freely adjust larger collections of parameters simultaneously is thus also of extremely high value.
  • [0224]
    A very few examples of this category of application where the invention may be useful include many forms of robotics control, computer-control manufacturing tools, industrial optical and electron microscopy, camera control (pan, tilt, zoom, focus, and/or iris), plant process elements (heaters, pumps, values, stirrers, aerators, actuators, activators, etc.), and a wide range of other related and divergent possibilities apparent to those skilled in the art.
  • 4. Concluding Remarks
  • [0225]
    The present invention at its core provides for a wide range of systems and methods for realizing and applying user interfaces providing, in many cases, at least four widely variable simultaneously interactively adjustable parameters. In so doing, the invention more broadly encompasses novel user interface structures, metaphors, and applications readily suggested and enabled by the core of the invention but which may be indeed realized in ways not involving the core of the invention.
  • [0226]
    While the invention has been described in detail with reference to disclosed embodiments, various modifications within the scope of the invention will be apparent to those of ordinary skill in this technological field. It is to be appreciated that features described with respect to one embodiment typically may be applied to other embodiments. Therefore, the invention properly is to be construed with reference to the claims.
Patent Citations
Cited PatentFiling datePublication dateApplicantTitle
US3974493 *29 Apr 197410 Aug 1976Vydec, Inc.Cursor find system for the display of a word processing system
US4988981 *28 Feb 198929 Jan 1991Vpl Research, Inc.Computer data entry and manipulation apparatus and method
US5049863 *9 Feb 199017 Sep 1991Kabushiki Kaisha ToshibaCursor key unit for a computer having a mouse function removably mounted on a keyboard section of a base
US5095303 *27 Mar 199010 Mar 1992Apple Computer, Inc.Six degree of freedom graphic object controller
US5327161 *21 Oct 19915 Jul 1994Microtouch Systems, Inc.System and method for emulating a mouse input device with a touchpad input device
US5490039 *24 Nov 19936 Feb 1996Dell Usa, L.P.Integrated mouse tray and mouse for portable computer having cavity therefore
US5565891 *23 Feb 199515 Oct 1996Armstrong; Brad A.Six degrees of freedom graphics controller
US5586243 *15 Apr 199417 Dec 1996International Business Machines CorporationMultiple display pointers for computer graphical user interfaces
US5615083 *11 Dec 199525 Mar 1997Gateway 2000, Inc.Detachable joystick for a portable computer
US5666499 *4 Aug 19959 Sep 1997Silicon Graphics, Inc.Clickaround tool-based graphical interface with two cursors
US5670990 *21 Jan 199423 Sep 1997Logitech IncorporatedIntegral ball cage for pointing device
US5726684 *26 Jul 199610 Mar 1998Ncr CorporationDetachable convertible mouse-trackball pointing device for use with a computer
US5805144 *18 Jul 19968 Sep 1998Dell Usa, L.P.Mouse pointing device having integrated touchpad
US5852442 *2 Jul 199622 Dec 1998Fujitsu LimitedMethod of drawing a three-dimensional object
US5881366 *1 May 19969 Mar 1999Logitech, Inc.Wireless peripheral interface
US5917472 *15 May 199729 Jun 1999International Computers LimitedCursor control system with multiple pointing devices
US5999169 *30 Aug 19967 Dec 1999International Business Machines CorporationComputer graphical user interface method and system for supporting multiple two-dimensional movement inputs
US6057830 *17 Jan 19972 May 2000Tritech Microelectronics International Ltd.Touchpad mouse controller
US6061051 *17 Jan 19979 May 2000Tritech MicroelectronicsCommand set for touchpad pen-input mouse
US6128006 *26 Mar 19983 Oct 2000Immersion CorporationForce feedback mouse wheel and other control wheels
US6204837 *13 Jul 199820 Mar 2001Hewlett-Packard CompanyComputing apparatus having multiple pointing devices
US6205021 *26 Feb 199820 Mar 2001Micron Electronics, Inc.Method for operating an input device and a laptop computer
US6219037 *2 Oct 199817 Apr 2001Samsung Electronics Co., Ltd.Pointing device provided with two types of input means for a computer system
US6232958 *29 Dec 199815 May 2001Microsoft CorporationInput device with multiplexed switches
US6239790 *17 Aug 199929 May 2001Interlink ElectronicsForce sensing semiconductive touchpad
US6239803 *14 Apr 199929 May 2001Stanley W. DriskellMethod to achieve least effort selection from an item list of arbitrary length
US6281881 *29 Oct 199828 Aug 2001Microsoft CorporationSystem and method of adjusting display characteristics of a displayable data file using an ergonomic computer input device
US6295051 *2 Jun 199925 Sep 2001International Business Machines CorporationIntelligent boundless computer mouse system
US6323846 *25 Jan 199927 Nov 2001University Of DelawareMethod and apparatus for integrating manual input
US6329978 *1 Jun 199811 Dec 2001Fu-Kuo YehCursor control device
US6396477 *14 Sep 199828 May 2002Microsoft Corp.Method of interacting with a computer using a proximity sensor in a computer input device
US6424335 *2 Sep 199823 Jul 2002Fujitsu LimitedNotebook computer with detachable infrared multi-mode input device
US6456275 *14 Sep 199824 Sep 2002Microsoft CorporationProximity sensor in a computer input device
US6489948 *20 Apr 20003 Dec 2002Benny Chi Wah LauComputer mouse having multiple cursor positioning inputs and method of operation
US6525713 *9 May 200025 Feb 2003Alps Electric Co., Ltd.Coordinate input device capable of inputting z-coordinate of image object
US6556150 *24 Mar 200029 Apr 2003Microsoft CorporationErgonomic computer input device
US6570078 *19 Mar 200127 May 2003Lester Frank LudwigTactile, visual, and array controllers for real-time control of music signal processing, mixing, video, and lighting
US6570557 *10 Feb 200127 May 2003Finger Works, Inc.Multi-touch system and method for emulating modifier keys via fingertip chords
US6580420 *15 Mar 200017 Jun 2003Yanqing WangConvertible computer input device
US6587091 *23 Apr 20011 Jul 2003Michael Lawrence SerpaStabilized tactile output mechanism for computer interface devices
US6590564 *24 Mar 20008 Jul 2003Microsoft CorporationErgonomic computer input device
US6623194 *24 Oct 200023 Sep 2003Chung Ching LipPosition encoder system
US6646632 *1 Dec 200011 Nov 2003Logitech Europe S.A.Tactile force feedback device
US6677932 *28 Jan 200113 Jan 2004Finger Works, Inc.System and method for recognizing touch typing under limited tactile feedback conditions
US6704003 *15 Dec 20009 Mar 2004Logitech Europe S.A.Adaptable input device support
US6714221 *3 Aug 200030 Mar 2004Apple Computer, Inc.Depicting and setting scroll amount
US6888536 *31 Jul 20013 May 2005The University Of DelawareMethod and apparatus for integrating manual input
US6909422 *6 Feb 200121 Jun 2005Matsushita Electric Industrial Co., Ltd.Track ball device and electronic apparatus using the same
US7168047 *28 May 200223 Jan 2007Apple Computer, Inc.Mouse having a button-less panning and scrolling switch
US7256770 *13 Mar 200114 Aug 2007Microsoft CorporationMethod for displaying information responsive to sensing a physical presence proximate to a computer input device
US7358956 *13 Mar 200115 Apr 2008Microsoft CorporationMethod for providing feedback responsive to sensing a physical presence proximate to a control of an electronic device
US7463239 *21 Jan 20049 Dec 2008Microsoft CorporationInput device including a wheel assembly for scrolling an image in multiple directions
US7557797 *24 Nov 20047 Jul 2009Ludwig Lester FMouse-based user interface device providing multiple parameters and modalities
US7620915 *13 Feb 200417 Nov 2009Ludwig Lester FElectronic document editing employing multiple cursors
US7808479 *2 Sep 20035 Oct 2010Apple Inc.Ambidextrous mouse
US20010033268 *20 Feb 200125 Oct 2001Jiang Jiong JohnHandheld ergonomic mouse
US20020005108 *19 Mar 200117 Jan 2002Ludwig Lester FrankTactile, visual, and array controllers for real-time control of music signal processing, mixing, video, and lighting
US20020080112 *27 Sep 200127 Jun 2002Braun Adam C.Directional tactile feedback for haptic feedback interface devices
US20020084981 *2 May 20014 Jul 2002Flack James F.Cursor navigation system and method for a display
US20020113776 *21 Feb 200122 Aug 2002Ran MeriazTrackball mouse
US20020190930 *13 Nov 200119 Dec 2002Fujitsu Hitachi Plasma Display LimitedMethod of driving plasma display panel
US20030006961 *26 Jun 20029 Jan 2003Yuly ShipilevskyMethod and system for increasing computer operator's productivity
US20030063062 *20 Mar 20023 Apr 2003Makoto TsumuraImage display device
US20030107552 *11 Dec 200112 Jun 2003Koninklijke Philips Electronics N.V.Computer mouse with dual functionality
US20030107603 *12 Dec 200112 Jun 2003Intel CorporationScroll notification system and method
US20030169216 *5 Mar 200311 Sep 2003Lg Electronics Inc.Method and apparatus for driving plasma display panel
US20040017358 *16 Jul 200329 Jan 2004Kuo Hsiao MingHousing assembly for computer mouse
US20040126171 *16 Dec 20031 Jul 2004Microsoft CorporationKeyboard with improved lateral region
US20040147318 *20 Jan 200429 Jul 2004Shahoian Erik J.Increasing force transmissibility for tactile feedback interface devices
US20040155865 *16 Dec 200312 Aug 2004Swiader Michael CErgonomic data input and cursor control device
US20040189605 *3 Nov 200330 Sep 2004Ken ShihMouse with multi-axis inputting device
US20050179650 *24 Nov 200418 Aug 2005Ludwig Lester F.Extended parameter-set mouse-based user interface device offering offset, warping, and mixed-reference features
US20050179652 *10 Dec 200418 Aug 2005Ludwig Lester F.Mouse-based user interface device employing user-removable modules
US20050179655 *13 Feb 200418 Aug 2005Ludwig Lester F.Electronic document editing employing multiple cursors
US20050179663 *22 Mar 200418 Aug 2005Ludwig Lester F.Freely rotating trackball providing additional control parameter modalities
US20050193321 *3 May 20051 Sep 2005Microsoft CorporationInsertion point bungee space tool
US20050275637 *25 Aug 200515 Dec 2005Microsoft CorporationMethod for displaying information responsive to sensing a physical presence proximate to a computer input device
US20060125803 *10 Feb 200615 Jun 2006Wayne WestermanSystem and method for packing multitouch gestures onto a hand
US20080088602 *28 Dec 200717 Apr 2008Apple Inc.Multi-functional hand-held device
US20080128182 *30 Jul 20075 Jun 2008Apple Inc.Sensor arrangement for use with a touch sensor
US20080158169 *3 Jan 20073 Jul 2008Apple Computer, Inc.Noise detection in multi-touch sensors
US20080163130 *15 Jun 20073 Jul 2008Apple IncGesture learning
US20080165140 *13 Jun 200710 Jul 2008Apple Inc.Detecting gestures on multi-event sensitive devices
US20080165141 *13 Jun 200710 Jul 2008Apple Inc.Gestures for controlling, manipulating, and editing of media files using touch sensitive devices
US20080168364 *5 Jan 200710 Jul 2008Apple Computer, Inc.Adaptive acceleration of mouse cursor
US20080204426 *15 Apr 200828 Aug 2008Apple Inc.Gestures for touch sensitive input devices
US20080211775 *9 May 20084 Sep 2008Apple Inc.Gestures for touch sensitive input devices
US20080211783 *9 May 20084 Sep 2008Apple Inc.Gestures for touch sensitive input devices
US20080211784 *9 May 20084 Sep 2008Apple Inc.Gestures for touch sensitive input devices
US20080211785 *9 May 20084 Sep 2008Apple Inc.Gestures for touch sensitive input devices
US20080231610 *9 May 200825 Sep 2008Apple Inc.Gestures for touch sensitive input devices
US20080309632 *13 Jun 200718 Dec 2008Apple Inc.Pinch-throw and translation gestures
US20100156818 *6 Apr 200924 Jun 2010Apple Inc.Multi touch with multi haptics
US20100214250 *11 May 201026 Aug 2010Synaptics IncorporatedTouch screen with user interface enhancement
Referenced by
Citing PatentFiling datePublication dateApplicantTitle
US843973316 Jun 200814 May 2013Harmonix Music Systems, Inc.Systems and methods for reinstating a player within a rhythm-action game
US844446430 Sep 201121 May 2013Harmonix Music Systems, Inc.Prompting a player of a dance game
US844448620 Oct 200921 May 2013Harmonix Music Systems, Inc.Systems and methods for indicating input actions in a rhythm-action game
US846536629 May 200918 Jun 2013Harmonix Music Systems, Inc.Biasing a musical performance input to a part
US8542110 *3 Aug 200924 Sep 2013Lg Electronics Inc.Mobile terminal and object displaying method using the same
US855090816 Mar 20118 Oct 2013Harmonix Music Systems, Inc.Simulating musical instruments
US856240310 Jun 201122 Oct 2013Harmonix Music Systems, Inc.Prompting a player of a dance game
US856823416 Mar 201129 Oct 2013Harmonix Music Systems, Inc.Simulating musical instruments
US863657216 Mar 201128 Jan 2014Harmonix Music Systems, Inc.Simulating musical instruments
US86630138 Jul 20094 Mar 2014Harmonix Music Systems, Inc.Systems and methods for simulating a rock band experience
US867889516 Jun 200825 Mar 2014Harmonix Music Systems, Inc.Systems and methods for online band matching in a rhythm action game
US867889614 Sep 200925 Mar 2014Harmonix Music Systems, Inc.Systems and methods for asynchronous band interaction in a rhythm action game
US868626931 Oct 20081 Apr 2014Harmonix Music Systems, Inc.Providing realistic interaction to a player of a music-based video game
US869067016 Jun 20088 Apr 2014Harmonix Music Systems, Inc.Systems and methods for simulating a rock band experience
US87024855 Nov 201022 Apr 2014Harmonix Music Systems, Inc.Dance game and tutorial
US887424316 Mar 201128 Oct 2014Harmonix Music Systems, Inc.Simulating musical instruments
US8918739 *24 Aug 200923 Dec 2014Kryon Systems Ltd.Display-independent recognition of graphical user interface control
US90241669 Sep 20105 May 2015Harmonix Music Systems, Inc.Preventing subtractive track separation
US909831324 Aug 20094 Aug 2015Kryon Systems Ltd.Recording display-independent computerized guidance
US9152315 *21 Nov 20116 Oct 2015Samsung Electronics Co., LtdMethod and apparatus for operating an electronic book function in a mobile device
US927828627 Oct 20148 Mar 2016Harmonix Music Systems, Inc.Simulating musical instruments
US935845614 Mar 20137 Jun 2016Harmonix Music Systems, Inc.Dance competition game
US940555824 Aug 20092 Aug 2016Kryon Systems Ltd.Display-independent computerized guidance
US95692319 Feb 200914 Feb 2017Kryon Systems Ltd.Device, system, and method for providing interactive guidance with execution of operations
US970346230 Nov 201411 Jul 2017Kryon Systems Ltd.Display-independent recognition of graphical user interface control
US20090310027 *16 Jun 200817 Dec 2009James FlemingSystems and methods for separate audio and video lag calibration in a video game
US20100009750 *8 Jul 200914 Jan 2010Harmonix Music Systems, Inc.Systems and methods for simulating a rock band experience
US20100060475 *3 Aug 200911 Mar 2010Lg Electronics Inc.Mobile terminal and object displaying method using the same
US20100205529 *9 Feb 200912 Aug 2010Emma Noya ButinDevice, system, and method for creating interactive guidance with execution of operations
US20100205530 *9 Feb 200912 Aug 2010Emma Noya ButinDevice, system, and method for providing interactive guidance with execution of operations
US20110047462 *24 Aug 200924 Feb 2011Emma ButinDisplay-independent computerized guidance
US20110047488 *24 Aug 200924 Feb 2011Emma ButinDisplay-independent recognition of graphical user interface control
US20110047514 *24 Aug 200924 Feb 2011Emma ButinRecording display-independent computerized guidance
US20120127104 *21 Nov 201124 May 2012Samsung Electronics Co., Ltd.Method and apparatus for operating an electronic book function in a mobile device
US20120146923 *7 Oct 201114 Jun 2012Basir Mossab OTouch screen device
Classifications
U.S. Classification715/770, 345/173, 715/862, 715/769
International ClassificationG09G5/00, G06F3/033, G06F3/048, G09G5/08, G06F3/041
Cooperative ClassificationG06F3/04883, G06F1/266, G06F3/03549, G06F2203/04104, G06F2203/0384, G06F3/0383, G06F3/03543, G06F2203/04106, G06F3/04847, G06F3/0414, G06F3/0346, G06F3/04845, G06F3/038, G06F3/04815, G06F3/03547
European ClassificationG06F3/0354P, G06F3/0484P, G06F3/038, G06F3/0354M, G06F3/0481E, G06F3/0346, G06F3/0484M
Legal Events
DateCodeEventDescription
7 Dec 2011ASAssignment
Owner name: BIORAM TECH L.L.C., DELAWARE
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:LUDWIG, LESTER F.;REEL/FRAME:027345/0099
Effective date: 20111114
11 Dec 2015ASAssignment
Owner name: CHEMTRON RESEARCH LLC, DELAWARE
Free format text: MERGER;ASSIGNOR:BIORAM TECH L.L.C.;REEL/FRAME:037271/0905
Effective date: 20150826