CA2200716C - A speech interface for an automated endoscopic system - Google Patents
A speech interface for an automated endoscopic system Download PDFInfo
- Publication number
- CA2200716C CA2200716C CA002200716A CA2200716A CA2200716C CA 2200716 C CA2200716 C CA 2200716C CA 002200716 A CA002200716 A CA 002200716A CA 2200716 A CA2200716 A CA 2200716A CA 2200716 C CA2200716 C CA 2200716C
- Authority
- CA
- Canada
- Prior art keywords
- voice
- recited
- surgical instrument
- command
- controller
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Lifetime
Links
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00039—Operational features of endoscopes provided with input arrangements for the user
- A61B1/0004—Operational features of endoscopes provided with input arrangements for the user for electronic operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00002—Operational features of endoscopes
- A61B1/00039—Operational features of endoscopes provided with input arrangements for the user
- A61B1/00042—Operational features of endoscopes provided with input arrangements for the user for mechanical operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00147—Holding or positioning arrangements
- A61B1/00149—Holding or positioning arrangements using articulated arms
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B1/00—Instruments for performing medical examinations of the interior of cavities or tubes of the body by visual or photographical inspection, e.g. endoscopes; Illuminating arrangements therefor
- A61B1/00147—Holding or positioning arrangements
- A61B1/0016—Holding or positioning arrangements using motor drive units
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/70—Manipulators specially adapted for use in surgery
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B17/00—Surgical instruments, devices or methods, e.g. tourniquets
- A61B2017/00017—Electrical control of surgical instruments
- A61B2017/00203—Electrical control of surgical instruments with speech control or speech recognition
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
Abstract
A robotic system (10) which controls the movement of a surgical instrument (16) in response to voice commands from the user. The robotic system (10) has a computer controlled arm (14) that holds the surgical instrument (16). The user provides voice commands to the computer (20) through a microphone (40).
The computer (20) contains a phrase recognizer that matches the user's speech with words stored in the computer (20). Matched words are then processed to determine whether the user has spoken a robot command. If the user has spoken a recognized robot command the computer (20) will move the robotic arm (14) in accordance with the command.
The computer (20) contains a phrase recognizer that matches the user's speech with words stored in the computer (20). Matched words are then processed to determine whether the user has spoken a robot command. If the user has spoken a recognized robot command the computer (20) will move the robotic arm (14) in accordance with the command.
Description
WO 96/09587 PCTJL1S95l12064 ~~ QQ 716 A SPEECH INTERFACE FOR AN AUTOMATED
ENDOSCOPIC SYSTEM
BACKGROUND OF THE INVENTION
1. FIELD OF THE INVENTION
The present invention relates to a robotic system that moves a surgical instrument in response to voice commands from the user.
ENDOSCOPIC SYSTEM
BACKGROUND OF THE INVENTION
1. FIELD OF THE INVENTION
The present invention relates to a robotic system that moves a surgical instrument in response to voice commands from the user.
2. DESCRIPTION OF RELATED AR,T
To reduce the invasiveness of surgery, endoscopes are commonly utilized to view the internal organs of a patient. One end of the endoscope contains a lens which is inserted into the patient through a small incision in the skin. The lens focuses an image that is transmitted by fiber optic cable to a camera located at the opposite end of the endoscope. The camera is coupled to a monitor that displays the image of the patient.
The endoscope can be used in conjunction with another surgical instrument that is inserted into the patient. An assistant typically hold the endoscope while the surgeon manipulates the surgical instrument.
The assistant moves the endoscope in response to instructions from the surgeon. Any miscommunication between the surgeon and the assistant may result in an error in the movement of the endoscope, thereby requiring the surgeon to repeat the instruction. Additionally, _7_ holding the endoscope for a significant amount of time may cause the assistant to become fatigued.
U.S. Patent 5,515,478 discloses a robotic arm that holds and moves an endoscope. The surgeon can move the robotic arm by depressing a foot pedal.
The foot pedal is connected to a cornputer which moves the arm and the scope.
Although the system disclosed by U.S. Patent 5,51 _5,478 effectively moves the endoscope, the surgeon must continually manipulate the toot pedal, a process which may detract from the surgical procedure. It would be desirable to provide a robotic endoscopic system that can be controlled by voice c:omrnands from the user.
To reduce the invasiveness of surgery, endoscopes are commonly utilized to view the internal organs of a patient. One end of the endoscope contains a lens which is inserted into the patient through a small incision in the skin. The lens focuses an image that is transmitted by fiber optic cable to a camera located at the opposite end of the endoscope. The camera is coupled to a monitor that displays the image of the patient.
The endoscope can be used in conjunction with another surgical instrument that is inserted into the patient. An assistant typically hold the endoscope while the surgeon manipulates the surgical instrument.
The assistant moves the endoscope in response to instructions from the surgeon. Any miscommunication between the surgeon and the assistant may result in an error in the movement of the endoscope, thereby requiring the surgeon to repeat the instruction. Additionally, _7_ holding the endoscope for a significant amount of time may cause the assistant to become fatigued.
U.S. Patent 5,515,478 discloses a robotic arm that holds and moves an endoscope. The surgeon can move the robotic arm by depressing a foot pedal.
The foot pedal is connected to a cornputer which moves the arm and the scope.
Although the system disclosed by U.S. Patent 5,51 _5,478 effectively moves the endoscope, the surgeon must continually manipulate the toot pedal, a process which may detract from the surgical procedure. It would be desirable to provide a robotic endoscopic system that can be controlled by voice c:omrnands from the user.
SUMMARY OF THE INVENTION
The present invention is a robotic system which controls the movement of a surgical instrument in response to voice commands from the user. The robotic system has a computer controlled arm that holds the surgical instrument. The user provides voice commands to the computer through a microphone. The computer contains a phrase recognizer that matches the user's speech with words stored in the computer.
Matched words are then processed to determine whether the user has spoken a robot command. If the user has spoken a recognized robot command the computer will move the robotic arm in accordance r~~ith the command.
Accordingly, in one aspect, the present invention provides a method for moving a surgical instrument, comprising the steps of: a) generating a voice qualifier and a voice command to move a mechanism that holds the surgical instrument; b) determining whether the voice qualifier matches a first robot command; c) determining whether the voice command matches a second robot command if the voice qualifier matched the first robot command; and, d) moving the surgical instrument with said mechanism only if the voice qualifier matches the first robot command and the voice command matches the second robot command.
In a further aspect, the present invention provides a medical robotic system that controls a surgical instrument, comprising: an actuator; a linkage attached to said actuator and that can be coupled to the surgical instrument; an audio input device that receives a voice qualifier and a voice command from a user; and, a controller that is coupled to said audio input device and provides a movement output signal to said -3a-actuator to move said linkage in accordance with the voice command only when the voice qualifier accompanies the voice command.
In a still further aspect, the present invention provides a robotic system that controls a surgical instrument, comprising: a mechanism that can move the surgical instrument; audio input means fbr receiving a voice qualifier and a voice command from a user to move the surgical instrument; and, controller means for moving said mechanism in accordance with the voice command only when the voice qualifier accompanies the voice command.
The present invention is a robotic system which controls the movement of a surgical instrument in response to voice commands from the user. The robotic system has a computer controlled arm that holds the surgical instrument. The user provides voice commands to the computer through a microphone. The computer contains a phrase recognizer that matches the user's speech with words stored in the computer.
Matched words are then processed to determine whether the user has spoken a robot command. If the user has spoken a recognized robot command the computer will move the robotic arm in accordance r~~ith the command.
Accordingly, in one aspect, the present invention provides a method for moving a surgical instrument, comprising the steps of: a) generating a voice qualifier and a voice command to move a mechanism that holds the surgical instrument; b) determining whether the voice qualifier matches a first robot command; c) determining whether the voice command matches a second robot command if the voice qualifier matched the first robot command; and, d) moving the surgical instrument with said mechanism only if the voice qualifier matches the first robot command and the voice command matches the second robot command.
In a further aspect, the present invention provides a medical robotic system that controls a surgical instrument, comprising: an actuator; a linkage attached to said actuator and that can be coupled to the surgical instrument; an audio input device that receives a voice qualifier and a voice command from a user; and, a controller that is coupled to said audio input device and provides a movement output signal to said -3a-actuator to move said linkage in accordance with the voice command only when the voice qualifier accompanies the voice command.
In a still further aspect, the present invention provides a robotic system that controls a surgical instrument, comprising: a mechanism that can move the surgical instrument; audio input means fbr receiving a voice qualifier and a voice command from a user to move the surgical instrument; and, controller means for moving said mechanism in accordance with the voice command only when the voice qualifier accompanies the voice command.
The objects and advantages of the present invention will become more readily apparent to those ordinarily skilled in the art after reviewing the following detailed description and accompanying drawings, wherein:
Figure 1 is a perspective view of a robotic endoscope system of the present invention;
Figure 2 is a schematic of an endoscope within two separate coordinate systems;
Figure 3 is a top view of a foot pedal;
Figure 4 is a schematic of a computer system;
Figure 5 is a schematic of a grammar process;
Figure 6 is a schematic of a robotic arm.
Figure 1 is a perspective view of a robotic endoscope system of the present invention;
Figure 2 is a schematic of an endoscope within two separate coordinate systems;
Figure 3 is a top view of a foot pedal;
Figure 4 is a schematic of a computer system;
Figure 5 is a schematic of a grammar process;
Figure 6 is a schematic of a robotic arm.
DETAILED DESCI~.IPTION OF THE INVENTION
Referring to the drawings more particularly by reference numbers, Figure 1 show:; si robotic system 10 of the present invention.
The system 10 is typically used in a sterile operating room where a surgeon performs a surgical procedure on a patient. The patient is placed on a operating tables 12. Attached to the table 12 is a robotic arm assembly 14 which can move a surgical instrument 16 relative to the table 12 and the patient. The surgical instrument 16 is typically an endoscope which is inserted into the abdomen of the patient. The endoscope 16 enters the patient through a cannula, wherein the scope 16 rotate about a cannul<:~ pivot point. The endoscope is typically connected to a monitor 18 vrhich allows the surgeon to view the organs, etc. of the patient. Althoug~~ an endoscope is described and shown, it is ' to be understood that the present invention c;m be used with other surgical instruments.
The robotic arm assembly 14 controlled by a computer 20. In the preferred embodiment, the robotic arm assembly 14 includes a linear actuator 24 fixed to the table 12. The linear actuator 24 is connected to a linkage arm assembly 2fi and adapted to move the linkage assembly 26 along the z axis of a first coordinate system.. The first coordinate system also has an x axis and a y axis.
The linkage arm asserzbly 26 includes a first linkage arm 28 attached to a rust rotary actuator 30 and an end effector 32. The first rotary actuator 30 is adapted to rotate the first linkage arm 28 and end effector 32 in a plane perpendicular to the z axis (x-y plane). The first rotary actuator 30 is connected to a second rotary actuator 34 by a ~~ ~0 l16 second linkage arm 36. The second actuator 34 is adapted to rotate the first actuator 30 in the x-y plane. The second rotary actuator 34 is connected to the output shaft of the linear actuator 24. The actuators 24, 30 and 34 rotate in response to output signals provided by the computer 20. As shown in Figure 2, the junction of the endoscope 16 and the end ef~'ector 32 define a second coordinate system which has an x' axis, a y' axis and a z' axis. The junction of the end effector 32 and endoscope 18 also define the origin of a third coordinate system which has a x" axis, a y" axis and a z" axis. The z" axis parallel with the longitudinal axis of the endoscope 16.
The arm assembly may have a pair of passive joints that allow the end effector to be rotated in the direction indicated by the arrows.
The actuators 24, 30 and 34, and joints of the arm may each have position sensors (not shown) that are connected to the computer 20.
The sensors provide positional feedback signals of each corresponding arm component.
The system has a microphone 40 that is connected to the computer 20. The system may also have a speaker 42 that is connected to the computer 20. The microphone 40 and speaker 42 may be mounted to a headset 44 that is worn by the user. Placing the microphone 40 in close proximity to the user reduces the amount of background noise provided to the computer and decreases the probability of an inadvertent input command.
As shown in Figure 3, the system may also have a foot pedal 50.
The foot pedal 22 has a housing 56 that supports a pair of outer first foot switches 58 and a second foot switch 60. One outer foot switch 58 has a first pressure transducer 62 and the other switch has a second 22 ~4 716 pressure transducer 64. The second foot switch 60 has third 66, fourth 68, fifth 70 and sixth 72 pressure transducers. The transducers are each connected to a corresponding operational amplifier that provides a voltage input to the computer 20. The pressure transducers 62-72 are preferably constructed so that the resistance of each transducer decreases as the surgeon increases the pressure on the foot switches.
Such a transducer is sold by Interlink Electronics. The decreasing transducer resistance increases the input voltage provided to the computer 20 from the operational amplifier. Each transducer corresponds to a predetermined direction within the image displayed by the monitor. In the preferred embodiment, the first pressure transducer 62 corresponds to moving the endoscope toward the image viewed by the surgeon. The second transducer 64 moves the scope away from the image. The third 66 and fourth 68 transducers move the image "up" and "down", respectively, and the fifth 70 and sixth 72 transducers move the image "left" and "right", respectively. The pedal may have a button 73 that enables the foot pedal 50 and disable the voice command feature, or vice versa.
Figure 4 shows a schematic of the computer 20. The computer 20 has a multiplexer 74 which is connected to the pressure transducers of the foot pedal 50 and the position sensors of the arm. The multiplexer 74 is connected to a single analog to digital (A/D) converter 76. The computer 20 also has a processor 78 and memory 80.
The processor 78 is connected to an address decoder 82 and separate digital to analog (D/A) converters 84. Each D/A converter is connected to an actuator 24, 30 and 34. The D/A converters 84 provide analog output signals to the actuators in response to output signals _8_ received from the processor 78. The analog output signals have a sufficient voltage level to energize the electric motors and move the robotic arm assembly. The decoder 82 correlates the addresses provided by the processor with a corresponding D/A converter, so that the correct motors) is driven. The address decoder 82 also provides an address for the input data from the A!D converter 76 so that the data is associated with the correct input channel.
The computer 20 has a phrase recognizer 86 connected to the microphone 40 and the processor 78. The phrase recognizer 86 digitizes voice commands provided by the user through the microphone 40. The voice commands are then processed to convert the spoken words into electronic form. The electronic words are typically generated by matching the user's speech with words stored within the computer 20.
In the preferred embodiment, the recognizer 86 is an electronic board with accompanying software that is marketed by Scott Instruments of Demon, Texas under the trademark "Coretechs Technology".
The electronic words are provided to the processor 78. The processor 78 compares a word, or a combination of words to predefined robot commands that are stored within a library in the memory 80 of the computer 20. If a word, or combination of words match a word or combination of words in the library, the processor 78 provides output commands to the D/A converter 84 to move the robotic arm in accordance with the command.
Figure 5 shows exemplary words and combinations of words that provide robot commands. A grammar process is performed to determine whether the voice commands satisfy certain conditions. The process contains a number of states advanced by the satisfaction of a condition.
_ 220 ~~6 If the voice command provided by the user satisfies a first condition, then the process proceeds to the first state. If a condition of a next state is satisfied then the process proceeds to the next corresponding state, and so forth and so on. For example, to prevent a robot command from being inadvertently spoken, it is desirable to predicate all voice commands with a qualifier. For example, the qualifier may be a name given to the robot such as "AESOP". Therefore when the user provides a voice command, the process initially determines whether the spoken word is AESOP. If the spoken word is not AESOP then the process ends. The term "stop" may be an exception to this rule, wherein the computer will stop arm movement when the user provides a simple "stop" voice command.
If the spoken word is AESOP the process continues to state 1.
The process next determines whether the user has spoken a word that satisfies a condition to advance to states 2-6. These words include "move", "step", "save", "return", "speed", "track instrument" and "track head". The track instrument command is for a system which has the ability to move an endoscope to automatically track the movement of a second instrument that is inserted into the patient. The track head command may enable the system so that the endoscope movement tracks the user's eyes. For example, if the user looks to the right of the image displayed by the monitor, the robot will move the endoscope to move the image in a rightward direction. The move and step commands induce movement of the scope in a desired direction. The save command saves the position of the endoscope within the memory of the computer.
The return command will return the scope to a saved position.
zz oo ~~s From states 2-6 the process will determine whether the user has spoken words that meet the next condition and so forth and so on. When a certain number of conditions have been met, the processor 78 will provide an output command to the D/A converter 84 in accordance with the voice commands. For example, if the user says "AESOP move left", the processor 78 will provide output commands to move the endoscope 12, so that the image displayed by the monitor moves in a leftward direction. The microphone 40 phrase recognizer 86 and grammar process essentially provide the same input function as the foot pedal 50, multiplexer 74 and A/D converter 76.
The processor 78 can also provide the user with feedback regarding the recognized command through the speaker 42 or the monitor 18. For example, when the user states "AESOP move right", after processing the speech, the processor 78 can provide an audio message through the speaker 42, or a visual message on the monitor 18, "AESOP move right". Additionally, the processor 78 can provide messages regarding system errors, or the present state of the system such as "speed is set for slow".
Referring to Figure 6, the processor 78 typically computes the movement of the robotic arm assembly 16 in accordance with the following equations.
(1) a3= n-cos-' x2 +y2 Llz +L22 _, x2 + y2 + L12 - L22 0 = cos 2 - Ll x2 + y2 22 ~0 7~~
a0 = tan-' ZC y x a2=a0+/-D
where;
a2 = angle between the second linkage arm 36 and the x axis.
a3 = angle between the first linkage arm 28 and the longitudinal axis of the second linkage arm 36.
L1 = length of the second linkage arm.
L2 = length of the first linkage arm.
x = x coordinate of the end effector in the first coordinate system.
y = y coordinate of the end effector in the first coordinate system.
To move the end effector to a new location of the x-y plane the processor 78 computes the change in angles a2~and a3 and then provides output signals to move the actuators accordingly. The original angular position of the end effector is provided to the processor 78 by the position sensors. The processor moves the linkage arms an angle that corresponds to the difference between the new location and the original location of the end effector. A differential angle ~a 2 corresponds to the amount of angular displacement provided by the second actuator 34, a differential angle Da3 corresponds to the amount of angular displacement provided by the first actuator 30.
To improve the effectiveness of the system 10, the system is constructed so that the desired movement of the surgical instrument correlates to a direction relative to the image displayed by the monitor.
Thus when the surgeon commands the scope to move up, the scope always appears to move in the up direction. To accomplish this result, the processor 78 converts the desired movement of the end of the endoscope in the third coordinate system to coordinates in the second coordinate system, and then converts the coordinates of the second coordinate system into the coordinates of the first coordinate system.
Referring to Fig. 2, the desired movement of the endoscope is converted from the third coordinate system to the second coordinate system by using the following transformation matrix;
fix' cos(a6) 0 -sin(a6) fix"
(2) ~y' -sin(a5)sin(a6) cos(a5) -sin(a5)cos(a6) ~y"
cos(a5)sin(a6) sin(a5) cos(a5)cos(a6) ~z"
where;
Ox" = the desired incremental movement of the scope along the x" axis of the third coordinate system.
0y" - the desired incremental movement of the scope along the y" axis of the third coordinate system.
Az "= the desired incremental movement of the scope along the z" axis of the third coordinate system.
a5 = the angle between the z' axis and the scope in the y'-z' plane.
a6 = the angle between the z' axis and the scope in the x'-z' plane.
fix' = the computed incremental movement of the scope along the x' axis of the second coordinate system.
Dy' = the computed incremental movement of the scope along the y' axis of the second coordinate system.
~z' = the computed incremental movement of the scope along the z' axis of the second coordinate system.
The angles a5 and a6 are provided by position sensors located on the end effector 32. The angles a5 and a6 are shown in Fig. 2.
The desired movement of the endoscope is converted from the second coordinate system to the first coordinate system by using the following transformation matrix;
~x cos(ec) -sin(n) 0 fix' (3) Dy = sin(n) cos(n) 0 ~y' Oz 0 0 1 where;
fix' = the computed incremental movement of the scope along the x' axis of the second coordinate system.
0y' = the computed incremental movement of the scope along the y' axis of the second coordinate system.
4z' = the computed incremental movement of the scope along the z' axis of the second coordinate system.
n = is the angle between the first linkage arm and the x axis of the first coordinate system.
~x = the computed incremental movement of the scope along the x axis of the first coordinate system.
Dy = the computed incremental movement of the scope along the y axis of the first coordinate system.
~z = the computed incremental movement of the scope along the z axis of the first coordinate system.
The incremental movements ~x and Dy are inserted into the algorithms described above for computing the angular movements ( D a2 and 0 a3 ) of the robotic arm assembly to determine the amount of rotation that is to be provided by each electric motor. The value Oz is used to determine the amount of linear movement provided by the linear actuator 24.
The surgical instrument is typically coupled to a camera and a viewing screen so that any spinning of the instrument about its own longitudinal axis will result in a corresponding rotation of the image on the viewing screen. Rotation of the instrument and viewing image may disorient the viewer. It is therefore desirable to maintain the orientation of the viewing image. In the preferred embodiment, the end effector has a worm gear (not shown) which rotates the surgical instrument about the longitudinal axis of the instrument. To insure proper orientation of the endoscope 16, the worm gear rotates the instrument 16 about its longitudinal axis an amount Dq6 to insure that the y" axis is oriented in the most vertical direction within the fixed coordinate system. Dq6 is computed from the following cross-products.
Dq6 = zi" x (yo" x yi") 2200 7~s where;
Dq6 = the angle that the instrument is to be rotated about the z"
axis.
yo" = is the vector orientation of the y" axis when the instrument is in the first position.
yi" = is the vector orientation of the y" axis when the instrument is in the second position.
zi" = is the vector orientation of the z" axis when the instrument is in the second position.
The vectors of the yi" and zi" axis are computed with the following algorithms.
cosa6 0 -sin a6 0 (zi"~= -sina5sina6 cosa5 -sina5cosa6 0 cosa5sina6 sina5 cosa5cosa6 1 xi"= z X z1"
yi"= zi" xxi"
where;
a5 = is the angle between the instrument and the z axis in the y-z plane.
a6 = is the angle between the instrument and the z axis in the x-z plane.
z = is the unit vector of the z axis in the first coordinate system.
WO 96/09587 PC"T/US95/12064 -is- 22 QO 7 1 6 The angles a5 and a6 are provided by position sensors. The vector yo" is computed using the angles a5 and a6 of the instrument in the original or first position. For the computation of yi" the angles a5 and a6 of the second position are used in the transformation matrix.
After each arm movement yo" is set to yi" and a new yi" vector and corresponding Dq6 angle are computed and used to re-orient the endoscope. Using the above described algorithms, the worm gear continuously rotates the instrument about its longitudinal axis to insure that the pivotal movement of the endoscope does not cause a corresponding rotation of the viewing image.
The system may have a memory feature to store desired instrument positions within the patient. The memory feature may be enabled either by voice commands or through a button on an input device such as the foot pedal. When a save command is spoken, the coordinates of the end effector in the first coordinate system are saved in a dedicated addresses) of the computer memory. When a return command is spoken, the processor retrieves the data stored in memory and moves the end effector to the coordinates of the effector when the save command was enabled.
The memory feature allows the operator to store the coordinates of the end effector in a first position, move the end effector to a second position and then return to the first position with a simple command.
By way of example, the surgeon may take a wide eye view of the patient from a predetermined location and store the coordinates of that location in memory. Subsequently, the surgeon may manipulate the endoscope to enter cavities, etc. which provide a more narrow view. The surgeon can rapidly move back to the wide eye view by merely stating "AESOP
return to one".
In operation, the user provides spoken words to the microphone.
The phrase recognizer 86 matches the user's speech with stored words and provides matched electronic words to the processor 78. The processor performs a grammar process to determine whether the spoken words are robot commands. If the words are commands, the computer energizes the actuators and moves the endoscope, accordingly. The system also allows the user to control the movement of the endoscope with a foot pedal if voice commands are not desired.
While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those ordinarily skilled in the art.
Referring to the drawings more particularly by reference numbers, Figure 1 show:; si robotic system 10 of the present invention.
The system 10 is typically used in a sterile operating room where a surgeon performs a surgical procedure on a patient. The patient is placed on a operating tables 12. Attached to the table 12 is a robotic arm assembly 14 which can move a surgical instrument 16 relative to the table 12 and the patient. The surgical instrument 16 is typically an endoscope which is inserted into the abdomen of the patient. The endoscope 16 enters the patient through a cannula, wherein the scope 16 rotate about a cannul<:~ pivot point. The endoscope is typically connected to a monitor 18 vrhich allows the surgeon to view the organs, etc. of the patient. Althoug~~ an endoscope is described and shown, it is ' to be understood that the present invention c;m be used with other surgical instruments.
The robotic arm assembly 14 controlled by a computer 20. In the preferred embodiment, the robotic arm assembly 14 includes a linear actuator 24 fixed to the table 12. The linear actuator 24 is connected to a linkage arm assembly 2fi and adapted to move the linkage assembly 26 along the z axis of a first coordinate system.. The first coordinate system also has an x axis and a y axis.
The linkage arm asserzbly 26 includes a first linkage arm 28 attached to a rust rotary actuator 30 and an end effector 32. The first rotary actuator 30 is adapted to rotate the first linkage arm 28 and end effector 32 in a plane perpendicular to the z axis (x-y plane). The first rotary actuator 30 is connected to a second rotary actuator 34 by a ~~ ~0 l16 second linkage arm 36. The second actuator 34 is adapted to rotate the first actuator 30 in the x-y plane. The second rotary actuator 34 is connected to the output shaft of the linear actuator 24. The actuators 24, 30 and 34 rotate in response to output signals provided by the computer 20. As shown in Figure 2, the junction of the endoscope 16 and the end ef~'ector 32 define a second coordinate system which has an x' axis, a y' axis and a z' axis. The junction of the end effector 32 and endoscope 18 also define the origin of a third coordinate system which has a x" axis, a y" axis and a z" axis. The z" axis parallel with the longitudinal axis of the endoscope 16.
The arm assembly may have a pair of passive joints that allow the end effector to be rotated in the direction indicated by the arrows.
The actuators 24, 30 and 34, and joints of the arm may each have position sensors (not shown) that are connected to the computer 20.
The sensors provide positional feedback signals of each corresponding arm component.
The system has a microphone 40 that is connected to the computer 20. The system may also have a speaker 42 that is connected to the computer 20. The microphone 40 and speaker 42 may be mounted to a headset 44 that is worn by the user. Placing the microphone 40 in close proximity to the user reduces the amount of background noise provided to the computer and decreases the probability of an inadvertent input command.
As shown in Figure 3, the system may also have a foot pedal 50.
The foot pedal 22 has a housing 56 that supports a pair of outer first foot switches 58 and a second foot switch 60. One outer foot switch 58 has a first pressure transducer 62 and the other switch has a second 22 ~4 716 pressure transducer 64. The second foot switch 60 has third 66, fourth 68, fifth 70 and sixth 72 pressure transducers. The transducers are each connected to a corresponding operational amplifier that provides a voltage input to the computer 20. The pressure transducers 62-72 are preferably constructed so that the resistance of each transducer decreases as the surgeon increases the pressure on the foot switches.
Such a transducer is sold by Interlink Electronics. The decreasing transducer resistance increases the input voltage provided to the computer 20 from the operational amplifier. Each transducer corresponds to a predetermined direction within the image displayed by the monitor. In the preferred embodiment, the first pressure transducer 62 corresponds to moving the endoscope toward the image viewed by the surgeon. The second transducer 64 moves the scope away from the image. The third 66 and fourth 68 transducers move the image "up" and "down", respectively, and the fifth 70 and sixth 72 transducers move the image "left" and "right", respectively. The pedal may have a button 73 that enables the foot pedal 50 and disable the voice command feature, or vice versa.
Figure 4 shows a schematic of the computer 20. The computer 20 has a multiplexer 74 which is connected to the pressure transducers of the foot pedal 50 and the position sensors of the arm. The multiplexer 74 is connected to a single analog to digital (A/D) converter 76. The computer 20 also has a processor 78 and memory 80.
The processor 78 is connected to an address decoder 82 and separate digital to analog (D/A) converters 84. Each D/A converter is connected to an actuator 24, 30 and 34. The D/A converters 84 provide analog output signals to the actuators in response to output signals _8_ received from the processor 78. The analog output signals have a sufficient voltage level to energize the electric motors and move the robotic arm assembly. The decoder 82 correlates the addresses provided by the processor with a corresponding D/A converter, so that the correct motors) is driven. The address decoder 82 also provides an address for the input data from the A!D converter 76 so that the data is associated with the correct input channel.
The computer 20 has a phrase recognizer 86 connected to the microphone 40 and the processor 78. The phrase recognizer 86 digitizes voice commands provided by the user through the microphone 40. The voice commands are then processed to convert the spoken words into electronic form. The electronic words are typically generated by matching the user's speech with words stored within the computer 20.
In the preferred embodiment, the recognizer 86 is an electronic board with accompanying software that is marketed by Scott Instruments of Demon, Texas under the trademark "Coretechs Technology".
The electronic words are provided to the processor 78. The processor 78 compares a word, or a combination of words to predefined robot commands that are stored within a library in the memory 80 of the computer 20. If a word, or combination of words match a word or combination of words in the library, the processor 78 provides output commands to the D/A converter 84 to move the robotic arm in accordance with the command.
Figure 5 shows exemplary words and combinations of words that provide robot commands. A grammar process is performed to determine whether the voice commands satisfy certain conditions. The process contains a number of states advanced by the satisfaction of a condition.
_ 220 ~~6 If the voice command provided by the user satisfies a first condition, then the process proceeds to the first state. If a condition of a next state is satisfied then the process proceeds to the next corresponding state, and so forth and so on. For example, to prevent a robot command from being inadvertently spoken, it is desirable to predicate all voice commands with a qualifier. For example, the qualifier may be a name given to the robot such as "AESOP". Therefore when the user provides a voice command, the process initially determines whether the spoken word is AESOP. If the spoken word is not AESOP then the process ends. The term "stop" may be an exception to this rule, wherein the computer will stop arm movement when the user provides a simple "stop" voice command.
If the spoken word is AESOP the process continues to state 1.
The process next determines whether the user has spoken a word that satisfies a condition to advance to states 2-6. These words include "move", "step", "save", "return", "speed", "track instrument" and "track head". The track instrument command is for a system which has the ability to move an endoscope to automatically track the movement of a second instrument that is inserted into the patient. The track head command may enable the system so that the endoscope movement tracks the user's eyes. For example, if the user looks to the right of the image displayed by the monitor, the robot will move the endoscope to move the image in a rightward direction. The move and step commands induce movement of the scope in a desired direction. The save command saves the position of the endoscope within the memory of the computer.
The return command will return the scope to a saved position.
zz oo ~~s From states 2-6 the process will determine whether the user has spoken words that meet the next condition and so forth and so on. When a certain number of conditions have been met, the processor 78 will provide an output command to the D/A converter 84 in accordance with the voice commands. For example, if the user says "AESOP move left", the processor 78 will provide output commands to move the endoscope 12, so that the image displayed by the monitor moves in a leftward direction. The microphone 40 phrase recognizer 86 and grammar process essentially provide the same input function as the foot pedal 50, multiplexer 74 and A/D converter 76.
The processor 78 can also provide the user with feedback regarding the recognized command through the speaker 42 or the monitor 18. For example, when the user states "AESOP move right", after processing the speech, the processor 78 can provide an audio message through the speaker 42, or a visual message on the monitor 18, "AESOP move right". Additionally, the processor 78 can provide messages regarding system errors, or the present state of the system such as "speed is set for slow".
Referring to Figure 6, the processor 78 typically computes the movement of the robotic arm assembly 16 in accordance with the following equations.
(1) a3= n-cos-' x2 +y2 Llz +L22 _, x2 + y2 + L12 - L22 0 = cos 2 - Ll x2 + y2 22 ~0 7~~
a0 = tan-' ZC y x a2=a0+/-D
where;
a2 = angle between the second linkage arm 36 and the x axis.
a3 = angle between the first linkage arm 28 and the longitudinal axis of the second linkage arm 36.
L1 = length of the second linkage arm.
L2 = length of the first linkage arm.
x = x coordinate of the end effector in the first coordinate system.
y = y coordinate of the end effector in the first coordinate system.
To move the end effector to a new location of the x-y plane the processor 78 computes the change in angles a2~and a3 and then provides output signals to move the actuators accordingly. The original angular position of the end effector is provided to the processor 78 by the position sensors. The processor moves the linkage arms an angle that corresponds to the difference between the new location and the original location of the end effector. A differential angle ~a 2 corresponds to the amount of angular displacement provided by the second actuator 34, a differential angle Da3 corresponds to the amount of angular displacement provided by the first actuator 30.
To improve the effectiveness of the system 10, the system is constructed so that the desired movement of the surgical instrument correlates to a direction relative to the image displayed by the monitor.
Thus when the surgeon commands the scope to move up, the scope always appears to move in the up direction. To accomplish this result, the processor 78 converts the desired movement of the end of the endoscope in the third coordinate system to coordinates in the second coordinate system, and then converts the coordinates of the second coordinate system into the coordinates of the first coordinate system.
Referring to Fig. 2, the desired movement of the endoscope is converted from the third coordinate system to the second coordinate system by using the following transformation matrix;
fix' cos(a6) 0 -sin(a6) fix"
(2) ~y' -sin(a5)sin(a6) cos(a5) -sin(a5)cos(a6) ~y"
cos(a5)sin(a6) sin(a5) cos(a5)cos(a6) ~z"
where;
Ox" = the desired incremental movement of the scope along the x" axis of the third coordinate system.
0y" - the desired incremental movement of the scope along the y" axis of the third coordinate system.
Az "= the desired incremental movement of the scope along the z" axis of the third coordinate system.
a5 = the angle between the z' axis and the scope in the y'-z' plane.
a6 = the angle between the z' axis and the scope in the x'-z' plane.
fix' = the computed incremental movement of the scope along the x' axis of the second coordinate system.
Dy' = the computed incremental movement of the scope along the y' axis of the second coordinate system.
~z' = the computed incremental movement of the scope along the z' axis of the second coordinate system.
The angles a5 and a6 are provided by position sensors located on the end effector 32. The angles a5 and a6 are shown in Fig. 2.
The desired movement of the endoscope is converted from the second coordinate system to the first coordinate system by using the following transformation matrix;
~x cos(ec) -sin(n) 0 fix' (3) Dy = sin(n) cos(n) 0 ~y' Oz 0 0 1 where;
fix' = the computed incremental movement of the scope along the x' axis of the second coordinate system.
0y' = the computed incremental movement of the scope along the y' axis of the second coordinate system.
4z' = the computed incremental movement of the scope along the z' axis of the second coordinate system.
n = is the angle between the first linkage arm and the x axis of the first coordinate system.
~x = the computed incremental movement of the scope along the x axis of the first coordinate system.
Dy = the computed incremental movement of the scope along the y axis of the first coordinate system.
~z = the computed incremental movement of the scope along the z axis of the first coordinate system.
The incremental movements ~x and Dy are inserted into the algorithms described above for computing the angular movements ( D a2 and 0 a3 ) of the robotic arm assembly to determine the amount of rotation that is to be provided by each electric motor. The value Oz is used to determine the amount of linear movement provided by the linear actuator 24.
The surgical instrument is typically coupled to a camera and a viewing screen so that any spinning of the instrument about its own longitudinal axis will result in a corresponding rotation of the image on the viewing screen. Rotation of the instrument and viewing image may disorient the viewer. It is therefore desirable to maintain the orientation of the viewing image. In the preferred embodiment, the end effector has a worm gear (not shown) which rotates the surgical instrument about the longitudinal axis of the instrument. To insure proper orientation of the endoscope 16, the worm gear rotates the instrument 16 about its longitudinal axis an amount Dq6 to insure that the y" axis is oriented in the most vertical direction within the fixed coordinate system. Dq6 is computed from the following cross-products.
Dq6 = zi" x (yo" x yi") 2200 7~s where;
Dq6 = the angle that the instrument is to be rotated about the z"
axis.
yo" = is the vector orientation of the y" axis when the instrument is in the first position.
yi" = is the vector orientation of the y" axis when the instrument is in the second position.
zi" = is the vector orientation of the z" axis when the instrument is in the second position.
The vectors of the yi" and zi" axis are computed with the following algorithms.
cosa6 0 -sin a6 0 (zi"~= -sina5sina6 cosa5 -sina5cosa6 0 cosa5sina6 sina5 cosa5cosa6 1 xi"= z X z1"
yi"= zi" xxi"
where;
a5 = is the angle between the instrument and the z axis in the y-z plane.
a6 = is the angle between the instrument and the z axis in the x-z plane.
z = is the unit vector of the z axis in the first coordinate system.
WO 96/09587 PC"T/US95/12064 -is- 22 QO 7 1 6 The angles a5 and a6 are provided by position sensors. The vector yo" is computed using the angles a5 and a6 of the instrument in the original or first position. For the computation of yi" the angles a5 and a6 of the second position are used in the transformation matrix.
After each arm movement yo" is set to yi" and a new yi" vector and corresponding Dq6 angle are computed and used to re-orient the endoscope. Using the above described algorithms, the worm gear continuously rotates the instrument about its longitudinal axis to insure that the pivotal movement of the endoscope does not cause a corresponding rotation of the viewing image.
The system may have a memory feature to store desired instrument positions within the patient. The memory feature may be enabled either by voice commands or through a button on an input device such as the foot pedal. When a save command is spoken, the coordinates of the end effector in the first coordinate system are saved in a dedicated addresses) of the computer memory. When a return command is spoken, the processor retrieves the data stored in memory and moves the end effector to the coordinates of the effector when the save command was enabled.
The memory feature allows the operator to store the coordinates of the end effector in a first position, move the end effector to a second position and then return to the first position with a simple command.
By way of example, the surgeon may take a wide eye view of the patient from a predetermined location and store the coordinates of that location in memory. Subsequently, the surgeon may manipulate the endoscope to enter cavities, etc. which provide a more narrow view. The surgeon can rapidly move back to the wide eye view by merely stating "AESOP
return to one".
In operation, the user provides spoken words to the microphone.
The phrase recognizer 86 matches the user's speech with stored words and provides matched electronic words to the processor 78. The processor performs a grammar process to determine whether the spoken words are robot commands. If the words are commands, the computer energizes the actuators and moves the endoscope, accordingly. The system also allows the user to control the movement of the endoscope with a foot pedal if voice commands are not desired.
While certain exemplary embodiments have been described and shown in the accompanying drawings, it is to be understood that such embodiments are merely illustrative of and not restrictive on the broad invention, and that this invention not be limited to the specific constructions and arrangements shown and described, since various other modifications may occur to those ordinarily skilled in the art.
Claims (26)
1. A robotic system that controls a surgical instrument, comprising:
a mechanism that can move the surgical instrument;
an audio input device that receives a voice qualifier and a voice command from a user to move the surgical instrument, and provides an output signal that correspond to the voice command; and, a controller that receives said output signal from said audio input device and provides a movement output signal to said mechanism to move said mechanism and the surgical instrument in accordance with the voice command only when the voice qualifier accompanies the voice command.
a mechanism that can move the surgical instrument;
an audio input device that receives a voice qualifier and a voice command from a user to move the surgical instrument, and provides an output signal that correspond to the voice command; and, a controller that receives said output signal from said audio input device and provides a movement output signal to said mechanism to move said mechanism and the surgical instrument in accordance with the voice command only when the voice qualifier accompanies the voice command.
2. The system as recited in claim 1, wherein said controller provides said movement output signal only when a plurality of voice commands are provided in a predetermined sequence.
3. The system as recited in claim 1, wherein said mechanism is a robotic arm assembly.
4. The system as recited in claim 1, further comprising a foot pedal that is coupled to said controller.
5. The system as recited in claim 1, further comprising a speaker coupled to said controller to generate an audio message.
6. The system as recited in claim 1, further comprising a monitor coupled to said controller to generate a visual message.
7. A robotic system that controls a surgical instrument, comprising:
a mechanism that can move the surgical instrument;
audio input means for receiving a voice qualifier and a voice command from a user to move the surgical instrument and providing an output signal that correspond to the voice command; and, controller means for receiving said output signal from said audio input means and providing a movement output signal to said mechanism to move said mechanism and the surgical instrument in accordance with the voice command only when the voice qualifier accompanies the voice command.
a mechanism that can move the surgical instrument;
audio input means for receiving a voice qualifier and a voice command from a user to move the surgical instrument and providing an output signal that correspond to the voice command; and, controller means for receiving said output signal from said audio input means and providing a movement output signal to said mechanism to move said mechanism and the surgical instrument in accordance with the voice command only when the voice qualifier accompanies the voice command.
8. The system as recited in claim 7, wherein said controller means provides said movement output signal only when a plurality of voice commands are provided in a predetermined sequence.
9. The system as recited in claim 7, wherein said mechanism is a robotic arm assembly.
10. The system as recited in claim 7, further comprising a foot pedal that is coupled to said controller means.
11. The system as recited in claim 7, further comprising speaker means for generating an audio message.
12. The system as recited in claim 7, further comprising monitor means for generating a visual message.
13. A method for moving a surgical instrument, comprising the steps of:
a) generating a voice qualifier and a voice command to move a mechanism that holds the surgical instrument;
b) determining whether the voice qualifier matches a first robot command;
c) determining whether the voice command matches a second robot command if the voice qualifier matched the first robot command; and, d) moving the surgical instrument with said mechanism only if the voice qualifier matches the first robot command and the voice command matches the second robot command.
a) generating a voice qualifier and a voice command to move a mechanism that holds the surgical instrument;
b) determining whether the voice qualifier matches a first robot command;
c) determining whether the voice command matches a second robot command if the voice qualifier matched the first robot command; and, d) moving the surgical instrument with said mechanism only if the voice qualifier matches the first robot command and the voice command matches the second robot command.
14. The method as recited in claim 13, wherein the surgical instrument is moved only when a plurality of voice commands are provided in a predetermined sequence.
15. A medical robotic system that controls a surgical instrument, comprising:
an actuator;
a linkage attached to said actuator and that can be coupled to the surgical instrument;
an audio input device that receives a voice qualifier and a voice command from a user; and, a controller that is coupled to said audio input device and provides a movement output signal to said actuator to move said linkage in accordance with the voice command only when the voice qualifier accompanies the voice command.
an actuator;
a linkage attached to said actuator and that can be coupled to the surgical instrument;
an audio input device that receives a voice qualifier and a voice command from a user; and, a controller that is coupled to said audio input device and provides a movement output signal to said actuator to move said linkage in accordance with the voice command only when the voice qualifier accompanies the voice command.
16. The system as recited in claim 1, wherein said controller provides said movement output signal only when a plurality of voice commands are provided in a predetermined sequence.
17. The system as recited in claim 1, wherein said mechanism is a robotic arm assembly.
18. The system as recited in claim 1, further comprising a foot pedal coupled to said controller.
19. The system as recited in claim 1, further comprising a speaker coupled to said controller to generate an audio message.
20. The system as recited in claim 1, further comprising a monitor coupled to said controller to generate a visual message.
21. A robotic system that controls a surgical instrument, comprising:
a mechanism that can move the surgical instrument;
audio input means for receiving a voice qualifier and a voice command from a user to move the surgical instrument; and, controller means for moving said mechanism in accordance with the voice command only when the voice qualifier accompanies the voice command.
a mechanism that can move the surgical instrument;
audio input means for receiving a voice qualifier and a voice command from a user to move the surgical instrument; and, controller means for moving said mechanism in accordance with the voice command only when the voice qualifier accompanies the voice command.
22. The system as recited in claim 21, wherein said controller means provides said movement output signal only when a plurality of voice commands are provided in a predetermined sequence.
23. The system as recited in claim 21, wherein said mechanism is a robotic arm assembly.
24. The system as recited in claim 21, further comprising a foot pedal coupled to said controller means.
25. The system as recited in claim 21, further comprising speaker means for generating an audio message.
26. The system as recited in claim 21, further comprising monitor means for generating a visual message.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US08/310,665 | 1994-09-22 | ||
US08/310,665 US6463361B1 (en) | 1994-09-22 | 1994-09-22 | Speech interface for an automated endoscopic system |
PCT/US1995/012064 WO1996009587A1 (en) | 1994-09-22 | 1995-09-21 | A speech interface for an automated endoscopic system |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2200716A1 CA2200716A1 (en) | 1996-03-28 |
CA2200716C true CA2200716C (en) | 2003-06-17 |
Family
ID=23203577
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002200716A Expired - Lifetime CA2200716C (en) | 1994-09-22 | 1995-09-21 | A speech interface for an automated endoscopic system |
Country Status (4)
Country | Link |
---|---|
US (3) | US6463361B1 (en) |
AU (1) | AU3721195A (en) |
CA (1) | CA2200716C (en) |
WO (1) | WO1996009587A1 (en) |
Families Citing this family (226)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6646541B1 (en) * | 1996-06-24 | 2003-11-11 | Computer Motion, Inc. | General purpose distributed operating room control system |
US6463361B1 (en) * | 1994-09-22 | 2002-10-08 | Computer Motion, Inc. | Speech interface for an automated endoscopic system |
US7053752B2 (en) * | 1996-08-06 | 2006-05-30 | Intuitive Surgical | General purpose distributed operating room control system |
US6714841B1 (en) * | 1995-09-15 | 2004-03-30 | Computer Motion, Inc. | Head cursor control interface for an automated endoscope system for optimal positioning |
US6496099B2 (en) | 1996-06-24 | 2002-12-17 | Computer Motion, Inc. | General purpose distributed operating room control system |
US6911916B1 (en) * | 1996-06-24 | 2005-06-28 | The Cleveland Clinic Foundation | Method and apparatus for accessing medical data over a network |
US6642836B1 (en) | 1996-08-06 | 2003-11-04 | Computer Motion, Inc. | General purpose distributed operating room control system |
US6364888B1 (en) | 1996-09-09 | 2002-04-02 | Intuitive Surgical, Inc. | Alignment of master and slave in a minimally invasive surgical apparatus |
US6231565B1 (en) | 1997-06-18 | 2001-05-15 | United States Surgical Corporation | Robotic arm DLUs for performing surgical tasks |
WO1999021165A1 (en) * | 1997-10-20 | 1999-04-29 | Computer Motion Inc. | General purpose distributed operating room control system |
US7758569B2 (en) | 1998-02-24 | 2010-07-20 | Hansen Medical, Inc. | Interchangeable surgical instrument |
US8414598B2 (en) | 1998-02-24 | 2013-04-09 | Hansen Medical, Inc. | Flexible instrument |
US7901399B2 (en) | 1998-02-24 | 2011-03-08 | Hansen Medical, Inc. | Interchangeable surgical instrument |
US7789875B2 (en) | 1998-02-24 | 2010-09-07 | Hansen Medical, Inc. | Surgical instruments |
US20020095175A1 (en) * | 1998-02-24 | 2002-07-18 | Brock David L. | Flexible instrument |
US7713190B2 (en) | 1998-02-24 | 2010-05-11 | Hansen Medical, Inc. | Flexible instrument |
US7169141B2 (en) | 1998-02-24 | 2007-01-30 | Hansen Medical, Inc. | Surgical instrument |
US7775972B2 (en) | 1998-02-24 | 2010-08-17 | Hansen Medical, Inc. | Flexible instrument |
US8303576B2 (en) | 1998-02-24 | 2012-11-06 | Hansen Medical, Inc. | Interchangeable surgical instrument |
EP1097447A1 (en) * | 1998-07-23 | 2001-05-09 | Siemens Aktiengesellschaft | Method and device for recognizing predetermined key words in spoken language |
US6493608B1 (en) | 1999-04-07 | 2002-12-10 | Intuitive Surgical, Inc. | Aspects of a control system of a minimally invasive surgical apparatus |
US6224542B1 (en) | 1999-01-04 | 2001-05-01 | Stryker Corporation | Endoscopic camera system with non-mechanical zoom |
ES2150880B1 (en) * | 1999-01-28 | 2001-08-16 | Univ Malaga | ROBOTIC ASSISTANT SYSTEM FOR LAPAROSCOPIC SURGERY. |
DE19915060A1 (en) * | 1999-04-01 | 2000-11-02 | Erbe Elektromedizin | Surgical instrument |
DE19915062C1 (en) | 1999-04-01 | 2001-02-08 | Erbe Elektromedizin | Surgical instrument |
DE60029234T2 (en) | 1999-05-10 | 2007-05-31 | Hansen Medical, Inc., Mountain View | Surgical instrument |
US6591239B1 (en) | 1999-12-09 | 2003-07-08 | Steris Inc. | Voice controlled surgical suite |
US7035716B2 (en) * | 2001-01-29 | 2006-04-25 | The Acrobot Company Limited | Active-constraint robots |
US7699835B2 (en) | 2001-02-15 | 2010-04-20 | Hansen Medical, Inc. | Robotically controlled surgical instruments |
US8414505B1 (en) | 2001-02-15 | 2013-04-09 | Hansen Medical, Inc. | Catheter driver system |
US20030135204A1 (en) | 2001-02-15 | 2003-07-17 | Endo Via Medical, Inc. | Robotically controlled medical instrument with a flexible section |
US7766894B2 (en) | 2001-02-15 | 2010-08-03 | Hansen Medical, Inc. | Coaxial catheter system |
US20020151837A1 (en) * | 2001-04-16 | 2002-10-17 | Surgicon Inc. | Surgical irrigation apparatus and methods for use |
US20020165524A1 (en) | 2001-05-01 | 2002-11-07 | Dan Sanchez | Pivot point arm for a robotic system used to perform a surgical procedure |
US20020166557A1 (en) * | 2001-05-09 | 2002-11-14 | David Cooper | Mask with a built-in microphone |
US6587750B2 (en) | 2001-09-25 | 2003-07-01 | Intuitive Surgical, Inc. | Removable infinite roll master grip handle and touch sensor for robotic surgery |
US7164968B2 (en) * | 2002-04-05 | 2007-01-16 | The Trustees Of Columbia University In The City Of New York | Robotic scrub nurse |
US7674270B2 (en) * | 2002-05-02 | 2010-03-09 | Laparocision, Inc | Apparatus for positioning a medical instrument |
US7347862B2 (en) * | 2002-05-02 | 2008-03-25 | Layer James H | Apparatus for positioning a medical instrument relative to a patient |
US6925357B2 (en) * | 2002-07-25 | 2005-08-02 | Intouch Health, Inc. | Medical tele-robotic system |
US7593030B2 (en) * | 2002-07-25 | 2009-09-22 | Intouch Technologies, Inc. | Tele-robotic videoconferencing in a corporate environment |
US20040162637A1 (en) | 2002-07-25 | 2004-08-19 | Yulun Wang | Medical tele-robotic system with a master remote station with an arbitrator |
AU2003257309A1 (en) | 2002-08-13 | 2004-02-25 | Microbotics Corporation | Microsurgical robot system |
US20040176751A1 (en) | 2002-08-14 | 2004-09-09 | Endovia Medical, Inc. | Robotic medical instrument system |
US7259906B1 (en) | 2002-09-03 | 2007-08-21 | Cheetah Omni, Llc | System and method for voice control of medical devices |
US7331967B2 (en) | 2002-09-09 | 2008-02-19 | Hansen Medical, Inc. | Surgical instrument coupling mechanism |
GB0222265D0 (en) * | 2002-09-25 | 2002-10-30 | Imp College Innovations Ltd | Control of robotic manipulation |
JP4096687B2 (en) * | 2002-10-09 | 2008-06-04 | 株式会社デンソー | EEPROM and method of manufacturing the same |
ES2237245B1 (en) * | 2002-11-07 | 2006-06-01 | Universidad De Alcala | Auxiliary robot for performing surgical work in operating room, has processing unit to control movement of robot arm using captured and digitized images of operating room and digitized vocal commands from operator |
GB0228937D0 (en) * | 2002-12-12 | 2003-01-15 | Univ Dundee | Control apparatus for imaging device |
GB2396905A (en) * | 2002-12-31 | 2004-07-07 | Armstrong Healthcare Ltd | A device for generating a control signal |
US7158859B2 (en) * | 2003-01-15 | 2007-01-02 | Intouch Technologies, Inc. | 5 degrees of freedom mobile robot |
US7171286B2 (en) * | 2003-02-24 | 2007-01-30 | Intouch Technologies, Inc. | Healthcare tele-robotic system with a robot that also functions as a remote station |
US7158860B2 (en) | 2003-02-24 | 2007-01-02 | Intouch Technologies, Inc. | Healthcare tele-robotic system which allows parallel remote station observation |
US7262573B2 (en) | 2003-03-06 | 2007-08-28 | Intouch Technologies, Inc. | Medical tele-robotic system with a head worn device |
US7736300B2 (en) * | 2003-04-14 | 2010-06-15 | Softscope Medical Technologies, Inc. | Self-propellable apparatus and method |
FR2853983A1 (en) * | 2003-04-17 | 2004-10-22 | Philippe Bellanger | Manual gesture assisting and training device for design field, has stimuli generator to inform position of tool relative to material, to operator by increase of reality of actions that his job implies |
US8007511B2 (en) | 2003-06-06 | 2011-08-30 | Hansen Medical, Inc. | Surgical instrument design |
US6888333B2 (en) * | 2003-07-02 | 2005-05-03 | Intouch Health, Inc. | Holonomic platform for a robot |
US7042184B2 (en) | 2003-07-08 | 2006-05-09 | Board Of Regents Of The University Of Nebraska | Microrobot for surgical applications |
US7960935B2 (en) | 2003-07-08 | 2011-06-14 | The Board Of Regents Of The University Of Nebraska | Robotic devices with agent delivery components and related methods |
US7247139B2 (en) * | 2003-09-09 | 2007-07-24 | Ge Medical Systems Global Technology Company, Llc | Method and apparatus for natural voice control of an ultrasound machine |
US7161322B2 (en) * | 2003-11-18 | 2007-01-09 | Intouch Technologies, Inc. | Robot with a manipulator arm |
US7292912B2 (en) * | 2003-12-05 | 2007-11-06 | Lntouch Technologies, Inc. | Door knocker control system for a remote controlled teleconferencing robot |
US7813836B2 (en) | 2003-12-09 | 2010-10-12 | Intouch Technologies, Inc. | Protocol for a remotely controlled videoconferencing robot |
US20050204438A1 (en) | 2004-02-26 | 2005-09-15 | Yulun Wang | Graphical interface for a remote presence system |
US8725525B2 (en) * | 2004-04-13 | 2014-05-13 | Olympus Corporation | Endoscope system |
US8077963B2 (en) | 2004-07-13 | 2011-12-13 | Yulun Wang | Mobile robot with a head-based movement mapping scheme |
US7752050B1 (en) * | 2004-09-03 | 2010-07-06 | Stryker Corporation | Multiple-user voice-based control of devices in an endoscopic imaging system |
US8443279B1 (en) | 2004-10-13 | 2013-05-14 | Stryker Corporation | Voice-responsive annotation of video generated by an endoscopic camera |
DE102004059166A1 (en) * | 2004-12-08 | 2006-06-29 | Siemens Ag | Operating method for support unit for medical-technical system entails support unit in reaction to speech input sending out both acoustic and visual output to enquirer |
US7222000B2 (en) * | 2005-01-18 | 2007-05-22 | Intouch Technologies, Inc. | Mobile videoconferencing platform with automatic shut-off features |
WO2006090402A1 (en) * | 2005-02-23 | 2006-08-31 | Hewlett-Packard Development Company, L.P. | System and method of voice communication with machines |
US9943372B2 (en) | 2005-04-18 | 2018-04-17 | M.S.T. Medical Surgery Technologies Ltd. | Device having a wearable interface for improving laparoscopic surgery and methods for use thereof |
US20060259193A1 (en) * | 2005-05-12 | 2006-11-16 | Yulun Wang | Telerobotic system with a dual application screen presentation |
US20070015999A1 (en) * | 2005-07-15 | 2007-01-18 | Heldreth Mark A | System and method for providing orthopaedic surgical information to a surgeon |
US20070078678A1 (en) * | 2005-09-30 | 2007-04-05 | Disilvestro Mark R | System and method for performing a computer assisted orthopaedic surgical procedure |
US9198728B2 (en) | 2005-09-30 | 2015-12-01 | Intouch Technologies, Inc. | Multi-camera mobile teleconferencing platform |
US7697827B2 (en) | 2005-10-17 | 2010-04-13 | Konicek Jeffrey C | User-friendlier interfaces for a camera |
US7519253B2 (en) | 2005-11-18 | 2009-04-14 | Omni Sciences, Inc. | Broadband or mid-infrared fiber light sources |
US7620553B2 (en) * | 2005-12-20 | 2009-11-17 | Storz Endoskop Produktions Gmbh | Simultaneous support of isolated and connected phrase command recognition in automatic speech recognition systems |
US20060253109A1 (en) * | 2006-02-08 | 2006-11-09 | David Chu | Surgical robotic helping hand system |
US7769492B2 (en) | 2006-02-22 | 2010-08-03 | Intouch Technologies, Inc. | Graphical interface for a remote presence system |
US8635082B2 (en) | 2006-05-25 | 2014-01-21 | DePuy Synthes Products, LLC | Method and system for managing inventories of orthopaedic implants |
US8849679B2 (en) | 2006-06-15 | 2014-09-30 | Intouch Technologies, Inc. | Remote controlled robot system that provides medical images |
US8015014B2 (en) * | 2006-06-16 | 2011-09-06 | Storz Endoskop Produktions Gmbh | Speech recognition system with user profiles management component |
CA3068216C (en) | 2006-06-22 | 2023-03-07 | Board Of Regents Of The University Of Nebraska | Magnetically coupleable robotic devices and related methods |
US9579088B2 (en) | 2007-02-20 | 2017-02-28 | Board Of Regents Of The University Of Nebraska | Methods, systems, and devices for surgical visualization and device manipulation |
US8679096B2 (en) | 2007-06-21 | 2014-03-25 | Board Of Regents Of The University Of Nebraska | Multifunctional operational component for robotic devices |
US8502876B2 (en) * | 2006-09-12 | 2013-08-06 | Storz Endoskop Producktions GmbH | Audio, visual and device data capturing system with real-time speech recognition command and control system |
US9514746B2 (en) * | 2006-09-26 | 2016-12-06 | Storz Endoskop Produktions Gmbh | System and method for hazard mitigation in voice-driven control applications |
US20080081956A1 (en) * | 2006-09-29 | 2008-04-03 | Jayesh Shah | System and method for integrating voice with a medical device |
US7761185B2 (en) * | 2006-10-03 | 2010-07-20 | Intouch Technologies, Inc. | Remote presence display through remotely controlled robot |
US8037179B2 (en) * | 2006-11-02 | 2011-10-11 | Storz Endoskop Produktions Gmbh | Device control system employing extensible markup language for defining information resources |
US20080167662A1 (en) * | 2007-01-08 | 2008-07-10 | Kurtz Anthony D | Tactile feel apparatus for use with robotic operations |
US8265793B2 (en) | 2007-03-20 | 2012-09-11 | Irobot Corporation | Mobile robot for telecommunication |
JP4916011B2 (en) * | 2007-03-20 | 2012-04-11 | 株式会社日立製作所 | Master / slave manipulator system |
US9160783B2 (en) | 2007-05-09 | 2015-10-13 | Intouch Technologies, Inc. | Robot system that operates through a network firewall |
EP3078344B1 (en) | 2007-07-12 | 2020-02-26 | Board of Regents of the University of Nebraska | Actuation in robotic devices |
JP2010536435A (en) | 2007-08-15 | 2010-12-02 | ボード オブ リージェンツ オブ ザ ユニバーシティ オブ ネブラスカ | Medical inflation, attachment and delivery devices and associated methods |
CA2695619C (en) | 2007-08-15 | 2015-11-24 | Board Of Regents Of The University Of Nebraska | Modular and cooperative medical devices and related systems and methods |
US8116910B2 (en) | 2007-08-23 | 2012-02-14 | Intouch Technologies, Inc. | Telepresence robot with a printer |
US8265949B2 (en) | 2007-09-27 | 2012-09-11 | Depuy Products, Inc. | Customized patient surgical plan |
ES2802126T3 (en) | 2007-09-30 | 2021-01-15 | Depuy Products Inc | Patient Specific Custom Orthopedic Surgical Instrument |
US8795326B2 (en) | 2007-10-05 | 2014-08-05 | Covidien Lp | Expanding seal anchor for single incision surgery |
GB2454664A (en) * | 2007-11-13 | 2009-05-20 | Sandor Mihaly Veres | Voice Actuated Robot |
US20090175464A1 (en) * | 2008-01-08 | 2009-07-09 | Somen Bryan M | Headset For Voice Activation/Confirmation Enabled Surgical Console |
US10875182B2 (en) | 2008-03-20 | 2020-12-29 | Teladoc Health, Inc. | Remote presence system mounted to operating room hardware |
US8179418B2 (en) | 2008-04-14 | 2012-05-15 | Intouch Technologies, Inc. | Robotic based health care system |
US8170241B2 (en) | 2008-04-17 | 2012-05-01 | Intouch Technologies, Inc. | Mobile tele-presence system with a microphone system |
US9193065B2 (en) | 2008-07-10 | 2015-11-24 | Intouch Technologies, Inc. | Docking system for a tele-presence robot |
US9842192B2 (en) | 2008-07-11 | 2017-12-12 | Intouch Technologies, Inc. | Tele-presence robot system with multi-cast features |
US8340819B2 (en) | 2008-09-18 | 2012-12-25 | Intouch Technologies, Inc. | Mobile videoconferencing robot system with network adaptive driving |
US20100100080A1 (en) * | 2008-10-16 | 2010-04-22 | Huculak John C | System and method for voice activation of surgical instruments |
US8996165B2 (en) | 2008-10-21 | 2015-03-31 | Intouch Technologies, Inc. | Telepresence robot with a camera boom |
US9138891B2 (en) | 2008-11-25 | 2015-09-22 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
US8463435B2 (en) | 2008-11-25 | 2013-06-11 | Intouch Technologies, Inc. | Server connectivity control for tele-presence robot |
FR2939022B1 (en) | 2008-11-28 | 2012-02-17 | Assistance Publique Hopitaux Paris | DEVICE FOR CONTROLLING THE DISPLACEMENT OF A SURGICAL INSTRUMENT. |
US8849680B2 (en) | 2009-01-29 | 2014-09-30 | Intouch Technologies, Inc. | Documentation through a remote presence robot |
DE102009010263B4 (en) * | 2009-02-24 | 2011-01-20 | Reiner Kunz | Method for navigating an endoscopic instrument during technical endoscopy and associated device |
US8897920B2 (en) | 2009-04-17 | 2014-11-25 | Intouch Technologies, Inc. | Tele-presence robot system with software modularity, projector and laser pointer |
US11399153B2 (en) | 2009-08-26 | 2022-07-26 | Teladoc Health, Inc. | Portable telepresence apparatus |
US8384755B2 (en) | 2009-08-26 | 2013-02-26 | Intouch Technologies, Inc. | Portable remote presence robot |
BR112012005693A2 (en) * | 2009-09-17 | 2016-02-23 | Fujifilm Corp | propelling device with active resizing capability |
KR101609281B1 (en) | 2009-10-01 | 2016-04-05 | 마코 서지컬 코포레이션 | Tool, kit-of-parts for multi-functional tool, and robotic system for same |
US8543240B2 (en) * | 2009-11-13 | 2013-09-24 | Intuitive Surgical Operations, Inc. | Master finger tracking device and method of use in a minimally invasive surgical system |
US8996173B2 (en) | 2010-09-21 | 2015-03-31 | Intuitive Surgical Operations, Inc. | Method and apparatus for hand gesture control in a minimally invasive surgical system |
US8935003B2 (en) | 2010-09-21 | 2015-01-13 | Intuitive Surgical Operations | Method and system for hand presence detection in a minimally invasive surgical system |
US8521331B2 (en) | 2009-11-13 | 2013-08-27 | Intuitive Surgical Operations, Inc. | Patient-side surgeon interface for a minimally invasive, teleoperated surgical instrument |
CA2784883A1 (en) | 2009-12-17 | 2011-06-23 | Board Of Regents Of The University Of Nebraska | Modular and cooperative medical devices and related systems and methods |
WO2011084863A2 (en) | 2010-01-07 | 2011-07-14 | Cheetah Omni, Llc | Fiber lasers and mid-infrared light sources in methods and systems for selective biological tissue processing and spectroscopy |
US11154981B2 (en) | 2010-02-04 | 2021-10-26 | Teladoc Health, Inc. | Robot user interface for telepresence robot system |
US8670017B2 (en) | 2010-03-04 | 2014-03-11 | Intouch Technologies, Inc. | Remote presence system including a cart that supports a robot face and an overhead camera |
US20110218550A1 (en) * | 2010-03-08 | 2011-09-08 | Tyco Healthcare Group Lp | System and method for determining and adjusting positioning and orientation of a surgical device |
US8918213B2 (en) | 2010-05-20 | 2014-12-23 | Irobot Corporation | Mobile human interface robot |
US8935005B2 (en) | 2010-05-20 | 2015-01-13 | Irobot Corporation | Operating a mobile robot |
US9014848B2 (en) | 2010-05-20 | 2015-04-21 | Irobot Corporation | Mobile robot system |
US10343283B2 (en) | 2010-05-24 | 2019-07-09 | Intouch Technologies, Inc. | Telepresence robot system that can be accessed by a cellular phone |
US10808882B2 (en) | 2010-05-26 | 2020-10-20 | Intouch Technologies, Inc. | Tele-robotic system with a robot face placed on a chair |
US8968267B2 (en) | 2010-08-06 | 2015-03-03 | Board Of Regents Of The University Of Nebraska | Methods and systems for handling or delivering materials for natural orifice surgery |
US20120065972A1 (en) * | 2010-09-12 | 2012-03-15 | Var Systems Ltd. | Wireless voice recognition control system for controlling a welder power supply by voice commands |
US9486189B2 (en) | 2010-12-02 | 2016-11-08 | Hitachi Aloka Medical, Ltd. | Assembly for use with surgery system |
US9264664B2 (en) | 2010-12-03 | 2016-02-16 | Intouch Technologies, Inc. | Systems and methods for dynamic bandwidth allocation |
US8930019B2 (en) | 2010-12-30 | 2015-01-06 | Irobot Corporation | Mobile human interface robot |
US9323250B2 (en) | 2011-01-28 | 2016-04-26 | Intouch Technologies, Inc. | Time-dependent navigation of telepresence robots |
KR102018763B1 (en) | 2011-01-28 | 2019-09-05 | 인터치 테크놀로지스 인코퍼레이티드 | Interfacing with a mobile telepresence robot |
US10769739B2 (en) | 2011-04-25 | 2020-09-08 | Intouch Technologies, Inc. | Systems and methods for management of information among medical providers and facilities |
US20140139616A1 (en) | 2012-01-27 | 2014-05-22 | Intouch Technologies, Inc. | Enhanced Diagnostics for a Telepresence Robot |
US9098611B2 (en) | 2012-11-26 | 2015-08-04 | Intouch Technologies, Inc. | Enhanced video interaction for a user interface of a telepresence network |
EP3977951B1 (en) | 2011-06-10 | 2023-11-08 | Board of Regents of the University of Nebraska | Surgical end effector |
EP2732344B1 (en) | 2011-07-11 | 2019-06-05 | Board of Regents of the University of Nebraska | Robotic surgical system |
US10866783B2 (en) * | 2011-08-21 | 2020-12-15 | Transenterix Europe S.A.R.L. | Vocally activated surgical control system |
US11561762B2 (en) * | 2011-08-21 | 2023-01-24 | Asensus Surgical Europe S.A.R.L. | Vocally actuated surgical control system |
US9757206B2 (en) | 2011-08-21 | 2017-09-12 | M.S.T. Medical Surgery Technologies Ltd | Device and method for assisting laparoscopic surgery—rule based approach |
US9204939B2 (en) | 2011-08-21 | 2015-12-08 | M.S.T. Medical Surgery Technologies Ltd. | Device and method for assisting laparoscopic surgery—rule based approach |
US9795282B2 (en) | 2011-09-20 | 2017-10-24 | M.S.T. Medical Surgery Technologies Ltd | Device and method for maneuvering endoscope |
JP5893330B2 (en) * | 2011-10-18 | 2016-03-23 | オリンパス株式会社 | Operation input device and method for initializing operation input device |
WO2013061857A1 (en) * | 2011-10-25 | 2013-05-02 | オリンパスメディカルシステムズ株式会社 | Endoscope operation system |
US8836751B2 (en) | 2011-11-08 | 2014-09-16 | Intouch Technologies, Inc. | Tele-presence system with a user interface that displays different communication links |
WO2013106569A2 (en) | 2012-01-10 | 2013-07-18 | Board Of Regents Of The University Of Nebraska | Methods, systems, and devices for surgical access and insertion |
US9445876B2 (en) | 2012-02-27 | 2016-09-20 | Covidien Lp | Glove with sensory elements incorporated therein for controlling at least one surgical instrument |
US9251313B2 (en) | 2012-04-11 | 2016-02-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
US8902278B2 (en) | 2012-04-11 | 2014-12-02 | Intouch Technologies, Inc. | Systems and methods for visualizing and managing telepresence devices in healthcare networks |
EP4357083A2 (en) | 2012-05-01 | 2024-04-24 | Board of Regents of the University of Nebraska | Single site robotic device and related systems and methods |
US9361021B2 (en) | 2012-05-22 | 2016-06-07 | Irobot Corporation | Graphical user interfaces including touchpad driving interfaces for telemedicine devices |
WO2013176758A1 (en) | 2012-05-22 | 2013-11-28 | Intouch Technologies, Inc. | Clinical workflows utilizing autonomous and semi-autonomous telemedicine devices |
EP3680071B1 (en) | 2012-06-22 | 2021-09-01 | Board of Regents of the University of Nebraska | Local control robotic surgical devices |
EP2882331A4 (en) | 2012-08-08 | 2016-03-23 | Univ Nebraska | Robotic surgical devices, systems, and related methods |
US9770305B2 (en) | 2012-08-08 | 2017-09-26 | Board Of Regents Of The University Of Nebraska | Robotic surgical devices, systems, and related methods |
US9060674B2 (en) | 2012-10-11 | 2015-06-23 | Karl Storz Imaging, Inc. | Auto zoom for video camera |
WO2014143276A2 (en) | 2012-12-31 | 2014-09-18 | Omni Medsci, Inc. | Short-wave infrared super-continuum lasers for natural gas leak detection, exploration, and other active remote sensing applications |
WO2014105520A1 (en) | 2012-12-31 | 2014-07-03 | Omni Medsci, Inc. | Near-infrared lasers for non-invasive monitoring of glucose, ketones, hba1c, and other blood constituents |
US9993159B2 (en) | 2012-12-31 | 2018-06-12 | Omni Medsci, Inc. | Near-infrared super-continuum lasers for early detection of breast and other cancers |
US9730621B2 (en) | 2012-12-31 | 2017-08-15 | Dexcom, Inc. | Remote monitoring of analyte measurements |
US9500635B2 (en) | 2012-12-31 | 2016-11-22 | Omni Medsci, Inc. | Short-wave infrared super-continuum lasers for early detection of dental caries |
CA2895982A1 (en) | 2012-12-31 | 2014-07-03 | Omni Medsci, Inc. | Short-wave infrared super-continuum lasers for early detection of dental caries |
US10660526B2 (en) | 2012-12-31 | 2020-05-26 | Omni Medsci, Inc. | Near-infrared time-of-flight imaging using laser diodes with Bragg reflectors |
US9585563B2 (en) | 2012-12-31 | 2017-03-07 | Dexcom, Inc. | Remote monitoring of analyte measurements |
US9743987B2 (en) | 2013-03-14 | 2017-08-29 | Board Of Regents Of The University Of Nebraska | Methods, systems, and devices relating to robotic surgical devices, end effectors, and controllers |
US9888966B2 (en) | 2013-03-14 | 2018-02-13 | Board Of Regents Of The University Of Nebraska | Methods, systems, and devices relating to force control surgical systems |
US10667883B2 (en) | 2013-03-15 | 2020-06-02 | Virtual Incision Corporation | Robotic surgical devices, systems, and related methods |
US9498291B2 (en) | 2013-03-15 | 2016-11-22 | Hansen Medical, Inc. | Touch-free catheter user interface controller |
JP6479790B2 (en) | 2013-07-17 | 2019-03-06 | ボード オブ リージェンツ オブ ザ ユニバーシティ オブ ネブラスカ | Robotic surgical device, system and related methods |
US20150058031A1 (en) * | 2013-08-20 | 2015-02-26 | Kyle Samani | Methods and systems for a medical checklist |
US9364635B2 (en) * | 2013-09-20 | 2016-06-14 | Covidien Lp | Computer controlled steerable tip guide catheter |
CN103654967A (en) * | 2013-12-06 | 2014-03-26 | 傅松青 | Speech-controlled auxiliary imaging device for minimally invasive operations |
CN105992568B (en) | 2014-02-12 | 2018-06-08 | 皇家飞利浦有限公司 | The robot control of operation instrument visibility |
EP3110305B1 (en) | 2014-02-27 | 2021-09-01 | University Surgical Associates, Inc. | Interactive display for surgery |
US9293141B2 (en) * | 2014-03-27 | 2016-03-22 | Storz Endoskop Produktions Gmbh | Multi-user voice control system for medical devices |
US10932657B2 (en) | 2014-04-02 | 2021-03-02 | Transenterix Europe S.A.R.L. | Endoscope with wide angle lens and adjustable view |
WO2015151094A1 (en) * | 2014-04-02 | 2015-10-08 | M.S.T. Medical Surgery Technologies Ltd. | Endoscope with wide angle lens and adjustable view |
CA3193139A1 (en) | 2014-05-05 | 2015-11-12 | Vicarious Surgical Inc. | Virtual reality surgical device |
CA2961213A1 (en) | 2014-09-12 | 2016-03-17 | Board Of Regents Of The University Of Nebraska | Quick-release end effectors and related systems and methods |
EP4286104A3 (en) | 2014-11-11 | 2024-02-14 | Board of Regents of the University of Nebraska | Robotic device with compact joint design and related systems and methods |
DE102014223701B4 (en) | 2014-11-20 | 2019-03-07 | Siemens Healthcare Gmbh | Predictable mobility for a robotic device |
JP6961146B2 (en) | 2015-08-03 | 2021-11-05 | バーチャル インシジョン コーポレイションVirtual Incision Corporation | Robotic surgical devices, systems and related methods |
EP3130276B8 (en) | 2015-08-12 | 2020-02-26 | TransEnterix Europe Sàrl | Endoscope with wide angle lens and adjustable view |
EP3335661A4 (en) | 2015-10-09 | 2019-05-22 | Sony Corporation | Surgical control device, surgical control method, and program |
CA3200794A1 (en) | 2015-12-28 | 2017-07-06 | Dexcom, Inc. | Systems and methods for remote and host monitoring communications |
JP2017156511A (en) | 2016-03-01 | 2017-09-07 | ソニー株式会社 | Information processing device, information processing method, and program |
CN109069221B (en) * | 2016-04-28 | 2021-09-17 | 索尼公司 | Control device, control method, program, and audio output system |
JP7176757B2 (en) | 2016-05-18 | 2022-11-22 | バーチャル インシジョン コーポレイション | ROBOTIC SURGICAL DEVICES, SYSTEMS AND RELATED METHODS |
WO2018039606A1 (en) | 2016-08-25 | 2018-03-01 | Virtual Incision Corporation | Quick-release tool coupler and related systems and methods |
EP3507065A4 (en) | 2016-08-30 | 2020-04-29 | Board of Regents of the University of Nebraska | Robotic device with compact joint design and an additional degree of freedom and related systems and methods |
CN115337111A (en) | 2016-11-22 | 2022-11-15 | 内布拉斯加大学董事会 | Improved coarse positioning apparatus and related systems and methods |
CN110462259B (en) | 2016-11-29 | 2022-10-28 | 虚拟切割有限公司 | User controller with user presence detection and related systems and methods |
WO2018112199A1 (en) | 2016-12-14 | 2018-06-21 | Virtual Incision Corporation | Releasable attachment device for coupling to medical devices and related systems and methods |
US10028794B2 (en) * | 2016-12-19 | 2018-07-24 | Ethicon Llc | Surgical system with voice control |
DE102017101782A1 (en) * | 2017-01-30 | 2018-08-02 | Aktormed Gmbh | Operation assistance system and method for generating control signals for voice control of a motor-controlled movable robot kinematics of such an operation assistance system |
EP3579736A4 (en) | 2017-02-09 | 2020-12-23 | Vicarious Surgical Inc. | Virtual reality surgical tools system |
US11862302B2 (en) | 2017-04-24 | 2024-01-02 | Teladoc Health, Inc. | Automated transcription and documentation of tele-health encounters |
US10483007B2 (en) | 2017-07-25 | 2019-11-19 | Intouch Technologies, Inc. | Modular telehealth cart with thermal imaging and touch screen user interface |
US11636944B2 (en) | 2017-08-25 | 2023-04-25 | Teladoc Health, Inc. | Connectivity infrastructure for a telehealth platform |
CA3075692A1 (en) | 2017-09-14 | 2019-03-21 | Vicarious Surgical Inc. | Virtual reality surgical camera system |
JP7405432B2 (en) | 2017-09-27 | 2023-12-26 | バーチャル インシジョン コーポレイション | Robotic surgical device with tracking camera technology and related systems and methods |
EP3735341A4 (en) | 2018-01-05 | 2021-10-06 | Board of Regents of the University of Nebraska | Single-arm robotic device with compact joint design and related systems and methods |
WO2019163906A1 (en) * | 2018-02-21 | 2019-08-29 | オリンパス株式会社 | Medical system and operation method of medical system |
US10617299B2 (en) | 2018-04-27 | 2020-04-14 | Intouch Technologies, Inc. | Telehealth cart that supports a removable tablet with seamless audio/video switching |
US11051829B2 (en) | 2018-06-26 | 2021-07-06 | DePuy Synthes Products, Inc. | Customized patient-specific orthopaedic surgical instrument |
US10896675B1 (en) * | 2018-06-29 | 2021-01-19 | X Development Llc | Multi-tiered command processing |
US10565977B1 (en) | 2018-08-20 | 2020-02-18 | Verb Surgical Inc. | Surgical tool having integrated microphones |
JP2022028086A (en) * | 2018-12-05 | 2022-02-15 | ソニーグループ株式会社 | Information processor, presentation method and surgery system |
CA3125742A1 (en) | 2019-01-07 | 2020-07-16 | Virtual Incision Corporation | Robotically assisted surgical system and related devices and methods |
US11502633B2 (en) * | 2019-05-03 | 2022-11-15 | Canon U.S.A., Inc. | Motor position control based on servo-to-edge direction feedback |
IT201900023826A1 (en) * | 2019-12-12 | 2021-06-12 | Advanced Vision And Inspection Systems S R L Avais | Control system for an endoscopic device, related control method and device |
DE102021127877A1 (en) | 2021-10-26 | 2023-04-27 | Karl Storz Se & Co. Kg | Handling device for handling medical and/or surgical instruments, and medical system with a handling device |
Family Cites Families (192)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US977825A (en) | 1910-01-08 | 1910-12-06 | George N Murphy | Surgical instrument. |
GB955005A (en) | 1961-07-21 | 1964-04-08 | Molins Machine Co Ltd | Apparatus for gripping and lifting articles |
US3300053A (en) * | 1964-04-08 | 1967-01-24 | Melville F Peters | Fluid separating device |
US3280991A (en) | 1964-04-28 | 1966-10-25 | Programmed & Remote Syst Corp | Position control manipulator |
US5196688A (en) | 1975-02-04 | 1993-03-23 | Telefunken Systemtechnik Gmbh | Apparatus for recognizing and following a target |
GB1569450A (en) * | 1976-05-27 | 1980-06-18 | Nippon Electric Co | Speech recognition system |
US4128880A (en) | 1976-06-30 | 1978-12-05 | Cray Research, Inc. | Computer vector register processing |
US4058001A (en) | 1976-08-02 | 1977-11-15 | G. D. Searle & Co. | Ultrasound imaging system with improved scan conversion |
US4216462A (en) | 1978-03-06 | 1980-08-05 | General Electric Company | Patient monitoring and data processing system |
US4207959A (en) | 1978-06-02 | 1980-06-17 | New York University | Wheelchair mounted control apparatus |
US4221997A (en) | 1978-12-18 | 1980-09-09 | Western Electric Company, Incorporated | Articulated robot arm and method of moving same |
DE3045295A1 (en) | 1979-05-21 | 1982-02-18 | American Cystoscope Makers Inc | Surgical instrument for an endoscope |
US4367998A (en) | 1979-09-13 | 1983-01-11 | United Kingdom Atomic Energy Authority | Manipulators |
JPS5688506A (en) * | 1979-12-21 | 1981-07-18 | Matsushita Electric Ind Co Ltd | Heater |
JPS5688503A (en) * | 1979-12-21 | 1981-07-18 | Matsushita Electric Ind Co Ltd | Heater |
FR2482508A1 (en) | 1980-05-14 | 1981-11-20 | Commissariat Energie Atomique | MANIPULATOR AND MOTORIZED ORIENTATION BRACKET FOR SUCH A MANIPULATOR |
US4348553A (en) * | 1980-07-02 | 1982-09-07 | International Business Machines Corporation | Parallel pattern verifier with dynamic time warping |
US4605080A (en) * | 1980-07-11 | 1986-08-12 | Lemelson Jerome H | Speech recognition control system and method |
FR2492304A1 (en) | 1980-10-17 | 1982-04-23 | Commissariat Energie Atomique | TELEMANIPULATION ASSEMBLY MOUNTED ON A MOBILE PLATFORM AND COMPRISING A RETRACTABLE TELESCOPIC CARRIER ASSEMBLY WITHIN A SEALED HOOD, AND METHOD FOR SETTING UP ON AN ENCLOSURE |
JPS57118299A (en) | 1981-01-14 | 1982-07-23 | Nissan Motor | Voice load driver |
JPS5862699A (en) * | 1981-10-09 | 1983-04-14 | シャープ株式会社 | Voice recognition equipment |
US4454586A (en) * | 1981-11-19 | 1984-06-12 | At&T Bell Laboratories | Method and apparatus for generating speech pattern templates |
JPS58130393A (en) | 1982-01-29 | 1983-08-03 | 株式会社東芝 | Voice recognition equipment |
JPS58134357A (en) | 1982-02-03 | 1983-08-10 | Hitachi Ltd | Array processor |
US4456961A (en) | 1982-03-05 | 1984-06-26 | Texas Instruments Incorporated | Apparatus for teaching and transforming noncoincident coordinate systems |
US4491135A (en) | 1982-11-03 | 1985-01-01 | Klein Harvey A | Surgical needle holder |
US4517963A (en) | 1983-01-04 | 1985-05-21 | Harold Unger | Image-erecting barrel rotator for articulated optical arm |
US4624008A (en) * | 1983-03-09 | 1986-11-18 | International Telephone And Telegraph Corporation | Apparatus for automatic speech recognition |
CS235792B1 (en) | 1983-04-06 | 1985-05-15 | Pavel Zahalka | Connection for direct heating of ionic melt especially of glass by alternating current passage with lower frequency than 50 hz |
US4482032A (en) * | 1983-04-25 | 1984-11-13 | Westinghouse Electric Corp. | Elevator emergency control system |
US4503854A (en) | 1983-06-16 | 1985-03-12 | Jako Geza J | Laser surgery |
US4641292A (en) | 1983-06-20 | 1987-02-03 | George Tunnell | Voice controlled welding system |
US4799171A (en) * | 1983-06-20 | 1989-01-17 | Kenner Parker Toys Inc. | Talk back doll |
US4604016A (en) | 1983-08-03 | 1986-08-05 | Joyce Stephen A | Multi-dimensional force-torque hand controller having force feedback |
JPS6055985A (en) * | 1983-09-05 | 1985-04-01 | 株式会社トミー | Sound recognizing toy |
US4586398A (en) | 1983-09-29 | 1986-05-06 | Hamilton Industries | Foot control assembly for power-operated tables and the like |
US4635292A (en) | 1983-12-19 | 1987-01-06 | Matsushita Electric Industrial Co., Ltd. | Image processor |
US4616637A (en) | 1984-09-14 | 1986-10-14 | Precision Surgical Instruments, Inc. | Shoulder traction apparatus |
US4676243A (en) | 1984-10-31 | 1987-06-30 | Aldebaran Xiii Consulting Company | Automated anterior capsulectomy instrument |
JPH055529Y2 (en) | 1985-03-25 | 1993-02-15 | ||
DE3514286A1 (en) * | 1985-04-19 | 1986-10-23 | Siemens AG, 1000 Berlin und 8000 München | SYSTEM FOR DETECTING INDIVIDUALLY SPOKEN WORDS |
JPS61279491A (en) | 1985-05-31 | 1986-12-10 | 株式会社安川電機 | Visual apparatus holder |
US4672963A (en) | 1985-06-07 | 1987-06-16 | Israel Barken | Apparatus and method for computer controlled laser surgery |
US4945479A (en) | 1985-07-31 | 1990-07-31 | Unisys Corporation | Tightly coupled scientific processing system |
US4725956A (en) * | 1985-10-15 | 1988-02-16 | Lockheed Corporation | Voice command air vehicle control system |
US4797924A (en) * | 1985-10-25 | 1989-01-10 | Nartron Corporation | Vehicle voice recognition method and apparatus |
US4757541A (en) * | 1985-11-05 | 1988-07-12 | Research Triangle Institute | Audio visual speech recognition |
US4776016A (en) * | 1985-11-21 | 1988-10-04 | Position Orientation Systems, Inc. | Voice control system |
US4817050A (en) | 1985-11-22 | 1989-03-28 | Kabushiki Kaisha Toshiba | Database system |
US4750136A (en) | 1986-01-10 | 1988-06-07 | American Telephone And Telegraph, At&T Information Systems Inc. | Communication system having automatic circuit board initialization capability |
JPH085018B2 (en) | 1986-02-26 | 1996-01-24 | 株式会社日立製作所 | Remote manipulation method and apparatus |
EP0239409A1 (en) | 1986-03-28 | 1987-09-30 | Life Technology Research Foundation | Robot for surgical operation |
US5078140A (en) | 1986-05-08 | 1992-01-07 | Kwoh Yik S | Imaging device - aided robotic stereotaxis system |
US4791934A (en) | 1986-08-07 | 1988-12-20 | Picker International, Inc. | Computer tomography assisted stereotactic surgery system and method |
SE464855B (en) | 1986-09-29 | 1991-06-24 | Asea Ab | PROCEDURE OF AN INDUSTRIAL BOTTOM FOR CALIBRATION OF A SENSOR |
DE3636678A1 (en) | 1986-10-28 | 1988-05-11 | Siemens Ag | X-RAY DIAGNOSTIC DEVICE |
US5157603A (en) | 1986-11-06 | 1992-10-20 | Storz Instrument Company | Control system for ophthalmic surgical instruments |
US4854301A (en) | 1986-11-13 | 1989-08-08 | Olympus Optical Co., Ltd. | Endoscope apparatus having a chair with a switch |
JPH0829509B2 (en) | 1986-12-12 | 1996-03-27 | 株式会社日立製作所 | Control device for manipulator |
US4791940A (en) | 1987-02-02 | 1988-12-20 | Florida Probe Corporation | Electronic periodontal probe with a constant force applier |
DE3889681T2 (en) | 1987-02-09 | 1994-09-08 | Sumitomo Electric Industries | Device for bending an elongated body. |
US4805219A (en) * | 1987-04-03 | 1989-02-14 | Dragon Systems, Inc. | Method for speech recognition |
US4860215A (en) | 1987-04-06 | 1989-08-22 | California Institute Of Technology | Method and apparatus for adaptive force and position control of manipulators |
US5065741A (en) | 1987-04-16 | 1991-11-19 | Olympus Optical Co. Ltd. | Extracoporeal ultrasonic lithotripter with a variable focus |
US4863133A (en) | 1987-05-26 | 1989-09-05 | Leonard Medical | Arm device for adjustable positioning of a medical instrument or the like |
US4762455A (en) | 1987-06-01 | 1988-08-09 | Remote Technology Corporation | Remote manipulator |
US4852083A (en) | 1987-06-22 | 1989-07-25 | Texas Instruments Incorporated | Digital crossbar switch |
JPH088933B2 (en) | 1987-07-10 | 1996-01-31 | 日本ゼオン株式会社 | Catheter |
US4794912A (en) | 1987-08-17 | 1989-01-03 | Welch Allyn, Inc. | Borescope or endoscope with fluid dynamic muscle |
JP2602240B2 (en) | 1987-08-28 | 1997-04-23 | 株式会社日立製作所 | Multi-processor system |
US4991579A (en) | 1987-11-10 | 1991-02-12 | Allen George S | Method and apparatus for providing related images over time of a portion of the anatomy using fiducial implants |
US5303148A (en) | 1987-11-27 | 1994-04-12 | Picker International, Inc. | Voice actuated volume image controller and display controller |
EP0326768A3 (en) | 1988-02-01 | 1991-01-23 | Faro Medical Technologies Inc. | Computer-aided surgery apparatus |
US5251127A (en) | 1988-02-01 | 1993-10-05 | Faro Medical Technologies Inc. | Computer-aided surgery apparatus |
US4815450A (en) | 1988-02-01 | 1989-03-28 | Patel Jayendra I | Endoscope having variable flexibility |
US5218969A (en) * | 1988-02-04 | 1993-06-15 | Blood Line Technology, Inc. | Intelligent stethoscope |
US4964062A (en) | 1988-02-16 | 1990-10-16 | Ubhayakar Shivadev K | Robotic arm systems |
US4930494A (en) | 1988-03-09 | 1990-06-05 | Olympus Optical Co., Ltd. | Apparatus for bending an insertion section of an endoscope using a shape memory alloy |
US4949717A (en) | 1988-03-17 | 1990-08-21 | Shaw Edward L | Surgical instrument with suture cutter |
US5019968A (en) | 1988-03-29 | 1991-05-28 | Yulan Wang | Three-dimensional vector processor |
US4989253A (en) | 1988-04-15 | 1991-01-29 | The Montefiore Hospital Association Of Western Pennsylvania | Voice activated microscope |
US4979949A (en) * | 1988-04-26 | 1990-12-25 | The Board Of Regents Of The University Of Washington | Robot-aided system for surgery |
US4979933A (en) | 1988-04-27 | 1990-12-25 | Kraft, Inc. | Reclosable bag |
US5142484A (en) | 1988-05-12 | 1992-08-25 | Health Tech Services Corporation | An interactive patient assistance device for storing and dispensing prescribed medication and physical device |
DE3819075A1 (en) * | 1988-06-04 | 1989-12-07 | Sartorius Gmbh | ELECTRONIC SCALE FOR DOSING |
US4883400A (en) | 1988-08-24 | 1989-11-28 | Martin Marietta Energy Systems, Inc. | Dual arm master controller for a bilateral servo-manipulator |
JPH079606B2 (en) | 1988-09-19 | 1995-02-01 | 豊田工機株式会社 | Robot controller |
CA2000818C (en) | 1988-10-19 | 1994-02-01 | Akira Tsuchihashi | Master slave manipulator system |
US5123095A (en) | 1989-01-17 | 1992-06-16 | Ergo Computing, Inc. | Integrated scalar and vector processors with vector addressing by the scalar processor |
US5098426A (en) | 1989-02-06 | 1992-03-24 | Phoenix Laser Systems, Inc. | Method and apparatus for precision laser surgery |
US4965417A (en) | 1989-03-27 | 1990-10-23 | Massie Philip E | Foot-operated control |
JPH034831A (en) | 1989-06-01 | 1991-01-10 | Toshiba Corp | Endoscope device |
US4980626A (en) | 1989-08-10 | 1990-12-25 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Method and apparatus for positioning a robotic end effector |
US5271384A (en) | 1989-09-01 | 1993-12-21 | Mcewen James A | Powered surgical retractor |
US5201325A (en) * | 1989-09-01 | 1993-04-13 | Andronic Devices Ltd. | Advanced surgical retractor |
US5182557A (en) * | 1989-09-20 | 1993-01-26 | Semborg Recrob, Corp. | Motorized joystick |
FR2652928B1 (en) * | 1989-10-05 | 1994-07-29 | Diadix Sa | INTERACTIVE LOCAL INTERVENTION SYSTEM WITHIN A AREA OF A NON-HOMOGENEOUS STRUCTURE. |
US5181823A (en) | 1989-10-27 | 1993-01-26 | Grumman Aerospace Corporation | Apparatus and method for producing a video display |
US5091656A (en) | 1989-10-27 | 1992-02-25 | Storz Instrument Company | Footswitch assembly with electrically engaged detents |
US5249121A (en) | 1989-10-27 | 1993-09-28 | American Cyanamid Company | Remote control console for surgical control system |
EP0427358B1 (en) | 1989-11-08 | 1996-03-27 | George S. Allen | Mechanical arm for and interactive image-guided surgical system |
DE4102196C2 (en) | 1990-01-26 | 2002-08-01 | Olympus Optical Co | Imaging device for tracking an object |
JP2964518B2 (en) | 1990-01-30 | 1999-10-18 | 日本電気株式会社 | Voice control method |
US5175694A (en) | 1990-02-08 | 1992-12-29 | The United States Of America As Represented By The Secretary Of The Navy | Centroid target tracking system utilizing parallel processing of digital data patterns |
US5097829A (en) | 1990-03-19 | 1992-03-24 | Tony Quisenberry | Temperature controlled cooling system |
US5343391A (en) * | 1990-04-10 | 1994-08-30 | Mushabac David R | Device for obtaining three dimensional contour data and for operating on a patient and related method |
FR2660852A1 (en) * | 1990-04-17 | 1991-10-18 | Cheval Freres Sa | LASER BEAM DENTAL INSTRUMENT. |
DE59006779D1 (en) * | 1990-05-09 | 1994-09-15 | Siemens Ag | Medical, in particular dental, facility. |
US5431645A (en) | 1990-05-10 | 1995-07-11 | Symbiosis Corporation | Remotely activated endoscopic tools such as endoscopic biopsy forceps |
US5086401A (en) | 1990-05-11 | 1992-02-04 | International Business Machines Corporation | Image-directed robotic system for precise robotic surgery including redundant consistency checking |
JPH0771288B2 (en) | 1990-08-24 | 1995-07-31 | 神田通信工業株式会社 | Automatic view adjustment method and device |
JPH04157889A (en) | 1990-10-20 | 1992-05-29 | Fujitsu Ltd | Automatic adjusting method for person image pickup position |
US5131105A (en) | 1990-11-21 | 1992-07-21 | Diasonics, Inc. | Patient support table |
US5145227A (en) | 1990-12-31 | 1992-09-08 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Electromagnetic attachment mechanism |
US5228429A (en) | 1991-01-14 | 1993-07-20 | Tadashi Hatano | Position measuring device for endoscope |
US5217453A (en) | 1991-03-18 | 1993-06-08 | Wilk Peter J | Automated surgical system and apparatus |
US5217003A (en) | 1991-03-18 | 1993-06-08 | Wilk Peter J | Automated surgical system and apparatus |
US5339799A (en) | 1991-04-23 | 1994-08-23 | Olympus Optical Co., Ltd. | Medical system for reproducing a state of contact of the treatment section in the operation unit |
US5166513A (en) | 1991-05-06 | 1992-11-24 | Coherent, Inc. | Dual actuation photoelectric foot switch |
US5313306A (en) | 1991-05-13 | 1994-05-17 | Telerobotics International, Inc. | Omniview motionless camera endoscopy system |
JP3173042B2 (en) | 1991-05-21 | 2001-06-04 | ソニー株式会社 | Robot numerical controller |
US5279309A (en) * | 1991-06-13 | 1994-01-18 | International Business Machines Corporation | Signaling device and method for monitoring positions in a surgical operation |
US5417210A (en) | 1992-05-27 | 1995-05-23 | International Business Machines Corporation | System and method for augmentation of endoscopic surgery |
US5182641A (en) | 1991-06-17 | 1993-01-26 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration | Composite video and graphics display for camera viewing systems in robotics and teleoperation |
US5735290A (en) | 1993-02-22 | 1998-04-07 | Heartport, Inc. | Methods and systems for performing thoracoscopic coronary bypass and other procedures |
US5458574A (en) | 1994-03-16 | 1995-10-17 | Heartport, Inc. | System for performing a cardiac procedure |
US5184601A (en) | 1991-08-05 | 1993-02-09 | Putman John M | Endoscope stabilizer |
DE4130631A1 (en) * | 1991-09-14 | 1993-03-18 | Philips Patentverwaltung | METHOD FOR RECOGNIZING THE SPOKEN WORDS IN A VOICE SIGNAL |
US5230623A (en) * | 1991-12-10 | 1993-07-27 | Radionics, Inc. | Operating pointer with interactive computergraphics |
US5289365A (en) | 1991-12-23 | 1994-02-22 | Donnelly Corporation | Modular network control system |
JP3583777B2 (en) | 1992-01-21 | 2004-11-04 | エス・アール・アイ・インターナシヨナル | Teleoperator system and telepresence method |
US6963792B1 (en) * | 1992-01-21 | 2005-11-08 | Sri International | Surgical method |
US5631973A (en) | 1994-05-05 | 1997-05-20 | Sri International | Method for telemanipulation with telepresence |
US5357962A (en) | 1992-01-27 | 1994-10-25 | Sri International | Ultrasonic imaging system and method wtih focusing correction |
US5345538A (en) | 1992-01-27 | 1994-09-06 | Krishna Narayannan | Voice activated control apparatus |
US5626595A (en) | 1992-02-14 | 1997-05-06 | Automated Medical Instruments, Inc. | Automated surgical instrument |
US5201743A (en) | 1992-05-05 | 1993-04-13 | Habley Medical Technology Corp. | Axially extendable endoscopic surgical instrument |
US5282806A (en) | 1992-08-21 | 1994-02-01 | Habley Medical Technology Corporation | Endoscopic surgical instrument having a removable, rotatable, end effector assembly |
DE9204118U1 (en) | 1992-03-27 | 1992-05-21 | Fa. Carl Zeiss, 7920 Heidenheim, De | |
JP3199130B2 (en) | 1992-03-31 | 2001-08-13 | パイオニア株式会社 | 3D coordinate input device |
US5305244B2 (en) * | 1992-04-06 | 1997-09-23 | Computer Products & Services I | Hands-free user-supported portable computer |
US5221283A (en) | 1992-05-15 | 1993-06-22 | General Electric Company | Apparatus and method for stereotactic surgery |
US5274862A (en) | 1992-05-18 | 1994-01-04 | Palmer Jr John M | Patient turning device and method for lateral traveling transfer system |
US5257999A (en) | 1992-06-04 | 1993-11-02 | Slanetz Jr Charles A | Self-oriented laparoscopic needle holder for curved needles |
US5758021A (en) * | 1992-06-12 | 1998-05-26 | Alcatel N.V. | Speech recognition combining dynamic programming and neural network techniques |
US5443484A (en) | 1992-06-16 | 1995-08-22 | Loma Linda University Medical Center | Trocar and method for endoscopic surgery |
GR930100244A (en) | 1992-06-30 | 1994-02-28 | Ethicon Inc | Flexible endoscopic surgical port |
AU4771893A (en) | 1992-07-14 | 1994-01-31 | Sierra Matrix, Inc. | Hands-free ultrasonic test view (hf-utv) |
US5458547A (en) | 1992-07-17 | 1995-10-17 | Tochigifujisangyo Kabushiki Kaisha | Differential apparatus with speed and torque sensitive differential limiting forces |
US5754741A (en) | 1992-08-10 | 1998-05-19 | Computer Motion, Inc. | Automated endoscope for optimal positioning |
US5515478A (en) | 1992-08-10 | 1996-05-07 | Computer Motion, Inc. | Automated endoscope system for optimal positioning |
US5657429A (en) | 1992-08-10 | 1997-08-12 | Computer Motion, Inc. | Automated endoscope system optimal positioning |
US5609560A (en) | 1992-08-19 | 1997-03-11 | Olympus Optical Co., Ltd. | Medical operation device control system for controlling a operation devices accessed respectively by ID codes |
US5337732A (en) | 1992-09-16 | 1994-08-16 | Cedars-Sinai Medical Center | Robotic endoscopy |
US5397323A (en) | 1992-10-30 | 1995-03-14 | International Business Machines Corporation | Remote center-of-motion robot for surgery |
US5304185A (en) | 1992-11-04 | 1994-04-19 | Unisurge, Inc. | Needle holder |
US5629594A (en) | 1992-12-02 | 1997-05-13 | Cybernet Systems Corporation | Force feedback system |
US5451924A (en) | 1993-01-14 | 1995-09-19 | Massachusetts Institute Of Technology | Apparatus for providing sensory substitution of force feedback |
US5303882A (en) * | 1993-02-22 | 1994-04-19 | The United States Of America As Represented By The Secretary Of The Navy | Corner vortex suppressor |
US5320630A (en) * | 1993-02-23 | 1994-06-14 | Munir Ahmed | Endoscopic ligating instrument for applying elastic bands |
DE4306786C1 (en) | 1993-03-04 | 1994-02-10 | Wolfgang Daum | Hand-type surgical manipulator for areas hard to reach - has distal components actuated by fingers via Bowden cables |
US5309717A (en) | 1993-03-22 | 1994-05-10 | Minch Richard B | Rapid shape memory effect micro-actuators |
JP3477781B2 (en) | 1993-03-23 | 2003-12-10 | セイコーエプソン株式会社 | IC card |
US5417701A (en) | 1993-03-30 | 1995-05-23 | Holmed Corporation | Surgical instrument with magnetic needle holder |
DE4310842C2 (en) | 1993-04-02 | 1996-01-25 | Viktor Dr Med Grablowitz | Device for performing minimally invasive operations |
US5410638A (en) | 1993-05-03 | 1995-04-25 | Northwestern University | System for positioning a medical instrument within a biotic structure using a micromanipulator |
JP2665052B2 (en) | 1993-05-14 | 1997-10-22 | エスアールアイ インターナショナル | Remote center positioning device |
US5395369A (en) | 1993-06-10 | 1995-03-07 | Symbiosis Corporation | Endoscopic bipolar electrocautery instruments |
AU7468494A (en) * | 1993-07-07 | 1995-02-06 | Cornelius Borst | Robotic system for close inspection and remote treatment of moving parts |
WO1995002426A1 (en) | 1993-07-13 | 1995-01-26 | Sims Deltec, Inc. | Medical pump and method of programming |
US5434457A (en) * | 1993-07-30 | 1995-07-18 | Josephs; Harold | Foot pedal safety switch and safety circuit |
CA2103626A1 (en) | 1993-08-09 | 1995-02-10 | Septimiu Edmund Salcudean | Motion scaling tele-operating system with force feedback suitable for microsurgery |
US5343385A (en) | 1993-08-17 | 1994-08-30 | International Business Machines Corporation | Interference-free insertion of a solid body into a cavity |
US5776126A (en) | 1993-09-23 | 1998-07-07 | Wilk; Peter J. | Laparoscopic surgical apparatus and associated method |
US5779623A (en) | 1993-10-08 | 1998-07-14 | Leonard Medical, Inc. | Positioner for medical instruments |
US5422521A (en) | 1993-11-18 | 1995-06-06 | Liebel-Flarsheim Co. | Foot operated control system for a multi-function device |
WO1995016396A1 (en) | 1993-12-15 | 1995-06-22 | Computer Motion, Inc. | Automated endoscope system for optimal positioning |
US5436542A (en) * | 1994-01-28 | 1995-07-25 | Surgix, Inc. | Telescopic camera mount with remotely controlled positioning |
US5645077A (en) | 1994-06-16 | 1997-07-08 | Massachusetts Institute Of Technology | Inertial orientation tracker apparatus having automatic drift compensation for tracking human head and other similarly sized body |
US5803089A (en) | 1994-09-15 | 1998-09-08 | Visualization Technology, Inc. | Position tracking and imaging system for use in medical applications |
US7053752B2 (en) * | 1996-08-06 | 2006-05-30 | Intuitive Surgical | General purpose distributed operating room control system |
US6646541B1 (en) * | 1996-06-24 | 2003-11-11 | Computer Motion, Inc. | General purpose distributed operating room control system |
US6463361B1 (en) * | 1994-09-22 | 2002-10-08 | Computer Motion, Inc. | Speech interface for an automated endoscopic system |
JP2996882B2 (en) * | 1994-10-20 | 2000-01-11 | 東芝機械株式会社 | Thermal displacement T-die |
US5544654A (en) * | 1995-06-06 | 1996-08-13 | Acuson Corporation | Voice control of a medical ultrasound scanning machine |
JPH0931483A (en) * | 1995-07-20 | 1997-02-04 | Tonen Corp | Lubricant composition |
US5715823A (en) * | 1996-02-27 | 1998-02-10 | Atlantis Diagnostics International, L.L.C. | Ultrasonic diagnostic imaging system with universal access to diagnostic information and images |
US5797900A (en) * | 1996-05-20 | 1998-08-25 | Intuitive Surgical, Inc. | Wrist mechanism for surgical instrument for performing minimally invasive surgery with enhanced dexterity and sensitivity |
US6911916B1 (en) * | 1996-06-24 | 2005-06-28 | The Cleveland Clinic Foundation | Method and apparatus for accessing medical data over a network |
US6496099B2 (en) * | 1996-06-24 | 2002-12-17 | Computer Motion, Inc. | General purpose distributed operating room control system |
US20020140665A1 (en) * | 2001-03-27 | 2002-10-03 | Gary Gordon | Method for framing viewports on a computer screen, and for pointing therein |
JP4150013B2 (en) * | 2005-03-31 | 2008-09-17 | Tdk株式会社 | Tunnel effect element |
-
1994
- 1994-09-22 US US08/310,665 patent/US6463361B1/en not_active Expired - Lifetime
-
1995
- 1995-09-21 CA CA002200716A patent/CA2200716C/en not_active Expired - Lifetime
- 1995-09-21 AU AU37211/95A patent/AU3721195A/en not_active Abandoned
- 1995-09-21 WO PCT/US1995/012064 patent/WO1996009587A1/en active Application Filing
-
2002
- 2002-03-11 US US10/095,488 patent/US6965812B2/en not_active Expired - Fee Related
-
2004
- 2004-09-16 US US10/942,374 patent/US7395249B2/en not_active Expired - Lifetime
Also Published As
Publication number | Publication date |
---|---|
US20050033580A1 (en) | 2005-02-10 |
US7395249B2 (en) | 2008-07-01 |
US6965812B2 (en) | 2005-11-15 |
US6463361B1 (en) | 2002-10-08 |
US20020183894A1 (en) | 2002-12-05 |
AU3721195A (en) | 1996-04-09 |
CA2200716A1 (en) | 1996-03-28 |
WO1996009587A1 (en) | 1996-03-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2200716C (en) | A speech interface for an automated endoscopic system | |
US5515478A (en) | Automated endoscope system for optimal positioning | |
US5754741A (en) | Automated endoscope for optimal positioning | |
US5841950A (en) | Automated endoscope system for optimal positioning | |
KR102345782B1 (en) | Surgical assistance device, control method therefor, and recording medium | |
Berkelman et al. | A compact, compliant laparoscopic endoscope manipulator | |
US6197017B1 (en) | Articulated apparatus for telemanipulator system | |
JP3540362B2 (en) | Surgical manipulator control system and control method | |
US6692485B1 (en) | Articulated apparatus for telemanipulator system | |
US6120433A (en) | Surgical manipulator system | |
Madhani | Design of teleoperated surgical instruments for minimally invasive surgery | |
US7074179B2 (en) | Method and apparatus for performing minimally invasive cardiac procedures | |
US20020156345A1 (en) | Method of guiding an endoscope for performing minimally invasive surgery | |
WO2002051329A1 (en) | Tendon actuated articulated members for a telemanipulator system | |
Kumar | An augmented steady hand system for precise micromanipulation | |
JP6546361B1 (en) | Surgery support device | |
JP6562174B1 (en) | Surgery support device | |
Dachs et al. | A novel surgical robot design: Minimizing the operating envelope within the sterile field | |
Ko et al. | Compact laparoscopic assistant robot using a bending mechanism | |
CN114327126A (en) | Curved surface three-dimensional touch operation device and operation strategy | |
Shin et al. | Development of a 5-DOF laparoscopic assistant robot | |
Neisius et al. | Robotic telemanipulator for laparoscopy | |
US20230256607A1 (en) | Robot system and control method thereof | |
US20220401088A1 (en) | Method for controlling an articulating instrument | |
Cuschieri et al. | Dumass—a surgical assist robot for minimal access surgery |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request | ||
MKEX | Expiry |
Effective date: 20150921 |