WO2009037576A2 - Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis - Google Patents
Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis Download PDFInfo
- Publication number
- WO2009037576A2 WO2009037576A2 PCT/IB2008/003323 IB2008003323W WO2009037576A2 WO 2009037576 A2 WO2009037576 A2 WO 2009037576A2 IB 2008003323 W IB2008003323 W IB 2008003323W WO 2009037576 A2 WO2009037576 A2 WO 2009037576A2
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- tool
- movement
- input device
- single axis
- move
- Prior art date
Links
- 230000033001 locomotion Effects 0.000 title claims abstract description 97
- 238000000034 method Methods 0.000 title abstract description 66
- 230000006870 function Effects 0.000 claims description 30
- 238000004088 simulation Methods 0.000 claims description 24
- 230000004044 response Effects 0.000 claims description 16
- 238000001356 surgical procedure Methods 0.000 claims description 12
- 230000000694 effects Effects 0.000 claims description 3
- 238000001574 biopsy Methods 0.000 abstract description 19
- 210000004556 brain Anatomy 0.000 abstract description 6
- 238000002513 implantation Methods 0.000 abstract description 4
- 210000000481 breast Anatomy 0.000 abstract description 3
- 238000012549 training Methods 0.000 description 11
- 238000002595 magnetic resonance imaging Methods 0.000 description 9
- 210000001519 tissue Anatomy 0.000 description 7
- 238000002406 microsurgery Methods 0.000 description 5
- 230000005540 biological transmission Effects 0.000 description 3
- 230000008859 change Effects 0.000 description 3
- 230000000881 depressing effect Effects 0.000 description 3
- 210000003811 finger Anatomy 0.000 description 3
- 238000003384 imaging method Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 230000003213 activating effect Effects 0.000 description 2
- 230000004913 activation Effects 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000011161 development Methods 0.000 description 2
- 238000003780 insertion Methods 0.000 description 2
- 230000037431 insertion Effects 0.000 description 2
- 210000003813 thumb Anatomy 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 206010028980 Neoplasm Diseases 0.000 description 1
- 208000008558 Osteophyte Diseases 0.000 description 1
- 238000007792 addition Methods 0.000 description 1
- 210000005013 brain tissue Anatomy 0.000 description 1
- 238000013500 data storage Methods 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 238000005553 drilling Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 229940079593 drug Drugs 0.000 description 1
- 239000012636 effector Substances 0.000 description 1
- 206010015037 epilepsy Diseases 0.000 description 1
- 210000005224 forefinger Anatomy 0.000 description 1
- 230000036571 hydration Effects 0.000 description 1
- 238000006703 hydration reaction Methods 0.000 description 1
- 239000000463 material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000008707 rearrangement Effects 0.000 description 1
- 210000000130 stem cell Anatomy 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1656—Programme controls characterised by programming, planning systems for manipulators
- B25J9/1664—Programme controls characterised by programming, planning systems for manipulators characterised by motion, path, trajectory planning
- B25J9/1666—Avoiding collision or forbidden zones
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/30—Surgical robots
- A61B34/37—Leader-follower robots
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J3/00—Manipulators of leader-follower type, i.e. both controlling unit and controlled unit perform corresponding spatial movements
- B25J3/04—Manipulators of leader-follower type, i.e. both controlling unit and controlled unit perform corresponding spatial movements involving servo mechanisms
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1656—Programme controls characterised by programming, planning systems for manipulators
- B25J9/1671—Programme controls characterised by programming, planning systems for manipulators characterised by simulation, either to verify existing program or to create and verify new program, CAD/CAM oriented, graphic oriented programming systems
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/10—Computer-aided planning, simulation or modelling of surgical operations
- A61B2034/107—Visualisation of planned trajectories or target regions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B34/00—Computer-aided surgery; Manipulators or robots specially adapted for use in surgery
- A61B34/25—User interfaces for surgical systems
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40128—Virtual tether, line on display connects end effector to destination point
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40399—Selection of master-slave operation mode
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/40—Robotics, robotics mapping to robotics vision
- G05B2219/40418—Presurgical planning, on screen indicate regions to be operated on
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/45—Nc applications
- G05B2219/45117—Medical, radio surgery manipulator
Definitions
- the present methods, devices, and systems relate generally to the field of surgical robotics, and more particularly to the non-mechanical restriction of a manipulator (e.g., a robotic arm with multiple degrees of freedom) to movement of a manipulator
- An example of a procedure that can be carried out according to the present methods, devices, and systems is an automated biopsy.
- An example of a surgical robot that can be used in a procedure to which the present methods, devices, and systems relate is disclosed in U.S. Patent No. 7,155,316 (the '"316 patent"), which is incorporated by reference.
- the pre-planned path is derived from an image taken hours before the procedure.
- the brain is not fixed within the cranial cavity and can shift as a result of damage, tumours, hydration, and body position changes. These relatively small brain shifts can be problematic in term of accuracy and pose a safety concern.
- post surgical images and other tools are used to ensure accurate and safe procedures with existing tools.
- attachment of a head frame to the patient's head is also required; this is both uncomfortable and time consuming.
- Embodiments of the present methods and systems enable a user, such as a surgeon, to set up and execute an automated move of a tool of one of the robotic arms (which includes a tool that is coupled to the robotic arm, as well as a tool that is integrated with the robotic arm) along a single axis, such as the longitudinal axis of the tool.
- a move may be particularly advantageous when implemented as an automated biopsy of tissue, such as brain or breast tissue.
- the automated move may be programmed to occur during a stereotactic procedure, when some or all of the robotic arm is positioned within the bore of an open or closed magnet of a magnetic resonance imaging machine, or during a microsurgical procedure during which one or both robotic arms may be set up to and execute such an automated move.
- Robots that may be manipulated according to the present techniques may be characterized as computer-assisted devices.
- the present systems take the form of a computer system useful in simulating, planning and/or executing an automated surgical procedure.
- the computer system is configured to perform at least the following functions: receive data designating a target location for a tool held by a medical robot; receive data designating a second location for the tool from which the tool will move toward the target location during an automated movement; and move the medical robot in response to a user command to begin the automated movement such that the tool moves along a single axis defined by the second location and the target location.
- the data designating the target location may comprise coordinates (e.g., Cartesian coordinates) of the tip of the tool, or coordinates of a location spaced away from the tool along a longitudinal axis of the tool, in any suitable coordinate system, or data sufficient to enable determination of such coordinates (such as joint values of the robotic arm that allow forward kinematics to be used to solve for the coordinates based on known parameters such as robotic arm link lengths).
- coordinates e.g., Cartesian coordinates
- coordinates of a location spaced away from the tool along a longitudinal axis of the tool in any suitable coordinate system, or data sufficient to enable determination of such coordinates (such as joint values of the robotic arm that allow forward kinematics to be used to solve for the coordinates based on known parameters such as robotic arm link lengths).
- the present devices take the form of a computer readable medium comprising machine readable instructions for receiving a command to restrict movement of an instrument held by or integral with a robotic arm along a single axis, the robotic arm being configured for use in surgery; receiving a position and orientation of an input device, the input device being linked to the robotic arm through a master-slave relationship in which the input device is the master, the difference between the position and orientation of the input device and a previous position and orientation of the input device corresponding to a desired movement of the instrument; and sending a signal or signals to effect a move of the instrument in accordance with the desired movement, where the move will be along the single axis and will not include any movement along any different axis from the single axis.
- the signal or signals may be any suitable form of data that includes information sufficient to cause the robotic arm to move appropriately.
- the signal or signals could represent a set of joint displacements and/or joint velocities outputted to a local controller for the robotic arm or directly to the individual joint actuators.
- the user may set up a procedure by delivering inputs to a computer system through an input device, such as a hand controller that is linked as a master to the robotic arm in a master-slave relationship.
- an input device such as a hand controller that is linked as a master to the robotic arm in a master-slave relationship.
- the user may also deliver inputs through one or more graphical user interfaces (GUIs) using any suitable input device, such as touch screen controls (e.g., buttons, slider bars, drop down menus, tabs, etc.), a mouse, or the like.
- GUIs graphical user interfaces
- Some embodiments of the present systems are computer systems that may be configured to display on a display screen a GUI that allows the user to select a simulation mode (e.g., through a control such as a button that can be selected via a touch, a mouse, or the like) for setting up the automated movement and otherwise for training.
- the computer system also may be configured to display on the GUI one or more controls (e.g., that can be selected via a touch, a mouse, or the like) for selecting the type of surgery, such as microsurgery, stereotaxy with one of the robotic arms, or stereotaxy with the other robotic arm.
- the computer system also may be configured to display on the GUI one or more controls (e.g., that can be selected via a touch, a mouse, or the like) for activating power to: the robotic arms (e.g., through separate buttons); a base motor for adjusting the height of the base on which the robotic arms sit during microsurgical procedures; a digitizing arm usable during the physical registration process for registering a structure (e.g., of a radio- frequency coil assembly) associate in a fixed relationship with a portion of a subject to one or both robotic arms; a field camera usable during microsurgery to capture images of the surgical field; and a bore camera or cameras to be positioned in the bore of a magnet of a magnetic resonance imaging machine.
- the robotic arms e.g., through separate buttons
- a base motor for adjusting the height of the base on which the robotic arms sit during microsurgical procedures
- a digitizing arm usable during the physical registration process for registering a structure (e.g., of a radio- frequency coil assembly) associate
- the computer system also may be configured to display on the GUI one or more controls (e.g., that can be selected via a touch, a mouse, or the like) for activating a single axis lock (e.g., a Z- axis lock) and another button or buttons for controlling which robotic arm to associate the single axis lock with.
- a single axis lock e.g., a Z- axis lock
- another button or buttons for controlling which robotic arm to associate the single axis lock with.
- the computer system also may be configured to display on one or more additional display screens one or more additional GUIs for displaying two- dimensional images (one at a time) of a portion of a subject and for displaying a three- dimensional representation (e.g., a set of 2D images that form a 3D dataset of images representing a volume) of a portion of a subject.
- a three- dimensional representation e.g., a set of 2D images that form a 3D dataset of images representing a volume
- the computer system may be configured to display one or more controls (e.g., buttons, tabs, or the like that can be selected via a touch, a mouse, or the like) that a user can select to display either 2D images (one at a time) or a 3D image.
- the computer system also may be configured to display a zoom button, slider bar, or the like (e.g., that can be selected/manipulated via a touch, a mouse, or the like) that will allow a user that has selected the 2D display to zoom in on a given 2D image, where the 2D image remains centered as it is enlarged or reduced in size.
- a zoom button e.g., that can be selected/manipulated via a touch, a mouse, or the like
- the computer system also may be configured to display controls (e.g., that can be selected via a touch, a mouse, or the like) that allow a user to turn on a tracking feature for one of the two robotic arms that will be displayed as crosshairs representative of the location of either (a) the working tip (e.g., the distal tip) of a tool of the robotic arm selected or (b) the end of a line that extends from the tool tip, and further may be configured to display controls (e.g., buttons, slider bars, or the like that can be selected via a touch, a mouse, or the like) that allow a user to activate the display of the extension line and control the length of the extension line.
- controls e.g., that can be selected via a touch, a mouse, or the like
- a given 2D image may comprise an oblique slice that is oriented perpendicular to the tool axis. Such slices interpolate pixels between the 2D slices to achieve off-axis images.
- the 3D image also changes in response by 2D slices that make up the 3D image being taken away or added depending on the depth of the tool/extension line into the subject.
- the computer system may also be configured to display, when either the 2D or 3D display is selected, a section corresponding to planning for an automated biopsy that includes a display of controls (e.g., that can be selected via a touch, a mouse, or the like) that can be used to set a target location (associated with a location of the crosshairs at the time when the target location button is selected) for an automated movement along a single axis (e.g., an automated biopsy); a second point (characterizable as a start point, though a given movement may not begin exactly at the start point; the second point being associated with a location of the crosshairs at the time when the second button is selected) that together with the target location defines a path for the tool movement; a tool alignment function, that can be used when a user desires to position the relevant robotic arm in place (e.g., within a preset distance, ranging from zero to some relatively small distance (e.g., 2 centimeters)) for the automated move procedure, and that when pressed will
- the computer system may also be configured to display an indicator (e.g., a colored circle) for the target location selected by the user; a line extending from the indicator and to the tool tip or extension line tip (whichever is used) following selection of the target location, the line being designed to show the user the path through the subject if the line is followed, the computer system being configured to alter the appearance of the line when a second point is selected (e.g., changing the line's color or shape).
- an indicator e.g., a colored circle
- the computer system may be configured to perform at least the following functions: receive a command (e.g., through a user's touch of the screen displaying the relevant GUI) identifying a target location for a tool used in an automated movement by a robotic arm; receive a command identifying a starting location for the tool; receive a command to execute an automated move along a path (e.g., a line) defined at least in part by the starting location and the target location; and execute the automated move such that the tool, which may have a longitudinal axis, travels along the path (e.g., along a single axis). That path also may be aligned with the tools longitudinal axis.
- a command e.g., through a user's touch of the screen displaying the relevant GUI
- receive a command identifying a starting location for the tool
- receive a command to execute an automated move along a path e.g., a line
- a path e.g., a line
- That path also may be align
- the computer system may also be configured to receive (e.g., prior to the command identifying the target location) a command selecting which robotic arm to use for the automated move. In some embodiments, the computer system may also be configured to receive (e.g., prior to the command identifying the target location) a command indicating a simulation and/or setup mode that disengages an input device that is linked in a master-slave relationship to a robotic arm holding or integrated with the tool, such that in the simulation mode movement of the input device does not cause movement of the robotic arm.
- the computer system may also be configured to receive a command (e.g., prior to the command identifying the target location) indicating a user's activation of the input device (such as through the user touching a button on in the input device with the user's hand), which activation allows the user to alter the position of the tracking indicator showing the location of the would-be tool tip relative to the image(s) of the subject as the user determines where to position the tracking indicator for selection of the target and starting locations.
- the computer system may also be configured to receive a command (e.g., after the command identifying the starting location) indicating a new (e.g., a second) target location.
- the computer system may also be configured to receive a command (e.g., after the command identifying the starting location) indicating a new (e.g., a second) starting location. In some embodiments, the computer system may also be configured to receive a command (e.g., after the command identifying the starting location) indicating termination of the simulation and/or setup mode. In some embodiments, the computer system may be configured to display on a GUI a control (e.g., that can be selected via a touch, a mouse, or the like) that can be used to select a mode in which the input device is engaged with the robotic arm in a master-slave relationship.
- a control e.g., that can be selected via a touch, a mouse, or the like
- the computer system may be configured to receive a command, when in the master-slave mode, enabling the input device (e.g., by holding the input device and pushing a button on the input device with a finger of the user).
- the computer system may be configured to receive a command to execute the automated move along a path that is defined at least in part by the starting and target locations, the computer system also be configured to cause the robotic arm in a way that moves the tool in a single axis along the path only after it has received a command indicating the input device is enabled (e.g., such that a user must be holding the input device in order for the automated move to proceed).
- the computer system may be configured to stop the robotic arm from completing the automated move if it receives a command to stop the automated move (e.g., through a user pushing the same button on the input device that otherwise enables the input device), and may also be configured to display on a GUI a message that includes buttons or the like (e.g., that can be selected via a touch, a mouse, or the like) for continuing with the automated move, reversing direction, or stopping, and may be configured to receive a command to either continue, reverse direction or stop, depending on the button or the like that is activated, provided it first receives a command indicating the input device is enabled (e.g., by holding the input device and pushing a button on the input device with the user's finger).
- a command to stop the automated move e.g., through a user pushing the same button on the input device that otherwise enables the input device
- buttons or the like e.g., that can be selected via a touch, a mouse, or the like
- any embodiment of any of the present methods, devices, and systems may consist of or consist essentially of — rather than comprise/include/contain/have — the described steps and/or features.
- the term “consisting of or “consisting essentially of may be substituted for any of the open-ended linking verbs recited above, in order to change the scope of a given claim from what it would otherwise be using the open-ended linking verb.
- FIG. IA is a perspective view of one embodiment of two input devices (hand controllers) that may be used consistent with the present techniques.
- FIG. IB is an enlarged view of a left-handed input device.
- FIGS. 1C- IE are different views showing a tool held by a robotic arm located in a first position of a stereotactic procedure.
- FIGS. 2A-2C are different views showing the tool from FIGS. 1C-1E in a second position of a stereotactic procedure.
- FIGS. 3A-3C are different views showing the tool from FIGS. 1C-1E in a third position of a stereotactic procedure.
- FIG. 4 is a perspective view of a workstation for use in planning and controlling the tool movement shown in FIGS. 1C-3C.
- FIG. 5 shows a graphical user interface that can be used in the set up and control of the tool movement shown in FIGS. 1C-3C.
- FIG. 6 shows the GUI of FIG. 5 in a training/simulation mode involving both robotic arms.
- FIG. 7 shows the GUI of FIG. 5 in a training/simulation mode in which the left arm has been chosen for stereotaxy.
- FIG. 8 shows the GUI of FIG. 5 in a training/simulation mode in which the right arm has been chosen for stereotaxy and the Z Axis Lock function for the tool of that arm has been activated.
- FIG. 9 shows the GUI of FIG. 5 in a training/simulation mode in which the microsurgery application has been chosen, both arms are enabled, and both tools have been chosen.
- FIG. 10 shows the GUI of FIG. 5 in a training/simulation mode in which the left arm has been chosen for stereotaxy (as in FIG. 7) and the Z Axis Lock function for the tool of that arm has been activated.
- FIG. 11 shows another GUI that can be used in the set up of the tool movement shown in FIGS. 1C-3C.
- An indicator in the form of crosshairs corresponding to the location of the tool tip is shown on a 2D image, representing the tool tip's location relative to the portion of the subject in the 2D image.
- FIG. 12 shows the GUI of FIG. 11 in a 3D mode in which a representation of the tool chosen for use in the training/simulation is displayed at the relevant depth within a 3D image of a portion of the subject.
- the GUI reflects that a user has selected the "Plane Cut” option, which results in oblique slices being cut away on the head to the relevant tool tip depth.
- FIG. 13 shows a warning box that can appear on a GUI such as the one in FIG. 5 if a problem is encountered during an automated procedure.
- a method, device, or system that "comprises,” “has,” “contains,” or “includes” one or more recited steps or elements possesses those recited steps or elements, but is not limited to possessing only those steps or elements; it may possess (i.e., cover) elements or steps that are not recited.
- an element of a method, device, or system that "comprises,” “has,” “contains,” or “includes” one or more recited features possesses those features, but is not limited to possessing only those features; it may possess features that are not recited.
- a computer readable medium “comprising" (or “encoded with”) machine readable instructions for performing certain steps is a computer readable medium that has machine readable instructions for implementing at least the recited steps, but also covers media having machine readable instructions for implementing additional, unrecited steps.
- a computer system that is configured to perform at least certain functions is not limited to performing only the recited functions, and may be configured in a way or ways that are not specified provided the system is configured to perform the recited functions.
- the invention is a software enabled single-axis lock for movement of a tool along the single axis by a robotic arm with multiple degrees of freedom.
- the software solution allows a robotic arm with an unlimited number of degrees of freedom to behave in the same fashion as a robot or device that is mechanically restricted to motion of its tool along the single axis.
- a command may be sent to the software to lock the motion by a given robotic arm of its tool (meaning a tool the robotic arm is holding or that is integral with the robotic arm; the present tools may be characterized more specifically as medical tools or surgical tools) in a single axis using any suitable input device, such as a button on a touch screen on a GUI, a button on an input device (e.g., a hand controller), or the like.
- the apparatus to which the inventive techniques may be applied may, in some embodiments, include a slave robotic arm commanded by a master input device, such as a hand controller.
- a master input device such as a hand controller.
- FIG. IA An example of a pair of input devices (in the form of hand controllers) that can be used to control two different robotic arms, respectively, of a medical or surgical robotic system are shown in FIG. IA.
- Input devices 20, which are mirror images of each other, each includes a stylus 25 that can be held like a long pen, lever 27 that can be squeezed toward stylus 25 to cause a tool integrated with or held by the slave robotic arm to actuate (e.g., squeezing lever 27 can cause forceps to close), and an enable/disable button 29 that can be touched and held for a short amount of time in order to activate the input device.
- One way to hold input devices 20 is to grasp stylus 25 so that lever 27 can be squeezed with the forefinger and so that button 29 can be touched with the thumb.
- FIG. IB shows an enlarged view of the left-handed input device 20.
- One manner of creating a non-mechanical single-axis tool movement lock involves the following: a) retrieving an input device (e.g., hand controller) command in tool tip space (e.g., Cartesian X, Y, Z, roll, pitch, yaw).
- an input device e.g., hand controller
- tool tip space e.g., Cartesian X, Y, Z, roll, pitch, yaw.
- This retrieving may comprise receiving a hand controller signal(s) (command(s), or data) signifying the position and orientation of the hand controller; determining (e.g., calculating) a delta value of the movement of the hand controller in a single axis (e.g., an axis that is related by a transformation to the single axis to which tool tip movement is restricted); and determining a corresponding delta value for the tool tip using that hand controller delta. If a transformation from delta values in hand controller space to delta values in tool tip space is determined, all tool tip delta values may be ignored except the delta along the relevant single axis.
- the rate of execution of the above loop may be arbitrarily small to produce linear motion at the tool tip.
- the movement of a tool along a single axis may be pre-programmed so as to be automated.
- FIGS. 1C-3C detailed views of a manipulator 100 and a surgical tool 150 are shown in various positions as the manipulator 100 causes movement of the tool along a single axis in a stereotactic procedure (such a movement also may be achieved in any other procedure, such as a microsurgical procedure).
- Manipulator 100 which is an example of a multi-degree of freedom robotic arm (specifically, manipulator 100 may be characterized as a six degree of freedom slave manipulator, and it is similar in functionality and operation to the robotic arms disclosed in the '316 patent), assembly 200 comprising a head clamp and radio- frequency coil device (which is coupled to the head clamp, and which can be further coupled to the operating room table by a fixable multi link arm), and cameras 190 (only one of which is visible (the other is on the opposite side of the extension board)) are coupled to an extension board 260.
- Extension board 260 may be coupled to any suitable table or other structure having a patient support surface (not shown). In the views shown, a schematic drawing of a patient's head 300 is shown held by the head clamp of assembly 200.
- FIGS. 1C and ID show manipulator 100 in a first position in which surgical tool 150 is located outside of head 300, near opening 350 in head 300, which may be a burr hole or any other suitable surgical opening. However, tip 160 of surgical tool 150 is outside of opening 350.
- FIG. IE is a side view of the position shown in FIGS. 1C and ID, and does not include assembly 200 for clarity.
- FIGS. 2 A and 2B show manipulator 100 moved to a second position in which tip 160 of surgical tool 150 has been advanced along axis 110, and no other axis, by manipulator 100 so that tip 160 has penetrated the boundary of opening 350.
- FIG. 2C is a side view of the position shown in FIGS. 2A and 2B, and does not include assembly 200 for clarity.
- FIG. 3 A and 3B show manipulator 100 moved to a third position in which tip 160 has moved along axis 110 into a location within head 300, which it can be further manipulated by a user/operator (e.g., a surgeon) to perform any of a number of functions, such as taking a biopsy of tissue.
- FIG. 3C is a side view of the position shown in FIGS. 3A and 3B, and does not include assembly 200 for clarity.
- Axis 110 is substantially aligned (perfectly aligned in the depicted embodiment) with the longitudinal axis (not separately shown) of tool 150. For other tools that have bends or angles, the tool and tool tip will still move along a single axis, however that axis may not coincide with a longitudinal axis of the tool itself.
- FIG. 4 illustrates a perspective view of a workstation 400 that can be used to control manipulator 100 (or two such manipulators) and surgical tool 150 (or two such surgical tools, one held by each of two such manipulators).
- workstation 400 comprises input devices 20 shown in FIGS. IA and IB to control movement of manipulator 100.
- Workstation 400 may include a table to which the input devices are secured as well as a series of display screens, including display screens 401 and 402, each of which can provide a graphical user interface (GUI) that can be used in setting up a procedure using manipulator 100.
- GUI graphical user interface
- the GUI shown on display screen 401 may be used to select two points that will define the axis (or path or trajectory) along which the tip of the relevant tool travels in an automated single axis movement (such a screen is referred to as a command status display (CSD) in this document) and the GUI shown on display screen 402 may be used to display one or more images from a three- dimensional dataset of images of a subject taken using a three-dimensional imaging device, such as a magnetic resonance imaging device, which may be viewed as a determination is made by an operator about which points to select on display screen 402 (such a screen is referred to in this document as a magnetic resonance image display (MRID)).
- the other display screens depicted in FIG. 4 may be used to show other images or displays associated with a given procedure involving one or both manipulators 100.
- FIGS. 5-11 illustrate various screen displays that can be used as GUIs for displays 401 and 402.
- multiple controls such as buttons, slider bars, radio buttons, check boxes, drop down menus, and the like
- buttons, slider bars, radio buttons, check boxes, drop down menus, and the like are provided on each screen for receiving user input through any suitable means, such as through touching the screen, manipulating an input device such as a mouse, or the like. Only those controls relevant to the features presented in this disclosure will be discussed.
- a computer system may be configured such that starting the primary application supported by the computer system brings the user to a startup screen as illustrated in FIG. 5.
- Those of ordinary skill in the art having the benefit of this disclosure will be able to write code (machine readable instructions, which can be implemented through software, hardware, firmware, or a combination of any two or more of these) without undue experimentation for accomplishing the features (including the graphical user interfaces) described below and shown in the figures.
- FIG. 5 illustrates a CSD 401 that can be used in setting up a desired mode for one or both manipulators (such as a "Z Axis Lock” representative of a manipulators ability to move its tool along only one axis) or a desired procedure, such as an automated move (e.g., along a single axis) of a surgical tool by a given manipulator 100.
- manipulators such as a "Z Axis Lock” representative of a manipulators ability to move its tool along only one axis
- a desired procedure such as an automated move (e.g., along a single axis) of a surgical tool by a given manipulator 100.
- This display includes options for selecting procedure types (microsurgery, stereotaxy left arm, stereotaxy right arm), as well as power selections for the right arm, left arm, a base motor for adjusting the height of the arms (manipulators 100) supported on a mobile and lockable base, a field camera for capturing images during microsurgery and a digitizing arm for use in the physical part of subject image-to- manipulator registration.
- the power buttons are shown in the "Mode Controls" tab, as is the "Surgery Type.”
- Manipulators 100 are shown in an unhighlighted manner on the GUI shown in FIG. 5, signifying that neither has been selected for using in either training/simulation or a procedure using the buttons at the bottom left of the screen.
- a suitable technique for registering one or more two-dimensional images of a portion of a subject with one or both manipulators 100 is disclosed in co-pending International Application No. PCT/US08/60538, which is incorporated by reference.
- suitable registration may include both an MRI registration aspect to locate the imaged subject to a physical structure and a physical registration aspect to register a given manipulator to that physical structure, set up may begin.
- a user may select a simulation mode by selecting the "Training Simulation Mode” button under the "User Settings” tab shown in CSD 401 of FIG. 6. Selecting the simulation mode can allow the user to view simulated movements of manipulator 100 and surgical tool 150 in response to movements of the input device, without causing actual movement of manipulator 100.
- the word “simulation” also appears near the bottom portion of the display, as shown for example in FIGS. 5-7.
- simulation mode a user can view a potential path of travel of surgical tool 150 that may be used in a surgical procedure. This allows a user to evaluate multiple potential paths of manipulator 100 and surgical tool 150 before defining one as described below for actual use in the procedure.
- CSD 401 in FIG. 7 illustrates the system in simulated stereotaxy mode with the left arm enabled.
- This version of CSD 401 now shows only one manipulator as a result of the left arm selection, and shows it in a highlighted state. It also shows an upper portion of an RF coil device (from assembly 200) positioned over a graphical representation of a subject's head (e.g., head 300). It also shows that the user has enabled power to the left arm and a "Bore Camera” (or cameras, such as camera 190 shown in FIGS. 1C-3C, which may be exposed without being affected to the magnetic field created in an MRI environment) and the digitizing arm (note that the unselected "Right Arm” and "Base Motor” buttons are unselected and grayed out).
- FIG. 8 illustrates a version of CSD 401 indicating that a user has selected to place the right arm in stereotaxy mode and Z Axis Lock mode, where the tool that has been selected for use by the right manipulator is shown on the right lower part of the screen (and is the same biopsy tool shown in FIGS. 1C-3C).
- the mode of the displayed manipulator shown in FIG. 8 was achieved through a user's selection of stereotaxy right arm (as shown in the buttons in FIG. 5), master/slave mode via selection of the Master/Slave button shown in FIG. 8, and the enablement of the right arm by selecting "Right Arm” in the "Arm Enable” box of the "Mode Controls" tab shown in FIG. 8.
- the user enables the input device associated with the right manipulator by depressing button 29 on right hand controller 20.
- the user can manipulate the enabled input device to put the manipulator into the position and orientation desired by the user for movement of the tool along a single axis.
- the user can disable control of the manipulator by again pushing button 29; otherwise, the user can proceed to enabling the z-axis lock for the tool held by that manipulator by (in the depicted embodiment) selecting "Right Tool” in the "Z Axis Lock” box of the "Mode Controls" tab shown in FIG. 8.
- the tool held by the manipulator will only travel along the axis defined (in the depicted embodiment) by the upper portion of the tool where it is held by the tool holder portions coupled to the end effector of the manipulator (which, in this embodiment, is a longitudinal axis that is centered in the entire length of tool), such travel occurring in the forward or backward directions depending on the user's motion of the input device.
- the user can push the same "Right Tool” button to disable that mode.
- FIG. 9 illustrates a version of CSD 401 indicating that a user has selected microsurgical mode and simulation mode, and enabled both manipulators (which are both highlighted) and selected tools for them.
- a user may enable the Z Axis Lock function for the tools of both arms from this version of CSD 401.
- the selected tool for each manipulator is shown to the side of the manipulator (bipolar forceps on the left and biopsy tool on the right).
- FIG. 10 illustrates a version of CSD 401 in which a simulated stereotaxy left arm mode has been selected (by, for example, selecting the "Stereotaxy Left Arm” button shown in FIG. 5), the left arm has been enabled, the Z Axis Lock function has been selected for the left tool.
- MRID 402 depicts a GUI that allows a user to toggle between 2D and 3D views taken with a 3-D imaging modality (such as an MRI machine) of a portion (such as the head) of a subject, as reflected in the 2D tabs "2D Tools” and “2D View” at the bottom left of the screen and in the 3D tabs "3D Tools” and “3D View” at the bottom right of the screen.
- a 3-D imaging modality such as an MRI machine
- an indicator in this example, crosshairs
- the location of the tip e.g., tip 160
- the relevant tool e.g., surgical tool 150, or, in other embodiments, the terminal end of an extension line that extends from the tool tip a distance selected using the slider bar shown underneath the "Tool Tip Extension:” box beneath the tool that is being tracked
- the slider bar shown underneath the "Tool Tip Extension:” box beneath the tool that is being tracked
- the 2D Tools tab has been selected, and a two-dimensional image is shown overlaid by the crosshairs indicator showing the location of the tip of the right tool within the subject.
- crosshairs appear in response to a user selecting the "Track" button beneath the section for the relevant tool(s).
- the Track option on MRID 402 a user can view the MRID as he or she manipulates the relevant input device to follow (or track) the location of the tool tip (or tool tip extension line end point, and regardless of whether the user is in simulation/training mode) relative to the subject, as it travels through the subject.
- the 3D Tools tab has been selected and the location of the tip of the tool relative to the subject's head is shown in 3D, where the 3D image is shown in this embodiment cut away on a plane that is normal to the axis along which the tool tip will travel, as a result of the selection of the "Plane Cut” button within the "Right Tool” box near the right of the screen.
- a user can manipulate the orientation of the 3D image through any suitable input device (e.g., a space ball) to move the displayed image and the overlaid tool so as to provide a desired view of the tissue affected by the proposed tool position and path.
- This overlay feature becomes available following the physical and MRI registration process and receipt of tool selection.
- Selection of the "Simple” button will replace the tool image with, for example, a thin red line of the same length as the tool so as not to obstruct the view of small structures.
- Selection of the "Wedge Cut” button will cut into the displayed 3D image at the location of the tool tip/extension line end by cutting away a wedge to reveal three orthogonal planes (e.g., sagital, axial, coronal), where the tip of the tool/extension line end is at the juncture of the three planes.
- a user may first select a mode on the CSD, such as Stereotaxy Left Arm Mode, and then enable the left arm and power on the bore camera(s). The user may then choose the Simulation Mode on the CSD to disengage the left manipulator (which may, for example, be a left version of manipulator 100 from FIGS. 1C-3C or one of the manipulators shown in the '316 patent) from the motion of the relevant input device (such as input device 20).
- a mode on the CSD such as Stereotaxy Left Arm Mode
- the user may then choose the Simulation Mode on the CSD to disengage the left manipulator (which may, for example, be a left version of manipulator 100 from FIGS. 1C-3C or one of the manipulators shown in the '316 patent) from the motion of the relevant input device (such as input device 20).
- the user may then select the 2D Tools tab and the Track mode/function in the "Left Tool” box, causing the crosshairs to appear overlaying the relevant 2D image of the subject when the 2D mode is selected.
- a user may select a non-zero "Tool Tip Extension” value, using the slider bar, if a tool tip extension line is desired. If the Tool Tip Extension function is set greater than 0.0 mm, the crosshairs will track the location of the end of the extension line. If this parameter is set at zero, the tracking function will illustrate crosshairs on the 2D slice image at the location of the tip (distal end) of the tool.
- subsequent 2D images e.g., 2D slices
- prior 2D slices are shown.
- the user can grasp the left input device and enable virtual or simulated motion of the tool by actuating (e.g., via use of the thumb) an enable button (e.g., button 29) on the input device.
- the user can then take the input device, and based on visual cues gained from toggling, as desired, between the 2D and 3D MRID views, move the virtual manipulator shown on the CSD and the manipulator's surgical tool to the area of the intended target.
- the CSD and the 2D and 3D MRID images can update in real time to show the location of the virtual (when in simulation mode) manipulator and its tool.
- the user may disable the input device so that movement of the input device does not lead to further movement of the virtual manipulator and its tool.
- a user can then push "Target” to select the target location for the procedure, which is stored in terms of X, Y and Z Cartesian coordinates of the tool tip in image space (e.g., magnetic resonance imaging space), which is then transformed to robot space.
- image space e.g., magnetic resonance imaging space
- a target location indicator will appear at the crosshairs location (for example, a red circle) in the 2D view and at the tool tip or extension line end location in the 3D view denoting the intended target.
- a user can then enable the input device if it has been disabled (and in the same way that the input device was disabled) and cause the tool tip or extension line end to move toward the intended insertion point for the subject.
- a path indicator (for example, a green line) can then be visible in the 3D view that links the tip/extension line end to the selected target so that the user can see the trajectory and any tissue that will be penetrated or disturbed by the proposed tool tip path.
- the user may then move the input device to the desired entry point (which could be, for example, at the surface of the brain or head, or a small distance outside the head). If a burr hole has already been made, a user may ensure that the path goes through the burr hole without contacting the head. The user may then, but need not, disable the input device when the entry point and trajectory are acceptable.
- a user may then push the button labeled "Start Point" on either the 2D or 3D version of the MRID and an indicator (e.g., a green circle) will appear at the crosshairs location in the 2D view and at the tool tip or extension line end location in the 3D view denoting the intended start point, which is stored in terms of X, Y and Z Cartesian coordinates of the tool tip in image space (e.g., magnetic resonance imaging space), which is then transformed to robot space. These coordinates are registered as the tool tip if the extension line value equals zero, or as the end of the extension line if that value is greater than zero.
- an indicator e.g., a green circle
- the indicator will change in some way (e.g., the green line will turn red) to denote that the line (or path) is set, and the start point, termination point and trajectory (or path) will appear on the CSD.
- the user can use the input device to move the simulated tool tip/extension line to a new location and push the "Start Point" or "Target” button again in either the 2D or 3D version of the MRID.
- the system will ask the user to push the relevant button a second time to confirm replacement of the old point with a new point. After an acceptable trajectory is chosen, the user can exit the simulation mode on the CSD by designating that button again (e.g., by touching it on the screen again).
- a user can execute the automated move by choosing the master/slave mode on the CSD, enabling the input device (e.g., by depressing button 29), and moving the input device to cause the manipulator (while watching the MRID and/or the bore camera or field camera image shown on display screen 403 shown in FIG. 4) to move to a location close to the start point selected for the movement and to be in an orientation that is as close to the selected trajectory as possible.
- the user may then disable the input device.
- the user can push the "Tool Align” button (see, e.g., FIGS. 11 and 12) and the manipulator will move to align with the programmed trajectory and place the tool tip at or near the selected start point (such as approximately two centimeters radially outward from the start point along the programmed trajectory).
- the user may then push the "Execute” button (under “Automated Biopsy"), and the user may be prompted to enable the input device to begin the automated movement (e.g., an automated biopsy).
- the user may then grasp the input device and enable the system to begin the automated biopsy by enabling the input device in the same way it has been previously enabled (e.g., by pushing button 29). Taking this step causes the user to hold the input device in order for the procedure to take place.
- the tool may move forward at a predetermined rate (which can be set in an initialization file) to the target location, at which point the surgical tool can perform a pre-programmed function, such as removing biopsy material.
- the surgical tool is a biopsy tool equipped with two small sharpened scoops that open away from each other about axes that are normal to the longitudinal axis of the tool
- the surgical tool's scoops will open, rotate 90 degrees clockwise, and close again, capturing tissue as a result.
- the surgical tool can then reverse direction straight out along the insertion trajectory.
- the user can disable the input device (e.g., by depressing button 29) to stop the move.
- the CSD will then present the user with a selection box, such as the one shown in FIG. 13, that includes options to stop, continue, and reverse direction. Once a selection is chosen, the tool will move again when the user enables the input device.
- embodiments of the present methods, devices, and systems may therefore also allow a user (e.g., a surgeon) to simulate multiple paths for a surgical tool prior to conducting the actual surgical procedure, evaluate those paths for the tissue they may affect, and choose a desired path by selecting a target point and a start point.
- the present devices and systems are configured to limit (electronically) the tool to a linear path; as a result, only a start point and a target point are needed to determine the tool path.
- Embodiments of the present devices and system may also comprise multiple safety features to allow the user to maintain control of the tool.
- Embodiments of the present methods may be coded as software stored on any suitable computer readable media (e.g., tangible computer readable media), such as any suitable form of memory or data storage device, including but not limited to hard drive media, optical media, RAM, SRAM, DRAM, SDRAM, ROM, EPROM, EEPROM, tape media, cartridge media, flash memory, memory stick, and/or the like.
- Tangible computer readable media includes any physical medium that can store or transfer information. Such embodiments may be characterized as tangible computer readable media having (or encoded with) computer executable (e.g., machine readable) instructions for performing certain step(s).
- tangible computer readable medium does not include wireless transmission media, such as carrier waves.
- the term "computer readable medium,” however, does cover wireless transmission media, and some embodiments of the present methods may include wireless transmission media carrying the computer readable instructions described above.
- the software can be written according to any technique known in the art. For instance, the software may be written in any one or more computer languages (e.g., ASSEMBLY, PASCAL, FORTRAN, BASIC, C, C++, C#, JAVA, Perl, Python) or using scientific packages like, but not limited to, Matlab®, R, S-plus®, and SAS®.
- the code may be to enable it to be compiled on all common platforms (e.g., Microsoft®, Linux®, Apple Macintosh® OS X, Unix®).
- the software may include a GUI, which may provide a user with a more intuitive feel when running the software.
- Different fields may be accessible by screen touching, a mouse and/or keyboard.
- Alarms, cues, and the like may be done via pop-up windows, audible alerts, or any other techniques known in the art.
- Some (up to all) of the steps described in the sections above may be implemented using a computer having a processor (e.g., one or more integrated circuits) programmed with firmware and/or running software.
- Some (up to all) of the steps described in the sections above may be implemented using a distributed computing environment, which is one example of a computer system.
- multiple computers may be used, such as those connected by any suitable number of connection mediums (e.g., a local area network (LAN), a wide area network (WAN), or other computer networks, including but not limited to Ethernets, enterprise-wide computer networks, intranets and the Internet, and the connections between computers can be wired or wireless).
- Servers and user terminals can be part of a given computer system.
- embodiments of suitable computer systems may be implemented on application specific integrated circuits (ASICs) or very large scale integrated (VLSI) circuits, and further (or alternatively) may be configured to use virtualization of resources, virtual computing, and/or cloud computing to achieve the specified functions.
- ASICs application specific integrated circuits
- VLSI very large scale integrated circuits
- persons of ordinary skill in the art may utilize any number of suitable structures capable of executing logical operations in order to achieve the functions described above in a computer system consistent with this disclosure.
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Surgery (AREA)
- Mechanical Engineering (AREA)
- Heart & Thoracic Surgery (AREA)
- Biomedical Technology (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Medical Informatics (AREA)
- Molecular Biology (AREA)
- Animal Behavior & Ethology (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Veterinary Medicine (AREA)
- Manipulator (AREA)
Abstract
Description
Claims
Priority Applications (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA2684459A CA2684459C (en) | 2007-04-16 | 2008-04-16 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
EP08832194A EP2142132B1 (en) | 2007-04-16 | 2008-04-16 | System for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
US12/596,426 US8560118B2 (en) | 2007-04-16 | 2008-04-16 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
JP2010503622A JP5543331B2 (en) | 2007-04-16 | 2008-04-16 | Method, apparatus, and system for non-mechanically limiting and / or programming movement along one axis of a manipulator tool |
IL201559A IL201559A0 (en) | 2007-04-16 | 2009-10-15 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
US14/054,664 US9131986B2 (en) | 2007-04-16 | 2013-10-15 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US91214607P | 2007-04-16 | 2007-04-16 | |
US60/912,146 | 2007-04-16 | ||
US2008060541 | 2008-04-16 | ||
USPCT/US2008/060541 | 2008-04-16 |
Related Child Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/596,426 A-371-Of-International US8560118B2 (en) | 2007-04-16 | 2008-04-16 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
US14/054,664 Continuation US9131986B2 (en) | 2007-04-16 | 2013-10-15 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
Publications (2)
Publication Number | Publication Date |
---|---|
WO2009037576A2 true WO2009037576A2 (en) | 2009-03-26 |
WO2009037576A3 WO2009037576A3 (en) | 2010-02-04 |
Family
ID=40468504
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/IB2008/003323 WO2009037576A2 (en) | 2007-04-16 | 2008-04-16 | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis |
Country Status (3)
Country | Link |
---|---|
JP (1) | JP5543331B2 (en) |
CA (1) | CA2684459C (en) |
WO (1) | WO2009037576A2 (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2010117685A2 (en) * | 2009-03-31 | 2010-10-14 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
EP2384714A1 (en) * | 2010-05-03 | 2011-11-09 | Universitat Politècnica de Catalunya | A method for defining working space limits in robotic surgery |
US8620473B2 (en) | 2007-06-13 | 2013-12-31 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
EP2704658A2 (en) * | 2011-05-05 | 2014-03-12 | The Johns Hopkins University | Method and system for analyzing a task trajectory |
US8864652B2 (en) | 2008-06-27 | 2014-10-21 | Intuitive Surgical Operations, Inc. | Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the positioning and orienting of its tip |
US8944070B2 (en) | 1999-04-07 | 2015-02-03 | Intuitive Surgical Operations, Inc. | Non-force reflecting method for providing tool force information to a user of a telesurgical system |
US9084623B2 (en) | 2009-08-15 | 2015-07-21 | Intuitive Surgical Operations, Inc. | Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide |
US9089256B2 (en) | 2008-06-27 | 2015-07-28 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
US9345387B2 (en) | 2006-06-13 | 2016-05-24 | Intuitive Surgical Operations, Inc. | Preventing instrument/tissue collisions |
WO2016089753A1 (en) * | 2014-12-03 | 2016-06-09 | Gambro Lundia Ab | Medical treatment system training |
US9492927B2 (en) | 2009-08-15 | 2016-11-15 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
US9572481B2 (en) | 2011-05-13 | 2017-02-21 | Intuitive Surgical Operations, Inc. | Medical system with multiple operating modes for steering a medical instrument through linked body passages |
US9622826B2 (en) | 2010-02-12 | 2017-04-18 | Intuitive Surgical Operations, Inc. | Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument |
US9629520B2 (en) | 2007-06-13 | 2017-04-25 | Intuitive Surgical Operations, Inc. | Method and system for moving an articulated instrument back towards an entry guide while automatically reconfiguring the articulated instrument for retraction into the entry guide |
US9718190B2 (en) | 2006-06-29 | 2017-08-01 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
CN107205789A (en) * | 2015-06-01 | 2017-09-26 | 奥林巴斯株式会社 | Medical manipulator system |
US9788909B2 (en) | 2006-06-29 | 2017-10-17 | Intuitive Surgical Operations, Inc | Synthetic representation of a surgical instrument |
US10008017B2 (en) | 2006-06-29 | 2018-06-26 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
US10258425B2 (en) | 2008-06-27 | 2019-04-16 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide |
US10271912B2 (en) | 2007-06-13 | 2019-04-30 | Intuitive Surgical Operations, Inc. | Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide |
CN110403701A (en) * | 2019-08-30 | 2019-11-05 | 山东威高手术机器人有限公司 | Input unit, minimally invasive surgical operation robot and micro-wound surgical operation control method |
US10507066B2 (en) | 2013-02-15 | 2019-12-17 | Intuitive Surgical Operations, Inc. | Providing information of tools by filtering image areas adjacent to or on displayed images of the tools |
EP3595565A4 (en) * | 2017-03-15 | 2021-04-14 | Covidien LP | Robotic surgical systems, instruments, and controls |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8903546B2 (en) | 2009-08-15 | 2014-12-02 | Intuitive Surgical Operations, Inc. | Smooth control of an articulated instrument across areas with different work space conditions |
WO2009042073A2 (en) * | 2007-09-21 | 2009-04-02 | Cabot Microelectronics Corporation | Polishing composition and method utilizing abrasive particles treated with an aminosilane |
JPWO2023140120A1 (en) * | 2022-01-21 | 2023-07-27 |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5078140A (en) * | 1986-05-08 | 1992-01-07 | Kwoh Yik S | Imaging device - aided robotic stereotaxis system |
US7074179B2 (en) * | 1992-08-10 | 2006-07-11 | Intuitive Surgical Inc | Method and apparatus for performing minimally invasive cardiac procedures |
US6120433A (en) * | 1994-09-01 | 2000-09-19 | Olympus Optical Co., Ltd. | Surgical manipulator system |
JPH08117238A (en) * | 1994-10-25 | 1996-05-14 | Olympus Optical Co Ltd | Surgical manipulator |
US5855583A (en) * | 1996-02-20 | 1999-01-05 | Computer Motion, Inc. | Method and apparatus for performing minimally invasive cardiac procedures |
US6436107B1 (en) * | 1996-02-20 | 2002-08-20 | Computer Motion, Inc. | Method and apparatus for performing minimally invasive surgical procedures |
JP3377740B2 (en) * | 1996-12-16 | 2003-02-17 | 株式会社三協精機製作所 | Control method of force assist device and control device using this method |
US6671538B1 (en) * | 1999-11-26 | 2003-12-30 | Koninklijke Philips Electronics, N.V. | Interface system for use with imaging devices to facilitate visualization of image-guided interventional procedure planning |
US20010025183A1 (en) * | 2000-02-25 | 2001-09-27 | Ramin Shahidi | Methods and apparatuses for maintaining a trajectory in sterotaxi for tracking a target inside a body |
US7121781B2 (en) * | 2003-06-11 | 2006-10-17 | Intuitive Surgical | Surgical instrument with a universal wrist |
JP2006312079A (en) * | 2006-08-09 | 2006-11-16 | Olympus Corp | Medical manipulator |
JP5078140B2 (en) * | 2007-09-03 | 2012-11-21 | 国立大学法人横浜国立大学 | Microstructure modeling method |
-
2008
- 2008-04-16 CA CA2684459A patent/CA2684459C/en not_active Expired - Fee Related
- 2008-04-16 JP JP2010503622A patent/JP5543331B2/en not_active Expired - Fee Related
- 2008-04-16 WO PCT/IB2008/003323 patent/WO2009037576A2/en active Application Filing
Non-Patent Citations (2)
Title |
---|
None |
See also references of EP2142132A4 |
Cited By (72)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8944070B2 (en) | 1999-04-07 | 2015-02-03 | Intuitive Surgical Operations, Inc. | Non-force reflecting method for providing tool force information to a user of a telesurgical system |
US10433919B2 (en) | 1999-04-07 | 2019-10-08 | Intuitive Surgical Operations, Inc. | Non-force reflecting method for providing tool force information to a user of a telesurgical system |
US10271909B2 (en) | 1999-04-07 | 2019-04-30 | Intuitive Surgical Operations, Inc. | Display of computer generated image of an out-of-view portion of a medical device adjacent a real-time image of an in-view portion of the medical device |
US9232984B2 (en) | 1999-04-07 | 2016-01-12 | Intuitive Surgical Operations, Inc. | Real-time generation of three-dimensional ultrasound image using a two-dimensional ultrasound transducer in a robotic system |
US9101397B2 (en) | 1999-04-07 | 2015-08-11 | Intuitive Surgical Operations, Inc. | Real-time generation of three-dimensional ultrasound image using a two-dimensional ultrasound transducer in a robotic system |
US9345387B2 (en) | 2006-06-13 | 2016-05-24 | Intuitive Surgical Operations, Inc. | Preventing instrument/tissue collisions |
US10008017B2 (en) | 2006-06-29 | 2018-06-26 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
US10773388B2 (en) | 2006-06-29 | 2020-09-15 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
US10730187B2 (en) | 2006-06-29 | 2020-08-04 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
US10737394B2 (en) | 2006-06-29 | 2020-08-11 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
US10137575B2 (en) | 2006-06-29 | 2018-11-27 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
US11638999B2 (en) | 2006-06-29 | 2023-05-02 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
US11865729B2 (en) | 2006-06-29 | 2024-01-09 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
US9801690B2 (en) | 2006-06-29 | 2017-10-31 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical instrument |
US9789608B2 (en) | 2006-06-29 | 2017-10-17 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
US9788909B2 (en) | 2006-06-29 | 2017-10-17 | Intuitive Surgical Operations, Inc | Synthetic representation of a surgical instrument |
US9718190B2 (en) | 2006-06-29 | 2017-08-01 | Intuitive Surgical Operations, Inc. | Tool position and identification indicator displayed in a boundary area of a computer display screen |
US9901408B2 (en) | 2007-06-13 | 2018-02-27 | Intuitive Surgical Operations, Inc. | Preventing instrument/tissue collisions |
US10695136B2 (en) | 2007-06-13 | 2020-06-30 | Intuitive Surgical Operations, Inc. | Preventing instrument/tissue collisions |
US12097002B2 (en) | 2007-06-13 | 2024-09-24 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US10271912B2 (en) | 2007-06-13 | 2019-04-30 | Intuitive Surgical Operations, Inc. | Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide |
US11432888B2 (en) | 2007-06-13 | 2022-09-06 | Intuitive Surgical Operations, Inc. | Method and system for moving a plurality of articulated instruments in tandem back towards an entry guide |
US11399908B2 (en) | 2007-06-13 | 2022-08-02 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US9629520B2 (en) | 2007-06-13 | 2017-04-25 | Intuitive Surgical Operations, Inc. | Method and system for moving an articulated instrument back towards an entry guide while automatically reconfiguring the articulated instrument for retraction into the entry guide |
US11751955B2 (en) | 2007-06-13 | 2023-09-12 | Intuitive Surgical Operations, Inc. | Method and system for retracting an instrument into an entry guide |
US10188472B2 (en) | 2007-06-13 | 2019-01-29 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US8620473B2 (en) | 2007-06-13 | 2013-12-31 | Intuitive Surgical Operations, Inc. | Medical robotic system with coupled control modes |
US9717563B2 (en) | 2008-06-27 | 2017-08-01 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxilary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
US11382702B2 (en) | 2008-06-27 | 2022-07-12 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
US9516996B2 (en) | 2008-06-27 | 2016-12-13 | Intuitive Surgical Operations, Inc. | Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the position and orienting of its tip |
US8864652B2 (en) | 2008-06-27 | 2014-10-21 | Intuitive Surgical Operations, Inc. | Medical robotic system providing computer generated auxiliary views of a camera instrument for controlling the positioning and orienting of its tip |
US10258425B2 (en) | 2008-06-27 | 2019-04-16 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide |
US11638622B2 (en) | 2008-06-27 | 2023-05-02 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view of articulatable instruments extending out of a distal end of an entry guide |
US10368952B2 (en) | 2008-06-27 | 2019-08-06 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
US9089256B2 (en) | 2008-06-27 | 2015-07-28 | Intuitive Surgical Operations, Inc. | Medical robotic system providing an auxiliary view including range of motion limitations for articulatable instruments extending out of a distal end of an entry guide |
JP2016101506A (en) * | 2009-03-31 | 2016-06-02 | インテュイティブ サージカル オペレーションズ, インコーポレイテッド | Synthetic representation of surgical robot |
KR20120004479A (en) * | 2009-03-31 | 2012-01-12 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | Integrated representation of surgical robots |
US10984567B2 (en) | 2009-03-31 | 2021-04-20 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
US11941734B2 (en) | 2009-03-31 | 2024-03-26 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
EP3246135A1 (en) * | 2009-03-31 | 2017-11-22 | Intuitive Surgical Operations Inc. | Synthetic representation of a surgical robot |
KR101705921B1 (en) * | 2009-03-31 | 2017-02-10 | 인튜어티브 서지컬 오퍼레이션즈 인코포레이티드 | Synthetic representation of a surgical robot |
CN102448680A (en) * | 2009-03-31 | 2012-05-09 | 直观外科手术操作公司 | Synthetic representation of a surgical robot |
US10282881B2 (en) | 2009-03-31 | 2019-05-07 | Intuitive Surgical Operations, Inc. | Rendering tool information as graphic overlays on displayed images of tools |
WO2010117685A2 (en) * | 2009-03-31 | 2010-10-14 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
WO2010117685A3 (en) * | 2009-03-31 | 2011-02-24 | Intuitive Surgical Operations, Inc. | Synthetic representation of a surgical robot |
US9492927B2 (en) | 2009-08-15 | 2016-11-15 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
US11596490B2 (en) | 2009-08-15 | 2023-03-07 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
US9084623B2 (en) | 2009-08-15 | 2015-07-21 | Intuitive Surgical Operations, Inc. | Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide |
US9956044B2 (en) | 2009-08-15 | 2018-05-01 | Intuitive Surgical Operations, Inc. | Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide |
US10959798B2 (en) | 2009-08-15 | 2021-03-30 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
US10772689B2 (en) | 2009-08-15 | 2020-09-15 | Intuitive Surgical Operations, Inc. | Controller assisted reconfiguration of an articulated instrument during movement into and out of an entry guide |
US10271915B2 (en) | 2009-08-15 | 2019-04-30 | Intuitive Surgical Operations, Inc. | Application of force feedback on an input device to urge its operator to command an articulated instrument to a preferred pose |
US9622826B2 (en) | 2010-02-12 | 2017-04-18 | Intuitive Surgical Operations, Inc. | Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument |
US10828774B2 (en) | 2010-02-12 | 2020-11-10 | Intuitive Surgical Operations, Inc. | Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument |
US10537994B2 (en) | 2010-02-12 | 2020-01-21 | Intuitive Surgical Operations, Inc. | Medical robotic system providing sensory feedback indicating a difference between a commanded state and a preferred pose of an articulated instrument |
EP2384714A1 (en) * | 2010-05-03 | 2011-11-09 | Universitat Politècnica de Catalunya | A method for defining working space limits in robotic surgery |
WO2011138289A1 (en) * | 2010-05-03 | 2011-11-10 | Universitat Politècnica De Catalunya | A method for defining working space limits in robotic surgery |
EP2704658A2 (en) * | 2011-05-05 | 2014-03-12 | The Johns Hopkins University | Method and system for analyzing a task trajectory |
EP2704658A4 (en) * | 2011-05-05 | 2014-12-03 | Univ Johns Hopkins | METHOD AND SYSTEM FOR ANALYZING A TASK PATH |
US9572481B2 (en) | 2011-05-13 | 2017-02-21 | Intuitive Surgical Operations, Inc. | Medical system with multiple operating modes for steering a medical instrument through linked body passages |
US11490793B2 (en) | 2011-05-13 | 2022-11-08 | Intuitive Surgical Operations, Inc. | Medical system with multiple operating modes for steering a medical instrument through linked body passages |
US11389255B2 (en) | 2013-02-15 | 2022-07-19 | Intuitive Surgical Operations, Inc. | Providing information of tools by filtering image areas adjacent to or on displayed images of the tools |
US10507066B2 (en) | 2013-02-15 | 2019-12-17 | Intuitive Surgical Operations, Inc. | Providing information of tools by filtering image areas adjacent to or on displayed images of the tools |
US11806102B2 (en) | 2013-02-15 | 2023-11-07 | Intuitive Surgical Operations, Inc. | Providing information of tools by filtering image areas adjacent to or on displayed images of the tools |
WO2016089753A1 (en) * | 2014-12-03 | 2016-06-09 | Gambro Lundia Ab | Medical treatment system training |
CN107205789A (en) * | 2015-06-01 | 2017-09-26 | 奥林巴斯株式会社 | Medical manipulator system |
US10201394B2 (en) | 2015-06-01 | 2019-02-12 | Olympus Corporation | Medical manipulator system |
US10335243B2 (en) | 2015-06-01 | 2019-07-02 | Olympus Corporation | Drape unit |
US10314663B2 (en) | 2015-06-01 | 2019-06-11 | Olympus Corporation | Medical overtube |
US10117716B2 (en) | 2015-06-01 | 2018-11-06 | Olympus Corporation | Medical manipulator system |
EP3595565A4 (en) * | 2017-03-15 | 2021-04-14 | Covidien LP | Robotic surgical systems, instruments, and controls |
CN110403701A (en) * | 2019-08-30 | 2019-11-05 | 山东威高手术机器人有限公司 | Input unit, minimally invasive surgical operation robot and micro-wound surgical operation control method |
Also Published As
Publication number | Publication date |
---|---|
WO2009037576A3 (en) | 2010-02-04 |
CA2684459A1 (en) | 2009-03-26 |
CA2684459C (en) | 2016-10-04 |
JP5543331B2 (en) | 2014-07-09 |
JP2010525838A (en) | 2010-07-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9131986B2 (en) | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis | |
CA2684459C (en) | Methods, devices, and systems for non-mechanically restricting and/or programming movement of a tool of a manipulator along a single axis | |
US11819301B2 (en) | Systems and methods for onscreen menus in a teleoperational medical system | |
US20220015832A1 (en) | Minimally invasive telesurgical systems with interative user interfaces for 3d operative images | |
CA2684472C (en) | Methods, devices, and systems for automated movements involving medical robots | |
US9795446B2 (en) | Systems and methods for interactive user interfaces for robotic minimally invasive surgical systems | |
CN106456251B (en) | For the system and method to imaging device and input control device recentralizing | |
Sutherland et al. | Robotics in the neurosurgical treatment of glioma | |
CN101193603B (en) | Laparoscopic ultrasound robotic surgical system | |
CN105188590A (en) | Collision avoidance during controlled movement of an image capture device and an actuatable device movable arm | |
CN113271884A (en) | System and method for integrating motion with an imaging device | |
US20240375282A1 (en) | Techniques for following commands of an input device using a constrained proxy | |
Abdurahiman et al. | Human-computer interfacing for control of angulated scopes in robotic scope assistant systems | |
WO2023192204A1 (en) | Setting and using software remote centers of motion for computer-assisted systems | |
WO2021211617A1 (en) | Automated rotation of a needle in a computer-assisted system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 08832194 Country of ref document: EP Kind code of ref document: A2 |
|
ENP | Entry into the national phase |
Ref document number: 2010503622 Country of ref document: JP Kind code of ref document: A |
|
WWE | Wipo information: entry into national phase |
Ref document number: 201559 Country of ref document: IL |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2684459 Country of ref document: CA |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2008832194 Country of ref document: EP |
|
WWE | Wipo information: entry into national phase |
Ref document number: 12596426 Country of ref document: US |