US20220009100A1 - Software Interface for Authoring Robotic Manufacturing Process - Google Patents
Software Interface for Authoring Robotic Manufacturing Process Download PDFInfo
- Publication number
- US20220009100A1 US20220009100A1 US17/379,920 US202117379920A US2022009100A1 US 20220009100 A1 US20220009100 A1 US 20220009100A1 US 202117379920 A US202117379920 A US 202117379920A US 2022009100 A1 US2022009100 A1 US 2022009100A1
- Authority
- US
- United States
- Prior art keywords
- robot
- safe zone
- tool
- node
- volume
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000004519 manufacturing process Methods 0.000 title description 30
- 230000033001 locomotion Effects 0.000 claims abstract description 185
- 238000000034 method Methods 0.000 claims abstract description 106
- 238000003860 storage Methods 0.000 claims description 9
- 230000001133 acceleration Effects 0.000 claims description 3
- 230000036461 convulsion Effects 0.000 claims description 3
- 230000001419 dependent effect Effects 0.000 claims 3
- 238000004590 computer program Methods 0.000 claims 2
- 230000009471 action Effects 0.000 abstract description 39
- 238000010276 construction Methods 0.000 abstract description 32
- 230000003993 interaction Effects 0.000 abstract description 7
- 230000008569 process Effects 0.000 description 60
- 230000006870 function Effects 0.000 description 28
- 239000000463 material Substances 0.000 description 28
- 239000012636 effector Substances 0.000 description 13
- 238000004891 communication Methods 0.000 description 12
- 230000008859 change Effects 0.000 description 9
- 238000010586 diagram Methods 0.000 description 9
- 150000001875 compounds Chemical class 0.000 description 7
- 230000005540 biological transmission Effects 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 5
- 210000000707 wrist Anatomy 0.000 description 5
- 238000013461 design Methods 0.000 description 4
- 238000012986 modification Methods 0.000 description 4
- 230000004048 modification Effects 0.000 description 4
- 238000012163 sequencing technique Methods 0.000 description 4
- 230000001360 synchronised effect Effects 0.000 description 4
- 230000009466 transformation Effects 0.000 description 4
- 238000000844 transformation Methods 0.000 description 4
- 238000012800 visualization Methods 0.000 description 4
- RTZKZFJDLAIYFH-UHFFFAOYSA-N Diethyl ether Chemical compound CCOCC RTZKZFJDLAIYFH-UHFFFAOYSA-N 0.000 description 3
- 239000003292 glue Substances 0.000 description 3
- 238000013178 mathematical model Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 238000012546 transfer Methods 0.000 description 3
- 241000254032 Acrididae Species 0.000 description 2
- XEEYBQQBJWHFJM-UHFFFAOYSA-N Iron Chemical compound [Fe] XEEYBQQBJWHFJM-UHFFFAOYSA-N 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000008520 organization Effects 0.000 description 2
- 239000013308 plastic optical fiber Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 230000002441 reversible effect Effects 0.000 description 2
- 238000003466 welding Methods 0.000 description 2
- RYGMFSIKBFXOCR-UHFFFAOYSA-N Copper Chemical compound [Cu] RYGMFSIKBFXOCR-UHFFFAOYSA-N 0.000 description 1
- 241000282806 Rhinoceros Species 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000006399 behavior Effects 0.000 description 1
- 239000004566 building material Substances 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 210000000080 chela (arthropods) Anatomy 0.000 description 1
- 239000010949 copper Substances 0.000 description 1
- 229910052802 copper Inorganic materials 0.000 description 1
- 238000005520 cutting process Methods 0.000 description 1
- 230000001934 delay Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 238000006073 displacement reaction Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 150000002170 ethers Chemical class 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000008571 general function Effects 0.000 description 1
- 238000000227 grinding Methods 0.000 description 1
- 229910052742 iron Inorganic materials 0.000 description 1
- 235000000396 iron Nutrition 0.000 description 1
- 230000007774 longterm Effects 0.000 description 1
- 238000007726 management method Methods 0.000 description 1
- 239000003550 marker Substances 0.000 description 1
- 238000005259 measurement Methods 0.000 description 1
- 238000003801 milling Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000002085 persistent effect Effects 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 230000001131 transforming effect Effects 0.000 description 1
- 238000013519 translation Methods 0.000 description 1
Images
Classifications
-
- B—PERFORMING OPERATIONS; TRANSPORTING
- B25—HAND TOOLS; PORTABLE POWER-DRIVEN TOOLS; MANIPULATORS
- B25J—MANIPULATORS; CHAMBERS PROVIDED WITH MANIPULATION DEVICES
- B25J9/00—Programme-controlled manipulators
- B25J9/16—Programme controls
- B25J9/1656—Programme controls characterised by programming, planning systems for manipulators
- B25J9/1664—Programme controls characterised by programming, planning systems for manipulators characterised by motion, path, trajectory planning
-
- G—PHYSICS
- G05—CONTROLLING; REGULATING
- G05B—CONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
- G05B2219/00—Program-control systems
- G05B2219/30—Nc systems
- G05B2219/35—Nc in input of data, input till input file format
- G05B2219/35216—Program, generate nc program, code from cad data
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10—TECHNICAL SUBJECTS COVERED BY FORMER USPC
- Y10S—TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y10S901/00—Robots
- Y10S901/30—End effector
- Y10S901/41—Tool
Definitions
- Automated manufacturing processes may involve the use of one or more robotic devices that may be used to construct an output product, such as a car, a wall, a piece of furniture, or any number of other physical fabrications.
- the robotic devices may be equipped with end-effector-mounted tools, such as a gripper or a drill, that may be used during a construction process.
- the robotic devices may be programmed with sequences of specific motion commands and commands for other operations in order to cause the robotic devices to complete a manufacturing process.
- the present disclosure provides methods and apparatuses that allow for use of a graphical interface in order to cause one or more robotic devices to construct an output product.
- input data specifying at least one motion path and at least one sequence of tool actions to occur at points within the motion path may be received.
- a number of different digital nodes may be generated within a display window.
- the digital nodes may include at least one robot node describing a robot actor.
- the digital nodes may also include at least one motion command node that determines robot motion commands and at least one tool command node that determines tool commands based on received input data.
- the at least one robotic device may then execute motion commands determined by the at least one motion command node and execute tool commands determined by the at least one tool command node in order to construct the output product.
- a method in one example, includes causing a graphical interface to be displayed on a display device.
- the method may further include receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data.
- the method may additionally include generating, based on the node configuration data, a plurality of digital nodes.
- the plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions.
- the method may further include providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- a system including at least one robot actor and a control system may be configured to cause a graphical interface to be displayed on a display device.
- the control system may be further configured to receive input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one construction path for use by the at least one robot actor, where the at least one construction path comprises at least one sequence of robot operations, where at least some of the robot operations comprise use of at least one tool, and where the input data further indicates node configuration data.
- the control system may additionally be configured to generate, based on the node configuration data, a plurality of digital nodes.
- the digital nodes may include (a) at least one robot node describing one or more attributes of a corresponding one of the at least one robot actor and (b) at least one command node that determines a sequence of robot commands to cause a robot actor corresponding to the at least one robot node to perform the at least one sequence of robot operations within the at least one construction path.
- the control system may also be configured to provide instructions for the at least one robot actor to execute the sequence of robot commands determined by the at least one command node.
- a non-transitory computer readable medium having stored therein instructions, that when executed by a computing system, cause the computing system to perform functions is disclosed.
- the functions may include causing a graphical interface to be displayed on a display device.
- the functions may further include receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data.
- the functions may additionally include generating, based on the node configuration data, a plurality of digital nodes.
- the plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions.
- the functions may further include providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- a system may include means for causing a graphical interface to be displayed on a display device.
- the system may further include means for receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data.
- the system may additionally include means for generating, based on the node configuration data, a plurality of digital nodes.
- the plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions.
- the system may further include means for providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- Figure I shows a block diagram of a manufacture control system, according to an example embodiment.
- FIG. 2A shows a view of a robot with 7 degrees of freedom, according to an example embodiment.
- FIG. 2B shows a view of a robot with an attached gripper, according to an example embodiment.
- FIG. 2C shows a view of a robot with an attached spindle, according to an example embodiment.
- FIG. 3A shows a view of a tool rack, according to an example embodiment.
- FIG. 3B shows a view of a tool rack and two robots, according to an example embodiment.
- FIG. 4A shows a graphical interface with a 3D model, according to an example embodiment.
- FIG. 4B shows additional graphical interfaces with 3D models, according to an example embodiment.
- FIG. 5 illustrates a node-based graphical interface and a visualization of a building process, according to an example embodiment.
- FIG. 6A illustrates a toolbar for a graphical interface, according to an example embodiment.
- FIG. 6B illustrates an organization of digital tools, according to an example embodiment.
- FIG. 6C is a block diagram of an example workflow, according to an example embodiment.
- FIG. 7 is a block diagram of an example method, according to an example embodiment.
- FIG. 8A illustrates a graphical interface with a digital robot actor, according to an example embodiment.
- FIG. 8B illustrates a graphical interface with a digital robot actor and a dropdown menu, according to an example embodiment.
- FIG. 8C illustrates a graphical interface with a digital robot actor and a visualization of the digital robot actor, according to an example embodiment.
- FIG. 9 is block diagram of an example node configuration, according to an example embodiment.
- Example methods and systems are described herein. Any example embodiment or feature described herein is not necessarily to be construed as preferred or advantageous over other embodiments or features.
- the example embodiments described herein are not meant to be limiting. It will be readily understood that certain aspects of the disclosed systems and methods can be arranged and combined in a wide variety of different configurations, all of which are contemplated herein.
- Example embodiments provide for a graphical interface that allows a user to configure one or more robotic devices in order to author a manufacturing process using the robotic devices.
- a user may be able to select and configure digital nodes via the graphical interface describing robot actors and other aspects of a physical workcell in order to control a building process in the physical world.
- the graphical interface may provide a layer of abstraction so that users may input motion paths and tool operations for robotic devices without having to specify particular robot commands.
- robotic devices may be represented within a graphical interface as digital robot nodes describing one or more attributes of robotic actors within a physical, workcell.
- digital robot actors may include a number of interchangeable component nodes.
- a component node may include a robot node representing different makes and models of commercial robots, along with different limitations (e.g., ranges of motion) of particular robotic devices.
- a component node may include a tool node representing different types of tools that may be used by a particular robotic device during construction, such as grippers or spindles.
- a component node may include an input/output (IO) node representing types or ranges of data input or output available to communicate with a particular robotic device.
- a component node may also include a track node representing different tracks of motion that a robot can move along within the physical workcell. Other component nodes are also possible.
- the graphical interface may enable a user to configure a stage node describing all of the available robot actors, tools, 10 types, and/or axes of motion within a physical workcell.
- the robot actors may include robots from a variety of different types and manufacturers.
- the construction tools may include a number of different tools usable by a robotic device in a building process, such as a gripper, spindle, or welding iron.
- users may configure the stage node to indicate which components such as robot actors or construction tools are currently available within a particular physical, workcell for a building process.
- users may also be able to select one or more motion command nodes which determine sequences of robot motion commands for one of the robot actors to follow based on inputted motion paths.
- a motion command node may determine robot commands for a particular robot actor in the stage that move the robot's end effector through a sequence of frames (positions and orientations) that correspond to an inputted motion path.
- a motion command node may determine a sequence of joint angles and/or joint movements that cause a robot actor to follow a particular motion path.
- a graphical interface may contain an interpolation slider that may be used to determine an interval at which robot motion commands are determined along the motion path.
- some or all of the motion paths may be determined based on input geometry. For instance, a motion path for a robot actor to follow (e.g., a sequence of positions and orientations through space) may be determined based on mathematically derived planes from parameters describing a desired output product. In some examples, one or more endpoints for one or more of the motion paths may be determined using target planes on the output product. In further examples, the graphical interface may provide a user with endpoints of motion paths and allow the user to adjust portions of the motion path between the endpoints.
- users may be able to select one or more tool command nodes that indicate certain tool actions (e.g., open a gripper) for one of the robot actors to take in order to construct the output product.
- the tool command nodes may determine a sequence of tool commands that cause a particular robot actor to execute the inputted tool actions.
- a tool command node may allow the user to configure a sequence of particular tool actions (e.g., pick up a stick, hold a stick, put down a stick, nail a stick, etc.) for the robotic device to take at certain points along its determined motion path in order to construct the output product.
- command nodes may be used to determine both motion commands and tool commands for robot actors. For instance, a construction path may be determined based on data received through a graphical interface for a particular robot actor. The construction path may indicate sequences of robot operation for the robot actor to take, which may include both motions and tool operations within a single construct. A command node may then determine robot operations for a particular robot actor to execute based on a received construction path. Other configurations are also possible.
- the set of motion commands and the set of tool commands determined by a motion command node and a tool command node may be paired up to create a step node that includes a sequence of movements and tool actions for a robot to follow to complete some action during construction (e.g., move to a stick and pick up the stick),
- export nodes may be used to export steps to a particular robot actor in order to cause the robot actor to execute the movements and tool actions within a physical workcell
- these step nodes may be reusable building blocks, within the graphical interface that allow for sequencing more complex sequences of motion.
- sync nodes may be used to cause commands from two or more step nodes to start at the same time, which may allow for multiple robotic devices to work in parallel.
- Example embodiments may provide for motion planning and control of multi-axis robotic systems for use in the manufacturing and making industries.
- Example design-to-production systems may allow users to change parameters describing an output product on the front end, with the effects propagated through to a product manufactured by one or more robotic devices using one or more tools.
- users may be provided with a graphical interface that allows for the configuration of the robot actors using a diverse toolset in order to automate the building process.
- robot motions may be abstracted so that users don't have to program specific robot commands (e.g., motion commands or tool commands) in order to control the building process. Accordingly, users may be able to design a building process without specific knowledge of commands for particular types of robots.
- users may be provided, with one or more interfaces that allow for varying amounts of control over specific robot operations within a manufacturing process, during offline motion programming and/or during runtime.
- users may be provided with a three-dimensional (3D) modeling graphical interface that allows the user to alter one or more variables describing a physical workcell and/or a desired output product that affect a building process in the physical world.
- 3D three-dimensional
- the user interface may provide abstract ways to represent physical objects digitally as nodes within a software environment.
- the user experience may enable users to select from an array of tools which can be configured and combined in a number of different ways to control different types of robot actors and hardware components within a physical workcell.
- the physical workcell may include a physical stage or stages on which a physical building process is planned or is occurring within the physical world.
- the physical workcell may include a variety of different robot actors and other hardware components as well as physical materials that may be used in the building process.
- the physical workcell may contain a tool rack and/or an automated tool changer.
- the physical workcell may contain one or more different types of sensors.
- the physical workcell may include any number of different dimensions, including platforms for particular building activities.
- Example embodiments may involve use of a manufacture control system to create automated programming of robotics arms during a building process.
- FIG. 1 describes an example manufacture control system 100 .
- Manufacture control system 100 may be part of a manufacturing environment used to control one or more robotic devices to use one or more tools to construct some output product.
- Manufacture control system 100 may comprise a master control 10 , input and feedback systems 20 , system devices 40 , and safety systems 90 . From the most basic perspective, manufacture control system 100 may function when an input system 20 provides instructions to one of system devices 40 via master control 10 .
- input and feedback systems 20 may include a database 22 , a master input 24 , a software control 26 , and an independent manual control 28 .
- database 22 may operate to provide a set of timing and position data to direct all or a portion of device actors 42 , 44 within system devices 40 . Two device actors 42 , 44 are shown in FIG. 1 , but any number of device actors could be used within manufacture control system 100 .
- database 22 may store data being created by manual or individual movement or data input related to operation and function of device actors 42 , 44 .
- Database 22 may also store data created independently of device actors 42 , 44 , such as data created using software modeling features of a software control 26 .
- a master input 24 may be any device that functions to operate all of the device actors 42 , 44 associated with a particular building process being executed by manufacture control system 100 .
- Master input 24 may function by sending input control signals to master control 10 .
- Master control 10 may then adapt the signal from master input 24 to send individual control signals to a plurality of robot actors operating as device actors 42 , 44 for a particular manufacturing process.
- every individual device of device actors 42 , 44 may be provided a control signal from master control 10 when a signal is received from master input 24 , including a signal to maintain a status quo or non-action to devices that are not operating as device actors 42 , 44 for a particular part of the manufacturing process.
- a portion of the device actors 42 , 44 connected as part of manufacture control system 100 may not be sent any signal from master control 10 as part of the operation of motion control system 100 for a particular part of the manufacturing process.
- software control 26 may act as a replacement for master input 24 in sending control signals to the plurality of device actors 42 , 44 via the master control 10 .
- software control 26 may control individual devices from among device actors 42 , 44 to control particular operations of the individual device.
- software control 26 may function to model the behavior of individual devices of device actors 42 , 44 within a virtual modeling environment representative of a physical workcell.
- software control 26 may contain a software model for an individual device, which allows control signals to be created for the device without actually sending the control signals to the device.
- the control signals may then be stored in the software control 26 , in database 22 , within a computer memory component that is part of master control 10 , or within computer memory that is part of the device of device actors 42 , 44 for which the controls are being created.
- a master control signal from software control 26 or from master input 24 may activate the control signal for the individual device to act in conjunction with other device actors 42 , 44 .
- certain devices of device actors 42 , 44 may additionally have an independent manual control 28 .
- control signals for an individual device may be created in software modeling.
- a device may have independent manual control 28 that may be used to operate a device of device actors 42 , 44 .
- the independent manual control 28 may be given input commands over time that are recorded to database 22 or a memory device of master control 10 .
- the independent manual control 28 may communicate directly with the associated device of device actors 42 , 44 .
- the independent manual control 28 may send a control signal to master control 10 , which then conveys the signal to the associated device of device actors 42 , 44 .
- the control signal may then be created either from the signal of the independent manual control 28 (e.g., a separate user interface), or from a measured feedback reading created by the operation of the associated device. Additionally, although in many situations, it may be preferable to have the independent manual control 28 actually control the associated device during control signal creation in real time, control signals may instead be created without controlling the device. For example, if input signals are expected for certain time marks, an independent manual control 28 may be operated independent of the related device, and the control operation may be recorded. Accordingly, instructions for individual device actors of device actors 42 , 44 from independent manual control may be integrated into a building process as part of manufacture control system 100 .
- master control 10 may allow for real-time control of components of a building system by providing a link between a virtual world (e.g., software control 26 ) and the physical world (e.g., a physical workcell containing device actors 42 , 44 ). Accordingly, movements of a physical robot within the physical world may be used to drive the current position of a corresponding virtual robot in real time. Similarly, movements of a virtual robot may be used to drive the current position of a physical robot in the physical world as well or instead.
- a virtual world e.g., software control 26
- the physical world e.g., a physical workcell containing device actors 42 , 44 .
- individual control signals for specific device actors may be coordinated into a single file within a memory of a master control with a common base time provided by a master clock within the master control.
- the master control may extract control signals for each device actor and provide individual control signals to 0065ach device actor at the appropriate intervals.
- the master control may maintain separate individual control signal files and timing data for different device actors, and synchronize the different control signals separately from the individual control files.
- control data for a portion of the device actors may be transferred by a master control to a memory within an associated individual device actor.
- device actors having control data within memory may receive a synchronization signal that indicates a location in a global timeline, a rate of progress through a global timeline, or both.
- Network support may also enable communications from master control 10 to one or more of system devices 40 .
- a network may comprise an EtherCAT network operating according to IEEE 1588.
- packets may be processed on the fly using a field bus memory management unit in each slave node.
- Each network node may read the data addressed to it, while the telegram is forwarded to the next device.
- input data may be inserted while the telegram passes through.
- the telegrams may only be delayed by a few nanoseconds.
- commercially available standard network interface cards or an on-board Ethernet controller can be used as a hardware interface. Using these interfaces, data transfer to the master control via direct memory access may be achieved with no CPU capacity taken up for the network access.
- the EtherCAT protocol uses an officially assigned Ether Type inside the Ethernet Frame.
- the use of this Ether Type may allow transport of control data directly within the Ethernet frame without redefining the standard Ethernet frame.
- the frame may consist of several sub-telegrams, each serving a particular memory area of the logical process images that can be up to 4 gigabytes in size. Addressing of the Ethernet terminals can be in any order because the data sequence may be independent of the physical order. Broadcast, multicast, and communication between slaves are possible.
- Transfer directly in the Ethernet frame may be used in cases where EtherCAT components are operated in the same subnet as the master controller and where the control software has direct access to the Ethernet controller. Wiring flexibility in EtherCAT may be further maximized through the choice of different cables.
- Flexible and inexpensive standard Ethernet patch cables transfer the signals optionally in Ethernet mode (100BASE-TX) or in E-Bus (LVDS) signal representation.
- Plastic optical fiber (POF) can be used in special applications for longer distances.
- the complete bandwidth of the Ethernet network, such as different fiber optics and copper cables, can be used in combination with switches or media converters.
- Fast Ethernet (100BASE-FX) or E-Bus can be selected based on distance requirements.
- EtherCAT supports an approach for synchronization with accurate alignment of distributed clocks, as described in the IEEE 1588 standard.
- distributed aligned clocks In contrast to fully synchronous communication, where synchronization quality suffers immediately in the event of a communication fault, distributed aligned clocks have a high degree of tolerance from possible fault-related delays within the communication system.
- data exchange may be completely done in hardware based on “mother” and “daughter” clocks.
- Each clock can simply and accurately determine the other clocks' run-time offset because the communication utilizes a logical and full-duplex Ethernet physical ring structure.
- the distributed clocks may be adjusted based on this value, which means that a very precise network-wide time base with a jitter of significantly less than 1 microsecond may be available.
- high-resolution distributed clocks are not only used for synchronization, but can also provide accurate information about the local timing of the data acquisition. For example, controls frequently calculate velocities from sequentially measured positions. Particularly with very short sampling times, even a small temporal jitter in the displacement measurement may lead to large step changes in velocity.
- the EtherCAT expanded data types timestamp data type, oversampling data type
- the local time may be linked to the measured value with a resolution of up to 10 ns, which is made possible by the large bandwidth offered by Ethernet. The accuracy of a velocity calculation may then no longer depend on the jitter of the communication system.
- a hot connect function may enable parts of the network to be linked and decoupled or reconfigured “on the fly”. Many applications require a change in I/O configuration during operation.
- the protocol structure of the EtherCAT system may take account these changing configurations.
- safety systems 90 may be provided for preventative safety in detecting potential collisions between device actors in modeling the motion of the actors through a global timeline. Further, such modeling through a global timeline may be used to set safety parameters for safety systems 90 . Modeling of locations and velocities of device actors through a global timeline may enable identification of unsafe zones and unsafe times in an area of a physical workcell. Such an identification may be used to set sensing triggers of object detectors that are part of an example safety system.
- a LIDAR detector may be configured to detect unexpected objects and movement within a 15 foot area of the device actor during operation, and to automatically create a safety shutdown if an object is detected.
- the LIDAR detector may be configured to create a warning signal if an object is detected in a periphery of the danger zone, and only to create a shutdown if the detected object is moving toward a potential impact zone.
- safety systems 90 may include modeling of actors and models of defined safe zones. Analysis of the motion of the actors in software control may allow a modeled safety check to see if any actor collides with a defined safe zone.
- safe zones may be defined by entry of fixed volumes of space into a software control, by image capture of a physical workcell. Safe zones may also be defined to be variable based on a detected motion, jerk, velocity, or acceleration of an object in a safe zone.
- a safe zone may be defined by input from transponder device data. For example, a transponder location device may be attached to a robotic device actor, and a safe zone defined by a distance from the transponder.
- the transponder may feed location data to software control, which may update safe zones within a software control or within a master safety control.
- software control may update safe zones within a software control or within a master safety control.
- fixed safe zones may be defined within software control, and published prior to a safety PLC within a master safety control prior to operation of a building process.
- system devices 40 may additionally include one or more sensors 46 and 48 , such as laser-based, infrared, or computer vision-based sensors.
- Master control 10 may stream data in from one or more different types of sensors located within the physical workcell. For instance, data from the sensors may reflect dimensions or other properties of parts and/or materials, within a physical workcell, as well as how the parts and/or materials are currently positioned within the real world. This data may then be streamed out to one or more robotic device actors 42 and 44 within the environment to control robotic actions, such as to accurately define a pick-up point or to adjust the pressure applied to a particular material to avoid damaging the material.
- robotic device actor 42 , 44 may be configurable to operate one or more tools for use in construction, such as spindles, grippers, drills, pincers, or welding irons.
- robotic device actors 42 , 44 may be able to switch between one or more tools during a building process using a tool rack and/or automated tool changer 50 .
- master control 10 may contain programming logic in order to automate the selection and equipping of tools from tool rack 50 .
- instructions to cause one of the robotic device actors 42 , 44 to change tools using the tool rack/tool changer 50 may come from independent manual control 28 as well or instead.
- system devices 40 including robotic device actors 42 , 44 and a tool rack/tool changer 50 will be described. Although these figures focus on the use of robotic arms, other types of device actors 42 , 44 or system devices 40 may be used in some examples as well or instead.
- FIG. 2A illustrates a robotic device actor, according to an example embodiment.
- robotic device actor 200 may include a robotic arm 202 with an end effector 204 capable of being equipped with one or more different tools.
- the robotic arm 202 may be capable of motion along six degrees of freedom, depicted in FIG. 2A as A 1 -A 6 .
- robotic device actor 200 may be further capable of motion along one or more axes AO, such as along a rail which is not shown that allows side to side movement.
- instructions may be given to position end effector 204 at a specific location, and the positions of the robotic arm 204 along A 1 -A 6 and/or of robotic device actor 200 along one or more axes AO may be calculated by a process of the related manufacture control system.
- position control of robotic device actor 200 and/or robotic arm 202 may require separate, individual settings and control commands. Robotic devices operating with fewer degrees of freedom may be used in some examples as well or instead.
- FIG. 2B illustrates robotic device actor 200 equipped with a gripper 206 .
- the gripper 206 may be placed at end effector 204 of the robotic arm 202 .
- the gripper 206 may be used for various functions during a building process, such as picking up objects or parts, moving objects or parts, holding objects or parts, and/or placing objects or parts.
- a variety of different types of grippers may be used, such as a vacuum gripper, a tumble gripper, or a passive centering gripper. Additionally, grippers with different dimensions or other properties may be used, possibly to coincide with different types of robot actors within a physical workcell.
- FIG. 2C illustrates robotic device actor 200 equipped with a spindle 208 .
- a spindle 208 may include a rotating axis for use in various functions within a building process, such as cutting materials, shaping materials, milling or routing.
- the spindle 208 could be a variety of different types, such as a grinding spindle, an electric spindle, a low-speed spindle, or a high-speed spindle.
- spindles with different dimensions or other properties may be used, depending on the different types of robot actors within a physical workcell. In some examples, other types of tools may be used by robotic device actors as well or instead.
- FIG. 3A illustrates a tool rack, according to an example embodiment.
- the tool rack may contain a number of different fabrication tools (e.g., spindles or grippers) and may be used along with an automated tool changer in order to equip robotic devices with particular tools to use within a physical workcell.
- the tool rack may contain several tool rack modules 302 , 304 , 306 , 308 positioned along a track 300 , with each of the tool rack modules 302 , 304 , 306 , 308 containing one or more particular tools.
- one or more of the tool rack modules 302 , 304 , 306 , 308 may be moveable along the track 300 .
- a tool rack module may be capable of interfacing with a slave module that allows for a particular tool to be selected from the tool rack module and then equipped onto a robotic device.
- tool rack module 302 may interface with slave module 310 and tool rack module 306 may interface with slave module 312 .
- the tool rack modules may be equipped with built-in safety sensors to minimize the risk of accidental tool fetch and drop commands.
- the tool change slave modules may include IO breakout boxes to simplify passing IO trigger signals to control tools.
- the IO breakout boxes may interface with a timing control system, such as master control 10 described with respect to FIG. 1 , that controls the robotic devices within a physical workcell. Master control 10 may be used to direct a tool change for a particular robotic device, which may be configured in advance using software control 26 and/or from independent manual control 28 during runtime.
- FIG. 3B illustrates use of the tool rack to equip robotic devices with tools, according to an example embodiment.
- a first robotic device 314 may move its end effector 316 to a position over a slave module 310 that interfaces with a tool rack module 302 of a tool rack.
- the robotic device 314 may currently be equipped with gripper 318 , and may be controlled to move to to the tool rack in order to place gripper 318 in the tool rack and equip a different tool held by tool rack module 302 .
- a second robotic device 320 may have positioned its end effector 322 on slave module 312 in order to equip spindle 324 , which may have been held by slave module 312 .
- robotic device 320 may then proceed to move away from the tool rack and complete operations using the spindle 324 .
- the tool rack modules may be positioned on the tool rack so that multiple robotic devices may equip or change tools at the same time.
- additional rack modules 304 , 308 may contain additional tools that may be equipped by one or more robotic devices.
- instructions from a control system may be used in order to instruct a robotic device how to equip a tool during runtime (e.g., to determine where a tool is within the tool rack and solve an end effector problem in real time in order to position the end effector over a slave module to enable the robotic device to pick up the tool).
- a drive system e.g., a VFD used to supply power drive a spindle
- FIG. 4A shows a graphical interface containing a 3D model, according to an example embodiment.
- a graphical interface 400 may contain an input, window 402 which may allow a user to enter parameters relating to an output product 406 , such as a wall built using individual sticks.
- the input window 402 may allow the user to enter parameters 404 that may relate to aspects of the output product, including dimensions, density, curvature properties, other geometric properties, materials to be used, and/or other numeric inputs.
- the inputs may be used to derive a parametric solution for an output product 406 . Additionally, the inputs may be used to generate a sketch of the output product 406 within a display window 408 of the graphical interface 400 .
- FIG. 4B shows three different output products based on different user input parameters, according to an example embodiment.
- a first view of the graphical interface 440 may contain an input window 402 and a display window 408 .
- the input window 402 may allow a user to enter parameters 404 relating to a desired output product, including product dimensions, density, curve offsets, amount or type of curvatures, and/or other geometric or numerical inputs.
- parameters 404 relating to a desired output product, including product dimensions, density, curve offsets, amount or type of curvatures, and/or other geometric or numerical inputs.
- a geometric representation of the output product 406 may be displayed within the display window 408 .
- a user may modify individual parameters 404 in order to change aspects of the output product 406 .
- a second view of the graphical interface 450 shows a different output product 406 within the display window 408 based on different input parameters 404 within the input window 402 .
- dimensions of the output product 406 and/or materials used to produce the output product 406 may be modified to produce an output product 406 with a greater height as shown in the second view 450 .
- a third view 460 shows another different output product 406 within the display window 408 based on different input parameters 404 within the input window 402 .
- parameters relating to the curvature of the output product may be modified by a user in order to produce another different output product 406 as shown in the third view 460 .
- FIG. 5 shows a graphical interface for architecting a robotic building process, according to an example embodiment
- the graphical interface may be part of software control 26 as described above with respect to FIG. 1 .
- a graphical interface 500 may contain an input window 502 which allows a user to control aspects of the building process, including nodes related to robot actors, tools, motion paths, and tool operations for use during construction.
- the graphical interface 500 may additionally contain a display window 510 which contains a 3D geometric view of the physical workcell, including components such as robot actors, tools, materials, and/or constructed output products.
- the input window 502 may provide a visual programming interface or different type of interface that may allow a user to enter parameters describing a desired output product and/or information about the actors and tools to be used in the building process.
- Input data collected using the input window 502 may be used to control geometry and/or other aspects of the physical workcell displayed within the display window 510 .
- a user may input parameters to control a building process using an input window 502 containing a visual programming interface, such as an interface built using a visual programming language, such as a commercial software program known as Grasshopper.
- the interface may allow a user to include one or more nodes 504 which may represent components of the building process, such as robot nodes representing different types and/or configurations of robots, tool nodes representing different types and/or configurations of tools, IO nodes representing types of available IO, track nodes representing possible tracks of motion of robot actors, and command nodes for determining motion commands and other types of commands for robot actors.
- individual nodes 504 may be connected together using connectors 506 .
- a connector 506 between two nodes may indicate that the output of a first node is to be used as an input to a second node.
- a single robot node may receive as inputs information from several different component nodes, such as nodes representing the type of robot, the type of tool used by the robot, a track the robot can travel along, and so on.
- the window 502 of FIG. 5 may contain a timeline 508 .
- the timeline 508 may have a cursor representing a current timestamp (e.g., 83 as shown in the figure) which may represent a particular point in time of the manufacturing process.
- the timeline 508 may contain buttons to play through the building process at a particular speed, or fast-forward or rewind through the building process.
- the timeline 508 may be used to control the point in time at which the geometry and/or other aspects of the physical workcell are displayed within the display window 510 . Further, the timeline 508 may be used to indicate a particular point in time either for purposes of simulating the building process or for visualizing within software an actual physical building process taking place within the physical world.
- the user interface may additionally contain a display window 510 which may display geometry and/or other aspects of the physical workcell based on inputs from the input window 502 .
- the display window 510 may include geometry relating to robot actors, tools, building materials, robotic motion paths, and output products, among other things.
- the display window 510 may be designed using a commercial 3D modeling software, such as Rhinoceros, as shown within FIG. 5 .
- the display window 510 may display geometry within a particular physical workcell 512 .
- the display window 510 may include options to change the perspective of the physical workcell 512 and/or to zoom in or zoom out a view of the physical workcell 512 .
- the physical workcell 512 may include one or more robot actors 514 .
- the robot actors 514 may be device actors 42 and/or 44 as described above with respect to FIG. 1 and/or robotic device 200 as described with respect to FIGS. 2A-2C .
- Support may be provided for numerous different types of multi-axis robotic systems of different types and/or from different manufacturers.
- one or more of the robot actors 514 may be traditional six-axis robots.
- other types of robots that may be configured to operate along fewer or more axes may be included for use within the physical workcell 512 in addition or instead.
- robot actors may be represented within a software interface as robot nodes, which may be put together from a number of interchangeable component nodes, including robot nodes representing different makes and models of commercial robots, tool nodes representing different types of physical tools that may be used for construction such as grippers or spindles, IO nodes representing different types 10 available to communicate with a robot actor and track nodes representing different types of axes that a robot can move along.
- individual tools and/or tooling parameters can be abstracted into components that can be assembled by a user into compound tools as well.
- the display window 510 may additionally contain one or more motion paths 516 representing paths of motion of individual robot actors 514 .
- the motion paths 516 may indicate paths to be taken by the robot actors 514 during the building process, such as to pick up materials and attach them to an object under construction.
- the motion paths 516 may further indicate points at which particular input or output actions will occur. For instance, an “x” on a motion path 516 may indicate a point at which a robot actor 514 uses a tool such as a gripper to pick up a particular type of material.
- the motion paths 516 may be synchronized with the timeline 508 from the input window 502 . Accordingly, in some examples, the robot actors 514 may be made to move along the motion paths 516 to positions at particular points in time based on the timestamp indicated by the timeline 508 .
- the physical workcell 512 may additionally contain one or more materials 518 to be used during the building process.
- the materials 518 consist of sticks used to construct a wall 520 .
- Motion paths 516 may be determined for the robot actor 514 to take in order to move the individual sticks 518 onto the wall 520 .
- connective materials such as glue
- the physical workcell 512 may also contain other components not shown in FIG. 5 that may be used in the building process.
- one or more sensors may be included to sense information about the robot actors and/or materials in the physical workcell in order to influence motion paths taken by the robot actors.
- a torque sensor may be used to determine if a particular piece of material is likely to break under stress.
- a control system such as master control 10 described above with respect to FIG. 1 , may be used to interface with the robot actors and/or sensors within the physical workcell.
- the display window 510 may provide users with multiple 3D views of the physical workcell, and may allow a user to change the orientation and/or zoom of a particular view. In other examples, the display window 510 may present other types of representations of the physical workcell, such as numerical representations, as well or instead.
- an input window may additionally contain a toolbar containing digital tools to control aspects of the building process.
- FIG. 6A shows a toolbar for a graphical interface, according to an example embodiment.
- the toolbar 602 may be equipped ⁇ -Vith a variety of different toolsets 604 that may be used to design or control a building process within an input window of a graphical interface.
- Toolsets 604 may be provided with digital tools relating to generating robot motion paths, transforming between different planes or axes, describing robot actors, describing physical building tools, sequencing individual robot motions, communicating data input and/or output to and/or from robot actors, mapping between a virtual software environment and a physical workcell, and/or enabling visualization of a building process, for example.
- FIG. 6B shows an organization of digital tools within a tool bar, according to an example embodiment.
- the digital tools may be divided into a number of different categories.
- the digital tools may then be used in combination to design a building process, as shown by FIG. 6C .
- FIG. 6C is a block diagram of an example workflow, according to an example embodiment.
- FIG. 6C shows workflow involving a number of digital tools, which may be accessible, within a toolbar as depicted in FIG. 6 i and FIG. 6B or by another means within a graphical interface.
- the digital tools may be divided into a number of different categories.
- One or more digital tools from a number of different categories may be selected by a user to affect particular aspects of the building process, including the robot actors and other components within a physical workcell that may be used in the process.
- a toolbar may include path tools 608 relating to generating target planes that may be used to determine motion paths of robot actors.
- the path tools 608 may take as input geometry 606 describing a desired output product, such as geometry generated by parametlic modeling software, Grasshopper.
- the output product geometry 606 may be generated based on user input within an input window specifying aspects of the output geometry, including dimensions, density, curvature, materials, and so on.
- the path tools 608 may then determine target planes for robot motion paths based on the output product geometry 606 .
- the output product geometry 606 may include particular splines, surfaces, and/or other geometrical constructions to be included within an output product.
- the path tools 608 may then provide shortcuts for generating target planes relating to aspects of the output product in a format that can be turned into particular robot trajectories that may be used to construct an output product containing the target planes. Motion paths for individual robot actors may then be derived as a function of the target planes in addition to tool definitions and material properties, for example.
- a toolbar may include transform tools 610 relating to transformations between different axis frames or offsets, as shown by FIG. 6B and FIG. 6C .
- the transform tools 610 may provide transformations between coordinate frames at the base or joints of a particular robot and a stage containing the output product.
- the transform tools 610 may additionally allow for transformations between multiple robots operating within different frames of reference as well. As shown m FIG. 6C , transformations may be applied before and/or after determining sequences of motion for individual robot actors.
- a toolbar may include stage tools 612 that represent aspects of a physical workcell, such as robot actors, tools, IO, and/or axes.
- stage tools 612 may also provide a mapping between virtual robots in software and physical robot actors within the physical workcell, as shown by FIG. 6B and FIG. 6C .
- the stage tools 612 may be used by engine node 624 to send trajectories for robot actors to take based on output from command tools 622 to a control system 628 .
- stage node 612 may be configured in order to specify the currently available robotic devices and/or tools within a particular physical workcell.
- the control system 626 may then command robot actors and/or other components within the physical world 630 based on information from stage tools 612 .
- stage tools 612 may take input from one or more robot nodes 614 representing attributes of individual robot actors within a physical workcell, as shown by FIG. 6B and FIG. 6C .
- a robot node 614 may be used to define attributes of a robot actor, such as traditional six-axis robots or other types of robots.
- the robot attributes may include link lengths of the robot and/or arm lengths of the robot, offsets of the robot and/or joints of the robot, and/or limits on robot joint movement or maximum torque that a robot joint can handle, for example.
- stage tools 612 may also take input from one or more tool nodes 616 for defining the attributes of physical building tools and/or a tool rack for holding the tools, as shown by FIG. 6B and FIG. 6C .
- attributes of building tools such as grippers or spindles may be specified by tool nodes, which may be used to configure an automatic tool changer so that robot actors can easily switch between tools.
- robot actors may switch between tools using a tool rack which stores the tools and facilitates a tool changing process, as described above with respect to FIGS. 3A and 3B .
- the tool nodes 616 may include support for compound tooling that may allow component tools to be assembled into compound tools.
- individual tooling parameters e.g., wrist mount offsets or tool center points
- individual tooling parameters may be abstracted into components that may be assembled into compound tools.
- multiple tools may be aggregated into one compound tool that has multiple tool orientations and/or center points that may be used depending on which component of the tool may be required at a particular time.
- a compound tool with an automatic tool changer may be represented by a master interface plus a number of different attachments, such as a spindle, a vacuum array, or a gripper.
- a compound tool may include a series of different tools, such as a gripper plus a tensioner plus a roller.
- Other examples of combining multiple tools and/or combining tools by abstracting tooling into parameters that define tool orientation and/or center points are also possible.
- stage tools 612 may also take input from one or more 10 nodes 618 .
- the IO nodes 618 may describe information relating to digital and/or analog input and/or output actions that may be taken by a robot actor, such as the type of action (e.g., gripping a material) and attributes associated with the action (e.g., width of material that can be gripped).
- the robot attributes may include one or more axis nodes 620 .
- the axis nodes 620 may describe one or more linear and/or rotational axes along which a robot can travel, including limitations on the robot's movements along the axes.
- a toolbar may include command tools 622 , as shown by FIGS. 6B and 6C .
- the command tools 622 may be used to determine robot commands to cause one or more of the robot actors to execute particular operations, which may include point-to-point motions, motions along external axes, and/or input or output events.
- one of command tools 622 may be used to direct a particular robot motion along one of six degrees of freedom, a particular robot motion along an external axis, or a particular input or output event, such as applying glue to a material in a particular manner.
- command tools 622 may be included for creating step nodes that instruct robot actors to take a particular sequence motion steps and carry out one or more tool actions.
- coordinated sequences of motions may be generated for multiple robot actors working together within a single timeframe.
- step nodes can be abstracted into reusable subroutines.
- a subroutine can be defined by connecting visual building blocks, which may represent particular motion commands or motion parameters. The subroutine could then be used to make one or more robots carry out the same action sequence multiple times within a single building process.
- steps can be synchronized across multiple robots so that multiple robots can, work in a shared environment simultaneously.
- Example systems may also include an engine node 624 , which may assign each of the steps to particular robotic devices within a stage.
- timeline 508 as illustrated and described with respect to FIG. 5 may also includes buttons to skip between steps on the timeline.
- digital bookmarks may be inserted by a user for particular steps. For instance, through the graphical interface, it may be possible to jump from the beginning of a “fetch stick” step to the beginning of a “nail stick” step. These bookmarks steps within the timeline may match the steps authored by the user by inputting motion commands, 10 commands, and/or other commands in a step node.
- the engine node 624 may communicate with control system 626 .
- the control system 626 may be a computing device capable of communicating wirelessly with robot actors and/or other components such as sensors within the physical, workcell in the physical world 630 .
- the control system 626 may provide access to real time data streams from all robot actors and devices, which may allow for precise control over the physical workcell at particular points in time.
- the control system could communicate with some or all of the actors or devices through wired connections or other types of communication channels as well or instead, including previously described network protocols.
- control system may additionally contain a physical control interface such as a touchscreen interface that may allow a user to interact with the control system to view live data or modify robot actions in real time.
- a stage file containing information about the physical workcell including actors, tools, materials, and environmental setup on the control system 626 may be accessible via a programming interface.
- a user who is watching a building process within the physical world 630 may then make modifications to the process before it is completed.
- a toolbar may include data input/output tools 628 that may allow the control system 626 to send and/or receive data to and/or from the virtual software environment that determines robot motion paths, as shown by FIG. 6B and FIG. 6C .
- telemetry from the control system 626 may be used to create a live link between the virtual world in software and the physical world 630 .
- the data input/output tools 628 may be used to process information from the control system 626 relating to the robot actors within the physical workcell and/or other components in the workcell such as sensors. Based on this information about the physical world 630 , the virtual robots within software may be updated with real-time feedback from the physical world 630 (e.g., motion paths for robot actors may be determined or modified based on real-time sensor data).
- the data input/output tools 628 may be used to send data back to the control system 626 so that the control system 626 can effectuate particular input or output actions within the physical world 630 , for example.
- the control system 626 may instruct a robot actor how use a tool in the physical world 630 (e.g., how to control a spindle) based on information from one or more digital tools within the software interface.
- engine node 624 include visualizer or simulation tools that may allow a user to simulate a building process through a user interface in software, as shown by FIG. 6B and FIG. 6C .
- the visualizer tools may display the building process as geometry drawn on a screen that shows the physical workcell.
- the visualizer tools may display the building process as curves representing particular data values as well or instead.
- the visualizer tools may also be used to visualize a building process in software as it is actually occurring within the physical world 630 .
- the visualizer tools may additionally provide a graphical representation of potential conflicts within a particular building process, such as when a robot's motion path extends outside its possible range of motion or when two robot actors may be going to collide based on the currently defined trajectories and/or step sequences.
- the visualizer component may allow a user to see simulations of the building process in advance and/or as the building takes place.
- the user may use the visualizer component offline to see robotic motion paths as well as input/output events over a series of sequential steps as geometry drawn within a viewing window.
- the user may be able to visualize a simulated playback as numerical data streams relating to the robot actors, materials, and/or other aspects of the physical workcell represented through curves in addition to or instead of visual geometry.
- the user may also be able to see particular data points at individual timesteps, such as robotic joint values, axis values, or input/output values.
- a user may also be able to use the visualizer component to visualize a building process that is occurring in the physical world in real time.
- the system may interface with a control system that receives real-time data streams from sensors that may be used to scan the physical workcell, individual robot actors, and/or parts used in construction as an output product is being built.
- the visualizer's user interfaces may be updated in real time to reflect real world dimensions, properties, and/or positions of objects and actors within the environment
- a method 700 is provided for causing one or more robot actors to construct an output product, according to an example embodiment.
- method 700 may be carried out by a control system, such as manufacture control system 100 , master control 10 , and/or software control 26 as described in reference to FIG. 1 .
- the control system may communicate with the robot actors using any of the network protocols or communication methods previously described.
- part or all of method 700 may be executed by any of the graphical interfaces or systems described and illustrated with respect to FIGS. 4A-4B, 5, and 6A-6C .
- part or all of method 700 may be carried out by one or more robotic devices, such as device actors 42 , 44 within system devices 40 as described in reference to FIG. 1 , or device actor 200 as illustrated and described in reference to FIGS. 2A-2C .
- robotic devices such as device actors 42 , 44 within system devices 40 as described in reference to FIG. 1 , or device actor 200 as illustrated and described in reference to FIGS. 2A-2C .
- various alternative embodiments may include any number and type of
- program code can be stored on any type of computer-readable medium, for example, such as a storage device including a disk or hard drive.
- each block of the flowchart shown in FIG. 7 may represent circuitry that is wired to perform the specific logical functions in the process. Unless specifically indicated, functions in the flowchart shown in FIG. 7 may be executed out of order from that shown or discussed, including substantially concurrent execution of separately described functions, or even in reverse order in some examples, depending on the functionality involved, so long as the overall functionality of the described method is maintained.
- method 700 may initially involve causing a graphical interface to be displayed on a display device.
- a graphical interface may be provided that allows for configuration of one or more robot actors for use during a building process.
- the graphical interface may include a display window containing a 3D model of a physical workcell, as described and illustrated with respect to FIG. 4A-4B .
- the graphical interface may additionally include an input window containing a configurable node-based interface as described with respect to FIG. 5 .
- Other types or variations of graphical interfaces could be used as well.
- the display window and the input window could be combined into a single graphical interface and/or divided into multiple additional sub-windows.
- the graphical interface could be displayed on a PC workstation, a touch-surface device, or on a number of other different types of computing devices.
- method 700 may additionally involve receiving input data corresponding to one or more interactions with the graphical interface.
- Users may interact with the graphical interface in a number of different ways.
- a user may be provided with one or more menus or toolbars that allow for configuration of aspects of a physical workcell and/or automated building process.
- users may be able to select and position digital nodes, which may be represented as shapes, icons, text, modules, or as other graphical representations on a display.
- the user may additionally be able to connect two or more nodes together (e.g., by using a pointing device or touch-surface interface) in order to indicate a logical relationship between the nodes (e.g., the output of one node may act as input for a second node).
- data may be received through the graphical interface in other ways. For instance, a user may be able to type or otherwise enter numerical input or other types of input relating to certain aspects of a building process as well.
- user may be able to interact with geometry in a 3D modeling display (e.g., by drawing a line between two positions in a modeling environment).
- the received input data may indicate at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path.
- a motion path may indicate a path of motion for an end effector of a robot actor to follow within the physical workcell.
- particular tool operations may be specified, such as to pick up an object, place an object, use a drill, or apply glue to a surface.
- An example motion path and sequence of tool actions is illustrated and described above with respect to FIG. 5 .
- a motion path may be defined by a sequence of frames, or planes with six degrees of freedom in Cartesian space.
- the sequence of frames may indicate a path for an end effector of a robot actor to follow in order to get from one point within the physical workcell to another point (e.g., to pick up a stick and then move to place the stick on a wall).
- a user may enter a motion path through the graphical interface by drawing a line from one point in a 3D display window to a second point, representing two points within the physical workcell.
- Motion paths may also be entered in other ways in some examples as well, such as by entering numerical data representing particular Cartesian coordinates within the physical workcell.
- a user may be provided with path tools, as described with respect to FIGS. 6A-6C , in order to assist the user in inputting motion paths through the graphical interface.
- a user may first determine a model of a desired output product, as described with respect to FIGS. 4A-4B . Based on parameters entered to describe the output product, mathematical models may be generated in order to represent the geometry of the output product as splines or surfaces.
- Path tools may be provided that determine target frames on the output product based on the mathematical models. In some examples, the target frames may then be used in order to assist the user in determining motion paths for robot actors to follow.
- a target frame may indicate a particular frame (e.g., position and orientation) in which a robot's end effector can be positioned in order to place a stick at a particular point on a stick wall.
- the user may be able to manipulate the motion path in between particular target frames.
- the received input data may further indicate node configuration data.
- node configuration data may indicate how to represent particular nodes within a display window.
- node configuration data may be received based on menu selections, commands from a pointing device (e.g., nodes may be drag and dropped), and/or through other means of user interaction.
- method 700 may further involve generating, based on the node configuration data, a plurality of digital nodes.
- a configuration of digital nodes may be determined and displayed within a display window.
- FIGS. 8A-8C collectively show an example configuration of digital nodes, according to an example embodiment. The simplified examples shown in FIGS. 8A-8C and elsewhere are meant for demonstration purposes and should not be viewed as limiting. In particular, other types and configurations of nodes besides those specifically illustrated here may be used as well or instead in some examples.
- a robot node may include any of the attributes previously described with respect to FIG. 6A-6C .
- An example robot node is shown in FIG. 8A .
- a robot node 806 may be generated and displayed within an input window 802 .
- the robot node 806 may be used to define attributes of a robot actor, such as traditional six-axis robots or other types of robots.
- robot node 806 may describe kinematic system characteristics of a particular make and model of robot, including robot attributes such as link lengths of the robot and/or arm lengths of the robot, offsets of the robot and/or joints of the robot, and/or limits on robot joint movement or maximum torque that a robot joint can handle, for example.
- robot node 806 may also include information relating to tool actions or other types of actions that may be taken by a robot actor, such as the type of action (e.g., gripping a material) and attributes associated with the action (e.g., width of material that can be gripped).
- one or more robot nodes may connected to a stage node 804 , which may describe available robot actors as well as other aspects of a physical workcell, including tools, IO, and/or axes. Using a stage node 804 , tools may be shared between robot actors and/or swapped on a particular robot actor at different points of a sequence of robot operations.
- a digital robot actor may further include input from one or more component nodes via stage node 804 .
- a component node may be a tool node 808 , which may indicate one or more tools (e.g., a spindle or grip gun) which may be operable by the robotic device represented by robot node 806 .
- a component node may be an IO node 810 , which may indicate one or more types of IO (e.g., a VFD switch or solanoid trigger) which may be operable by the robotic device represented by robot node 806 .
- another component node may be an axis node 812 that may include one or more linear and/or rotational axes along which the robot can travel, including limitations on the robot's movements along the axes. Other configurations of a robot node and other digital component nodes are also possible.
- a robot node and/or one or more component nodes may be configurable using a dropdown menu, as shown by FIG. 8B .
- a dropdown menu 814 may contain preconfigured options for particular types of robots, axes of motion, IO types, and/or tools (e.g., that coincide with particular types of industrial robots).
- one or more of robot node 806 , tool node 808 , IO node 810 , and/or axis node 812 may selected or configured using dropdown menu 814 .
- custom robots with custom arrangements of robot type, tools, IO, and/or axes may be designed and then saved for later use.
- a 3D model of a robotic device may be displayed corresponding to a particular digital robot actor.
- a digital robot actor may be generated within input window 802 via stage node 804 corresponding to robot node 806 , tool node 808 , IO node 810 , and axis node 812 .
- a 3D model of a robotic device 816 corresponding to the digital robot actor may be displayed. Parameters of robotic device 816 may be determined based on configurations of robot node 806 , tool node 808 , IO node 810 , and/or track node 812 .
- the type of robotic device 816 may be determined based on robot node 806 .
- a spindle 818 may be shcn.vn as an end-effector-mounted tool on the robotic device 816 based on tool node 808 . Accordingly, using display window 814 , a user may be able to visualize how a configured robot actor will look within a physical workcell.
- the digital nodes generated based on the node configuration data may include at least one motion command node.
- a motion command node may determine a sequence of robot motion commands for a particular robot actor in order to cause the robot actor to move along a received motion path.
- a motion command node may take as input a robot node and a motion path. The motion command node may then determine motion commands for the robotic device corresponding to the robot node in order to cause the robotic device (or an end effector of the robotic device) to move along the motion path.
- a user may specify a motion path in Cartesian space (e.g., XYZ coordinates in the physical workcell). For instance, a user may draw a line within a display
- the user may specify the position within the physical workcell for the end effector of the robotic device to move to (e.g., by specifying XYZ coordinates) as well as an angle of rotation with which to use an end-effector-mounted tool, such as a gripper (e.g., by specifying Euler angles).
- a user may be assisted in determining motion paths based on generated mathematical models, as previously described.
- a motion command node may determine how to command a particular robot, by specifying the joint angles and/or changes in joint angles at each of the robot's joints. For example, a typical industrial robot has six joints: three in the wrist for rotation, and three at the shoulder and elbow for positioning the wrist in Cartesian space. Because the robot has six joints, it can position an end-effector-mounted tool at an arbitrary six degree-of-freedom Cartesian pose (XYZ position and 3 Euler angles), within some limitations. The translation from Cartesian space to joint angles is called inverse kinematics (IK).
- IK inverse kinematics
- a motion command node may be referred to as an IK command node that solves the inverse kinematics problem in order to determine motion commands for a particular robot.
- the motion command node may take into account limitations and ranges of motion of different types of robotic devices in determining robot motion commands.
- a motion command node or IK command node may select from a number of different possible solutions to the inverse kinematics problem.
- an IK command node may select a solution by evaluating a history of previous IK solutions to find solutions that provided desired results.
- a user may optionally specify particular aspects of a solution, such as whether the robot elbow should be positioned up or down, or the robot wrist should be positioned up or down.
- an IK command node may allow the user to choose whether determined robot motions will be point-to-point or linear.
- Point-to-point motions may allow a robot some degree of freedom in moving between interpolation points while maximizing speed and/or smoothness. Accordingly, point-to-point motions may be curvilinear by allowing a robot to rotate its joints in between interpolation points.
- linear motion may be used to force a robot to move in straight lines in between interpolation points.
- a separate parameter may be included to cause a control system to interpolate between motions continuously.
- a user may be able to specify through the graphical interface an interval at which the motion command node determines robot motion commands for a given motion path.
- an interval may be chosen for more critical functions within a building process.
- the interval may be chosen using an interpolation slider, such that the interval changes for one or more motion command nodes based on the position of a graphical marker on a slider
- the interval may be selected in other ways in some examples as well.
- At least one tool command node may be generated based on the node configuration data.
- a tool command node may determine a sequence of tool commands in order to cause a robot actor to operate a tool according to the received sequence of tool actions. For instance, the user may specify through the graphical interface one or more points within a motion path where certain tool operations are to be performed (e.g., pick up a stick at one point and place the stick at a second point).
- a tool command node may then determine tool commands for a particular robot actor in order to make the robot actor perform those tool operations at the specified locations.
- a tool command node may take into account parameters describing the robot actor in order to determine particular commands for the robot actor to follow. For instance, a robot actor may be configured to receive certain types of input signals in order to activate end-effector-mounted tools. A tool command node may determine a sequence of digital or analog signals that will cause a particular robot actor to activate a particular physical tool in the specified manner at specified points within a motion path. In some examples, 10 values describing an aspect of tool operation (e.g., a frequency of a spindle) may be specified within the sequence of tool actions. In such examples, a tool command node may determine tool commands that cause a particular robot actor to operate a tool with the specified IO values.
- the tool command node may determine analog signals to command the robot actor to operate a spindle with particular specified frequencies at different points within a motion path.
- the tool command node may determine digital signals to command a robot actor to open and close a gripper at different points within a motion path.
- other types of tool commands or signals to activate physical tools at particular points could be determined by tool command nodes as well.
- method 700 may also involve providing instructions for the at least one robot actor to construct the output product
- a robot actor may be controlled to move based on the sequence of robot motion commands determined by a motion command node.
- the robot actor may be controlled to execute tool commands at particular points within the sequence of robot motion commands based on the sequence of tool commands determined by a tool command node.
- a step sequencer may be provided in order to sequence together motion commands and tool commands before exporting the commands to a robot actor to execute within the physical workcell.
- FIG. 9 is block diagram of an example node configuration, according to an example embodiment.
- FIG. 9 shows the generation of step nodes in order to sequence together robot motion commands and tool commands before exporting the commands to a robot actor.
- Output geometry 902 may first be determined describing a desired output product.
- a user may then determine motion paths and tool actions 904 for a robot actor to use in construction of output product based on output geometry 902 .
- the robot actor including any relevant parameters of the robotic device, tools, IO, or tracks, may be described using robot node 906 , tool node 908 , IO node 910 , and axis node 912 .
- Each of these component nodes may be input into stage node 914 , which may describe all of the available components within a particular physical workcell.
- the motion paths and tool actions 904 may be input into a motion command node 916 , along with robot node 906 and tool node 908 .
- motion command node 916 may determine a sequence of robot motion commands in order to cause a robotic device described by robot node 906 and tool node 908 to move its end-effector-mounted tool through the inputted motion paths 904 .
- the motion paths and tool actions 904 and IO node 910 may be input into a tool command node, or IO command node 918 .
- IO command node 918 may determine a sequence of IO triggers to cause the robotic device described by robot node 906 to execute tool operations at certain points along the motion path of the tool based on the inputted motion paths and tool actions 904 .
- axis node 912 may be input into axis command node 920 in order to determine a sequence of axis commands to cause the robotic device described by robot node 906 to move along an axis within the physical workcell described by axis node 912 .
- one or more robotic devices may be stationary within a physical workcell.
- the output from motion command node 916 , IO command node 918 , and axis command node 920 may combined within a step node 922 .
- the step node 922 may interleave the sequence of robot motion commands and tool commands in order to create a step that defines a step of a building process within the physical workcell.
- a step may indicate how the robotic device corresponding to robot node 906 may perform a particular function within a modeling process, such as moving to a pile of sticks, picking up a stick, moving to a stick wall under construction, and placing the stick on the stick wall.
- a duplicate step node 928 may be generated within a graphical interface.
- a user may be able to copy step node 922 in order to cause a robotic device to repeat particular operations or to cause a different robotic device to execute the operations.
- step nodes such as step node 922 may function as reusable building blocks within a graphical interface that may allow a user to design and configure more complex sequences of motion, possibly sequencing together a number of different steps before exporting to a robotic device for execution.
- output from the step node 922 and duplicate step node 928 may be input into engine node 924 .
- Engine node 924 may compile multiple steps to be performed by one or more robotic actors, and may additionally provide visualization and/or simulation functionality.
- stage node 914 may be defined by a user, describing each of the robotic devices, tools, IO, and/or axes available within a particular physical workcell. For instance, a user may configure stage node 914 to indicate the devices and/or components currently available for a building process. Then, the engine node 924 may interface with the stage node 914 in order to determine particular robot actors (including robotic devices, tools, IO, and/or axes) to perform particular steps within the building process.
- the stage node 914 may define other characteristics of a physical workcell as well, such as available sensors, materials for construction, or other hardware components.
- Output from engine node 924 may be input to an export node 926 , which may be used to send instructions to a robotic device within the physical workcell. Communication between a control system and a robotic device may be achieved using any of the methods previously described.
- the export node 926 may allow a user to select a particular robotic device to cause to execute the determined step operations.
- the export node 926 may contain a dropdown menu that allows for the selection of a robotic device from one or more robotic devices within a particular physical workcell.
- a sync node may be used to connect two or more steps by indicating that commands from each of the step nodes should start at the same time. Accordingly, each of the robotic devices may be prevented from executing commands within one of the step nodes attached to a sync node until all of the robotic devices are available and ready to begin executing commands specified by corresponding step nodes. Robot operations may be synchronized between different robot actors in different ways as well.
- data may be received through the graphical interface indicating at least one construction path for use by at least one robot actor.
- the construction path may include at least one sequence of robot operations, where at least some of the robot operations include use of a tool.
- a user may be able to specify both a motion path for a robot actor to follow as well as tool operations occurring along the motion path as a single construct.
- the user may be able to specify a construction path by drawing a line through a modeling environment with an “x” at particular points indicating tool operations, as illustrated and described with respect to FIG. 5 .
- At least one command node may be generated in order to cause a robot actor to perform robot operations within the construction path.
- a single type of command node may determine both robot motion commands and tool commands in order to cause a particular robot actor corresponding to a robot node to perform operations within the construction path.
- the command node may solve an inverse kinematics problem to determine particular robot joint angles or movements and also determine how to command the robot to perform particular tool actions at points within the received construction path.
- the command node may also perform a sequencing function in order to sequence motion commands with tool commands into a single sequence of robot commands.
- one or more of the command nodes may be conditional command nodes.
- a conditional command node may determine robot operations for a particular robot actor to follow based on sensor data received from one or more sensors within a physical workcell. For instance, a conditional command node may contain a conditional statement that will cause a robotic actor to only use a piece in construction if it will fit within a certain area. Further, sensor data received during construction may indicate the shape of a particular piece. Based on the sensor data, the conditional command node may determine whether or not to output robot operations to pick up and use the particular piece during the building process. Other types of conditional command nodes are also possible.
- each block and/or communication may represent a processing of information and/or a transmission of information in accordance with example embodiments.
- Alternative embodiments are included within the scope of these example embodiments.
- functions described as blocks, transmissions, communications, requests, responses, and/or messages may be executed out of order from that shown or discussed, including substantially concurrent or in reverse order, depending on the functionality involved.
- more or fewer blocks and/or functions may be used with any of the ladder diagrams, scenarios, and flow charts discussed herein, and these ladder diagrams, scenarios, and flow charts may be combined with one another, in part or in whole.
- a block that represents a processing of information may correspond to circuitry that can be configured to perform the specific logical functions of a herein-described method or technique.
- a block that represents a processing of information may correspond to a module, a segment, or a portion of program code (including related data).
- the program code may include one or more instructions executable by a processor for implementing specific logical functions or actions in the method or technique.
- the program code and/or related data may be stored on any type of computer readable medium such as a storage device including a disk or hard drive or other storage medium.
- the computer readable medium may also include non-transitory computer readable media such as computer-readable media that stores data for short periods of time like register memory, processor cache, and random access memory (RAM).
- the computer readable media may also include non-transitory computer readable media that stores program code and/or data for longer periods of time, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example.
- the computer readable media may also be any other volatile or non-volatile storage systems.
- a computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device.
- a block that represents one or more information transmissions may correspond to information transmissions between software and/or hardware modules in the same physical device. However, other information transmissions may be between software modules and/or hardware modules in different physical devices.
Landscapes
- Engineering & Computer Science (AREA)
- Robotics (AREA)
- Mechanical Engineering (AREA)
- Manipulator (AREA)
- Automation & Control Theory (AREA)
- Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- Fuzzy Systems (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
Abstract
Description
- This application is a continuation of U.S. patent application Ser. No. 15/598,655, filed Oct. 10, 2019, which is a continuation of U.S. patent application Ser. No. 15/616,096, filed Jun. 7, 2017, which is a continuation of U.S. patent application Ser. No. 14/242,828, filed Apr. 1, 2014, the contents of each are incorporated by reference herein.
- Automated manufacturing processes may involve the use of one or more robotic devices that may be used to construct an output product, such as a car, a wall, a piece of furniture, or any number of other physical fabrications. The robotic devices may be equipped with end-effector-mounted tools, such as a gripper or a drill, that may be used during a construction process. The robotic devices may be programmed with sequences of specific motion commands and commands for other operations in order to cause the robotic devices to complete a manufacturing process.
- The present disclosure provides methods and apparatuses that allow for use of a graphical interface in order to cause one or more robotic devices to construct an output product. In particular, using a graphical interface, input data specifying at least one motion path and at least one sequence of tool actions to occur at points within the motion path may be received. Then, based on node configuration data, a number of different digital nodes may be generated within a display window. The digital nodes may include at least one robot node describing a robot actor. The digital nodes may also include at least one motion command node that determines robot motion commands and at least one tool command node that determines tool commands based on received input data. The at least one robotic device may then execute motion commands determined by the at least one motion command node and execute tool commands determined by the at least one tool command node in order to construct the output product.
- In one example, a method is provided that includes causing a graphical interface to be displayed on a display device. The method may further include receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data. The method may additionally include generating, based on the node configuration data, a plurality of digital nodes. The plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions. The method may further include providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- In a further example, a system including at least one robot actor and a control system is disclosed. The control system may be configured to cause a graphical interface to be displayed on a display device. The control system may be further configured to receive input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one construction path for use by the at least one robot actor, where the at least one construction path comprises at least one sequence of robot operations, where at least some of the robot operations comprise use of at least one tool, and where the input data further indicates node configuration data. The control system may additionally be configured to generate, based on the node configuration data, a plurality of digital nodes. The digital nodes may include (a) at least one robot node describing one or more attributes of a corresponding one of the at least one robot actor and (b) at least one command node that determines a sequence of robot commands to cause a robot actor corresponding to the at least one robot node to perform the at least one sequence of robot operations within the at least one construction path. The control system may also be configured to provide instructions for the at least one robot actor to execute the sequence of robot commands determined by the at least one command node.
- In another example, a non-transitory computer readable medium having stored therein instructions, that when executed by a computing system, cause the computing system to perform functions is disclosed. The functions may include causing a graphical interface to be displayed on a display device. The functions may further include receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data. The functions may additionally include generating, based on the node configuration data, a plurality of digital nodes. The plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions. The functions may further include providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- In yet another example, a system may include means for causing a graphical interface to be displayed on a display device. The system may further include means for receiving input data corresponding to one or more interactions with the graphical interface, where the input data indicates at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path for use in construction of an output product, and where in the put data further indicates node configuration data. The system may additionally include means for generating, based on the node configuration data, a plurality of digital nodes. The plurality of digital nodes may include (a) at least one robot node describing one or more attributes of at least one robot actor, (b) at least one motion command node that determines a sequence of robot motion commands to cause the at least one robot actor to move along the at least one motion path, and (c) at least one tool command node that determines a sequence of tool commands to cause the at least one robot actor to operate at least one tool according to the at least one sequence of tool actions. The system may further include means for providing instructions for the at least one robot actor to move according to the sequence of robot motion commands determined by the at least one motion command node and execute the sequence of tool commands determined by the at least one tool command node to construct the output product.
- The foregoing summary is illustrative only and is not intended to be in any way limiting. In addition to the illustrative aspects, embodiments, and features described above, further aspects, embodiments, and features will become apparent by reference to the figures and the following detailed description and the accompanying drawings.
- Figure I shows a block diagram of a manufacture control system, according to an example embodiment.
-
FIG. 2A shows a view of a robot with 7 degrees of freedom, according to an example embodiment. -
FIG. 2B shows a view of a robot with an attached gripper, according to an example embodiment. -
FIG. 2C shows a view of a robot with an attached spindle, according to an example embodiment. -
FIG. 3A shows a view of a tool rack, according to an example embodiment. -
FIG. 3B shows a view of a tool rack and two robots, according to an example embodiment. -
FIG. 4A shows a graphical interface with a 3D model, according to an example embodiment. -
FIG. 4B shows additional graphical interfaces with 3D models, according to an example embodiment. -
FIG. 5 illustrates a node-based graphical interface and a visualization of a building process, according to an example embodiment. -
FIG. 6A illustrates a toolbar for a graphical interface, according to an example embodiment. -
FIG. 6B illustrates an organization of digital tools, according to an example embodiment. -
FIG. 6C is a block diagram of an example workflow, according to an example embodiment. -
FIG. 7 is a block diagram of an example method, according to an example embodiment. -
FIG. 8A illustrates a graphical interface with a digital robot actor, according to an example embodiment. -
FIG. 8B illustrates a graphical interface with a digital robot actor and a dropdown menu, according to an example embodiment. -
FIG. 8C illustrates a graphical interface with a digital robot actor and a visualization of the digital robot actor, according to an example embodiment. -
FIG. 9 is block diagram of an example node configuration, according to an example embodiment. - Example methods and systems are described herein. Any example embodiment or feature described herein is not necessarily to be construed as preferred or advantageous over other embodiments or features. The example embodiments described herein are not meant to be limiting. It will be readily understood that certain aspects of the disclosed systems and methods can be arranged and combined in a wide variety of different configurations, all of which are contemplated herein.
- Furthermore, the particular arrangements shown in the Figures should not be viewed as limiting. It should be understood that other embodiments might include more or less of each element shown in a given Figure. Further, some of the illustrated elements may be combined or omitted. Yet further, an example embodiment may include elements that are not illustrated in the Figures.
- Example embodiments provide for a graphical interface that allows a user to configure one or more robotic devices in order to author a manufacturing process using the robotic devices. In particular, a user may be able to select and configure digital nodes via the graphical interface describing robot actors and other aspects of a physical workcell in order to control a building process in the physical world. Additionally, the graphical interface may provide a layer of abstraction so that users may input motion paths and tool operations for robotic devices without having to specify particular robot commands.
- In some examples, robotic devices may be represented within a graphical interface as digital robot nodes describing one or more attributes of robotic actors within a physical, workcell. In further examples, digital robot actors may include a number of interchangeable component nodes. For instance, a component node may include a robot node representing different makes and models of commercial robots, along with different limitations (e.g., ranges of motion) of particular robotic devices. In further examples, a component node may include a tool node representing different types of tools that may be used by a particular robotic device during construction, such as grippers or spindles. In yet further examples, a component node may include an input/output (IO) node representing types or ranges of data input or output available to communicate with a particular robotic device. In additional examples, a component node may also include a track node representing different tracks of motion that a robot can move along within the physical workcell. Other component nodes are also possible.
- In further examples, the graphical interface may enable a user to configure a stage node describing all of the available robot actors, tools, 10 types, and/or axes of motion within a physical workcell. The robot actors may include robots from a variety of different types and manufacturers. The construction tools may include a number of different tools usable by a robotic device in a building process, such as a gripper, spindle, or welding iron. In additional examples, users may configure the stage node to indicate which components such as robot actors or construction tools are currently available within a particular physical, workcell for a building process.
- In some examples, users may also be able to select one or more motion command nodes which determine sequences of robot motion commands for one of the robot actors to follow based on inputted motion paths. In particular, a motion command node may determine robot commands for a particular robot actor in the stage that move the robot's end effector through a sequence of frames (positions and orientations) that correspond to an inputted motion path. For instance, a motion command node may determine a sequence of joint angles and/or joint movements that cause a robot actor to follow a particular motion path. In further examples, a graphical interface may contain an interpolation slider that may be used to determine an interval at which robot motion commands are determined along the motion path.
- In some examples, some or all of the motion paths may be determined based on input geometry. For instance, a motion path for a robot actor to follow (e.g., a sequence of positions and orientations through space) may be determined based on mathematically derived planes from parameters describing a desired output product In some examples, one or more endpoints for one or more of the motion paths may be determined using target planes on the output product. In further examples, the graphical interface may provide a user with endpoints of motion paths and allow the user to adjust portions of the motion path between the endpoints.
- In further examples, users may be able to select one or more tool command nodes that indicate certain tool actions (e.g., open a gripper) for one of the robot actors to take in order to construct the output product. The tool command nodes may determine a sequence of tool commands that cause a particular robot actor to execute the inputted tool actions. In particular, a tool command node may allow the user to configure a sequence of particular tool actions (e.g., pick up a stick, hold a stick, put down a stick, nail a stick, etc.) for the robotic device to take at certain points along its determined motion path in order to construct the output product.
- In an alternative embodiment, command nodes may be used to determine both motion commands and tool commands for robot actors. For instance, a construction path may be determined based on data received through a graphical interface for a particular robot actor. The construction path may indicate sequences of robot operation for the robot actor to take, which may include both motions and tool operations within a single construct. A command node may then determine robot operations for a particular robot actor to execute based on a received construction path. Other configurations are also possible.
- In additional examples, the set of motion commands and the set of tool commands determined by a motion command node and a tool command node may be paired up to create a step node that includes a sequence of movements and tool actions for a robot to follow to complete some action during construction (e.g., move to a stick and pick up the stick), In further examples, export nodes may be used to export steps to a particular robot actor in order to cause the robot actor to execute the movements and tool actions within a physical workcell In some examples, these step nodes may be reusable building blocks, within the graphical interface that allow for sequencing more complex sequences of motion. In further examples, sync nodes may be used to cause commands from two or more step nodes to start at the same time, which may allow for multiple robotic devices to work in parallel.
- Example embodiments may provide for motion planning and control of multi-axis robotic systems for use in the manufacturing and making industries. Example design-to-production systems may allow users to change parameters describing an output product on the front end, with the effects propagated through to a product manufactured by one or more robotic devices using one or more tools. In some examples, users may be provided with a graphical interface that allows for the configuration of the robot actors using a diverse toolset in order to automate the building process. In further examples, robot motions may be abstracted so that users don't have to program specific robot commands (e.g., motion commands or tool commands) in order to control the building process. Accordingly, users may be able to design a building process without specific knowledge of commands for particular types of robots. Additionally, users may be provided, with one or more interfaces that allow for varying amounts of control over specific robot operations within a manufacturing process, during offline motion programming and/or during runtime.
- In further examples, users may be provided with a three-dimensional (3D) modeling graphical interface that allows the user to alter one or more variables describing a physical workcell and/or a desired output product that affect a building process in the physical world. Additionally, the user interface may provide abstract ways to represent physical objects digitally as nodes within a software environment. In particular, the user experience may enable users to select from an array of tools which can be configured and combined in a number of different ways to control different types of robot actors and hardware components within a physical workcell.
- In further examples, the physical workcell may include a physical stage or stages on which a physical building process is planned or is occurring within the physical world. In some examples, the physical workcell may include a variety of different robot actors and other hardware components as well as physical materials that may be used in the building process. In further examples, the physical workcell may contain a tool rack and/or an automated tool changer. In additional examples, the physical workcell may contain one or more different types of sensors. Also, the physical workcell may include any number of different dimensions, including platforms for particular building activities.
- It should be understood that the present disclosure is not to be limited in terms of the particular embodiments described in this application, which are intended as illustrations of various aspects. Numerous components of example manufacturing systems are described herein. Systems that contain only some of those components or any combination of such components are contemplated as well. Many modifications and variations can be made without departing from the spirit and scope of the disclosed systems and methods. Functionally equivalent methods and apparatuses within the scope of the disclosure, in addition to those enumerated herein, will be apparent to those skilled in the art.
- Example embodiments may involve use of a manufacture control system to create automated programming of robotics arms during a building process.
FIG. 1 describes an examplemanufacture control system 100.Manufacture control system 100 may be part of a manufacturing environment used to control one or more robotic devices to use one or more tools to construct some output product.Manufacture control system 100 may comprise amaster control 10, input and feedback systems 20,system devices 40, andsafety systems 90. From the most basic perspective,manufacture control system 100 may function when an input system 20 provides instructions to one ofsystem devices 40 viamaster control 10. - In one potential embodiment as part of a
manufacture control system 100, input and feedback systems 20 may include adatabase 22, amaster input 24, asoftware control 26, and an independentmanual control 28. As part of the input and feedback systems 20,database 22 may operate to provide a set of timing and position data to direct all or a portion ofdevice actors system devices 40. Twodevice actors FIG. 1 , but any number of device actors could be used withinmanufacture control system 100. Alternatively,database 22 may store data being created by manual or individual movement or data input related to operation and function ofdevice actors Database 22 may also store data created independently ofdevice actors software control 26. - A
master input 24 may be any device that functions to operate all of thedevice actors manufacture control system 100.Master input 24 may function by sending input control signals tomaster control 10.Master control 10 may then adapt the signal frommaster input 24 to send individual control signals to a plurality of robot actors operating asdevice actors device actors master control 10 when a signal is received frommaster input 24, including a signal to maintain a status quo or non-action to devices that are not operating asdevice actors device actors manufacture control system 100 may not be sent any signal frommaster control 10 as part of the operation ofmotion control system 100 for a particular part of the manufacturing process. - In some examples,
software control 26 may act as a replacement formaster input 24 in sending control signals to the plurality ofdevice actors master control 10. Alternately,software control 26 may control individual devices from amongdevice actors software control 26 may function to model the behavior of individual devices ofdevice actors software control 26 may contain a software model for an individual device, which allows control signals to be created for the device without actually sending the control signals to the device. The control signals may then be stored in thesoftware control 26, indatabase 22, within a computer memory component that is part ofmaster control 10, or within computer memory that is part of the device ofdevice actors software control 26 and propagated to the appropriate storage location, a master control signal fromsoftware control 26 or frommaster input 24 may activate the control signal for the individual device to act in conjunction withother device actors - In further examples, certain devices of
device actors manual control 28. As described above with respect tosoftware control 26, control signals for an individual device may be created in software modeling. In addition or instead, a device may have independentmanual control 28 that may be used to operate a device ofdevice actors manual control 28 may be given input commands over time that are recorded todatabase 22 or a memory device ofmaster control 10. During creation of a set of instructions using independentmanual control 28, the independentmanual control 28 may communicate directly with the associated device ofdevice actors manual control 28 may send a control signal tomaster control 10, which then conveys the signal to the associated device ofdevice actors - The control signal may then be created either from the signal of the independent manual control 28 (e.g., a separate user interface), or from a measured feedback reading created by the operation of the associated device. Additionally, although in many situations, it may be preferable to have the independent
manual control 28 actually control the associated device during control signal creation in real time, control signals may instead be created without controlling the device. For example, if input signals are expected for certain time marks, an independentmanual control 28 may be operated independent of the related device, and the control operation may be recorded. Accordingly, instructions for individual device actors ofdevice actors manufacture control system 100. - In further examples,
master control 10 may allow for real-time control of components of a building system by providing a link between a virtual world (e.g., software control 26) and the physical world (e.g., a physical workcell containingdevice actors 42, 44). Accordingly, movements of a physical robot within the physical world may be used to drive the current position of a corresponding virtual robot in real time. Similarly, movements of a virtual robot may be used to drive the current position of a physical robot in the physical world as well or instead. - In one potential embodiment, individual control signals for specific device actors may be coordinated into a single file within a memory of a master control with a common base time provided by a master clock within the master control. During operation, the master control may extract control signals for each device actor and provide individual control signals to 0065ach device actor at the appropriate intervals. In an alternative embodiment, the master control may maintain separate individual control signal files and timing data for different device actors, and synchronize the different control signals separately from the individual control files.
- In another alternative embodiment, the control data for a portion of the device actors may be transferred by a master control to a memory within an associated individual device actor. During operation, device actors having control data within memory may receive a synchronization signal that indicates a location in a global timeline, a rate of progress through a global timeline, or both.
- Network support may also enable communications from
master control 10 to one or more ofsystem devices 40. In one potential embodiment, a network may comprise an EtherCAT network operating according to IEEE 1588. In such an embodiment, packets may be processed on the fly using a field bus memory management unit in each slave node. Each network node may read the data addressed to it, while the telegram is forwarded to the next device. Similarly, input data may be inserted while the telegram passes through. The telegrams may only be delayed by a few nanoseconds. On the master side, commercially available standard network interface cards or an on-board Ethernet controller can be used as a hardware interface. Using these interfaces, data transfer to the master control via direct memory access may be achieved with no CPU capacity taken up for the network access. The EtherCAT protocol uses an officially assigned Ether Type inside the Ethernet Frame. The use of this Ether Type may allow transport of control data directly within the Ethernet frame without redefining the standard Ethernet frame. The frame may consist of several sub-telegrams, each serving a particular memory area of the logical process images that can be up to 4 gigabytes in size. Addressing of the Ethernet terminals can be in any order because the data sequence may be independent of the physical order. Broadcast, multicast, and communication between slaves are possible. - Transfer directly in the Ethernet frame may be used in cases where EtherCAT components are operated in the same subnet as the master controller and where the control software has direct access to the Ethernet controller. Wiring flexibility in EtherCAT may be further maximized through the choice of different cables. Flexible and inexpensive standard Ethernet patch cables transfer the signals optionally in Ethernet mode (100BASE-TX) or in E-Bus (LVDS) signal representation. Plastic optical fiber (POF) can be used in special applications for longer distances. The complete bandwidth of the Ethernet network, such as different fiber optics and copper cables, can be used in combination with switches or media converters. Fast Ethernet (100BASE-FX) or E-Bus can be selected based on distance requirements.
- Further, such an embodiment using EtherCAT supports an approach for synchronization with accurate alignment of distributed clocks, as described in the IEEE 1588 standard. In contrast to fully synchronous communication, where synchronization quality suffers immediately in the event of a communication fault, distributed aligned clocks have a high degree of tolerance from possible fault-related delays within the communication system. Thus, data exchange may be completely done in hardware based on “mother” and “daughter” clocks. Each clock can simply and accurately determine the other clocks' run-time offset because the communication utilizes a logical and full-duplex Ethernet physical ring structure. The distributed clocks may be adjusted based on this value, which means that a very precise network-wide time base with a jitter of significantly less than 1 microsecond may be available.
- However, high-resolution distributed clocks are not only used for synchronization, but can also provide accurate information about the local timing of the data acquisition. For example, controls frequently calculate velocities from sequentially measured positions. Particularly with very short sampling times, even a small temporal jitter in the displacement measurement may lead to large step changes in velocity. In an embodiment comprising EtherCAT, the EtherCAT expanded data types (timestamp data type, oversampling data type) may be introduced. The local time may be linked to the measured value with a resolution of up to 10 ns, which is made possible by the large bandwidth offered by Ethernet. The accuracy of a velocity calculation may then no longer depend on the jitter of the communication system.
- Further, in an embodiment where a network comprises EtherCAT, a hot connect function may enable parts of the network to be linked and decoupled or reconfigured “on the fly”. Many applications require a change in I/O configuration during operation. The protocol structure of the EtherCAT system may take account these changing configurations.
- In further examples,
safety systems 90 may be provided for preventative safety in detecting potential collisions between device actors in modeling the motion of the actors through a global timeline. Further, such modeling through a global timeline may be used to set safety parameters forsafety systems 90. Modeling of locations and velocities of device actors through a global timeline may enable identification of unsafe zones and unsafe times in an area of a physical workcell. Such an identification may be used to set sensing triggers of object detectors that are part of an example safety system. For example, if an area within 5 feet of a certain device actor is determined to be at risk of collision, and a buffer zone of IO additional feet is required to insure safety during operation, a LIDAR detector may be configured to detect unexpected objects and movement within a 15 foot area of the device actor during operation, and to automatically create a safety shutdown if an object is detected. In an alternative embodiment, the LIDAR detector may be configured to create a warning signal if an object is detected in a periphery of the danger zone, and only to create a shutdown if the detected object is moving toward a potential impact zone. - In an alternate embodiment,
safety systems 90 may include modeling of actors and models of defined safe zones. Analysis of the motion of the actors in software control may allow a modeled safety check to see if any actor collides with a defined safe zone. In some examples, safe zones may be defined by entry of fixed volumes of space into a software control, by image capture of a physical workcell. Safe zones may also be defined to be variable based on a detected motion, jerk, velocity, or acceleration of an object in a safe zone. In an alternate embodiment, a safe zone may be defined by input from transponder device data. For example, a transponder location device may be attached to a robotic device actor, and a safe zone defined by a distance from the transponder. The transponder may feed location data to software control, which may update safe zones within a software control or within a master safety control. In another embodiment, fixed safe zones may be defined within software control, and published prior to a safety PLC within a master safety control prior to operation of a building process. - In some examples,
system devices 40 may additionally include one ormore sensors Master control 10 may stream data in from one or more different types of sensors located within the physical workcell. For instance, data from the sensors may reflect dimensions or other properties of parts and/or materials, within a physical workcell, as well as how the parts and/or materials are currently positioned within the real world. This data may then be streamed out to one or morerobotic device actors - In further examples,
robotic device actor robotic device actors automated tool changer 50. For instance,master control 10 may contain programming logic in order to automate the selection and equipping of tools fromtool rack 50. In other examples, instructions to cause one of therobotic device actors tool changer 50 may come from independentmanual control 28 as well or instead. - Referring now to
FIGS. 2A-2C and 3A -C, several non-limiting examples ofsystem devices 40, includingrobotic device actors tool changer 50 will be described. Although these figures focus on the use of robotic arms, other types ofdevice actors system devices 40 may be used in some examples as well or instead. -
FIG. 2A illustrates a robotic device actor, according to an example embodiment. In particular,robotic device actor 200 may include arobotic arm 202 with anend effector 204 capable of being equipped with one or more different tools. Therobotic arm 202 may be capable of motion along six degrees of freedom, depicted inFIG. 2A as A1-A6. In certain examples,robotic device actor 200 may be further capable of motion along one or more axes AO, such as along a rail which is not shown that allows side to side movement. In certain embodiments, instructions may be given to positionend effector 204 at a specific location, and the positions of therobotic arm 204 along A1-A6 and/or ofrobotic device actor 200 along one or more axes AO may be calculated by a process of the related manufacture control system. In alternative embodiments, position control ofrobotic device actor 200 and/orrobotic arm 202 may require separate, individual settings and control commands. Robotic devices operating with fewer degrees of freedom may be used in some examples as well or instead. -
FIG. 2B illustratesrobotic device actor 200 equipped with agripper 206. In particular, thegripper 206 may be placed atend effector 204 of therobotic arm 202. Thegripper 206 may be used for various functions during a building process, such as picking up objects or parts, moving objects or parts, holding objects or parts, and/or placing objects or parts. A variety of different types of grippers may be used, such as a vacuum gripper, a tumble gripper, or a passive centering gripper. Additionally, grippers with different dimensions or other properties may be used, possibly to coincide with different types of robot actors within a physical workcell. -
FIG. 2C illustratesrobotic device actor 200 equipped with a spindle 208. A spindle 208 may include a rotating axis for use in various functions within a building process, such as cutting materials, shaping materials, milling or routing. The spindle 208 could be a variety of different types, such as a grinding spindle, an electric spindle, a low-speed spindle, or a high-speed spindle. Additionally, spindles with different dimensions or other properties may be used, depending on the different types of robot actors within a physical workcell. In some examples, other types of tools may be used by robotic device actors as well or instead. -
FIG. 3A illustrates a tool rack, according to an example embodiment. The tool rack may contain a number of different fabrication tools (e.g., spindles or grippers) and may be used along with an automated tool changer in order to equip robotic devices with particular tools to use within a physical workcell. In some examples, the tool rack may contain severaltool rack modules track 300, with each of thetool rack modules tool rack modules track 300. In further examples, a tool rack module may be capable of interfacing with a slave module that allows for a particular tool to be selected from the tool rack module and then equipped onto a robotic device. For instance, referring toFIG. 3A ,tool rack module 302 may interface withslave module 310 andtool rack module 306 may interface withslave module 312. - In order to facilitate tool changing, the tool rack modules may be equipped with built-in safety sensors to minimize the risk of accidental tool fetch and drop commands. Additionally, the tool change slave modules may include IO breakout boxes to simplify passing IO trigger signals to control tools. In some examples, the IO breakout boxes may interface with a timing control system, such as
master control 10 described with respect toFIG. 1 , that controls the robotic devices within a physical workcell.Master control 10 may be used to direct a tool change for a particular robotic device, which may be configured in advance usingsoftware control 26 and/or from independentmanual control 28 during runtime. -
FIG. 3B illustrates use of the tool rack to equip robotic devices with tools, according to an example embodiment. In particular, a firstrobotic device 314 may move itsend effector 316 to a position over aslave module 310 that interfaces with atool rack module 302 of a tool rack. For instance, therobotic device 314 may currently be equipped withgripper 318, and may be controlled to move to to the tool rack in order to placegripper 318 in the tool rack and equip a different tool held bytool rack module 302. Additionally, a secondrobotic device 320 may have positioned itsend effector 322 onslave module 312 in order to equipspindle 324, which may have been held byslave module 312. After equippingspindle 324,robotic device 320 may then proceed to move away from the tool rack and complete operations using thespindle 324. The tool rack modules may be positioned on the tool rack so that multiple robotic devices may equip or change tools at the same time. In some examples,additional rack modules - In further examples, instructions from a control system, such as
master control 10 described with respect toFIG. 1 , may be used in order to instruct a robotic device how to equip a tool during runtime (e.g., to determine where a tool is within the tool rack and solve an end effector problem in real time in order to position the end effector over a slave module to enable the robotic device to pick up the tool). In additional examples, a drive system (e.g., a VFD used to supply power drive a spindle) may be mounted at a separate fixed location within a physical workcell in order to supply power on the tool changer system. -
FIG. 4A shows a graphical interface containing a 3D model, according to an example embodiment. As shown, agraphical interface 400 may contain an input,window 402 which may allow a user to enter parameters relating to anoutput product 406, such as a wall built using individual sticks. Theinput window 402 may allow the user to enterparameters 404 that may relate to aspects of the output product, including dimensions, density, curvature properties, other geometric properties, materials to be used, and/or other numeric inputs. The inputs may be used to derive a parametric solution for anoutput product 406. Additionally, the inputs may be used to generate a sketch of theoutput product 406 within adisplay window 408 of thegraphical interface 400. -
FIG. 4B shows three different output products based on different user input parameters, according to an example embodiment. A first view of thegraphical interface 440 may contain aninput window 402 and adisplay window 408. Theinput window 402 may allow a user to enterparameters 404 relating to a desired output product, including product dimensions, density, curve offsets, amount or type of curvatures, and/or other geometric or numerical inputs. Based on theinput parameters 404, a geometric representation of theoutput product 406 may be displayed within thedisplay window 408. In some examples, a user may modifyindividual parameters 404 in order to change aspects of theoutput product 406. - For instance, a second view of the
graphical interface 450 shows adifferent output product 406 within thedisplay window 408 based ondifferent input parameters 404 within theinput window 402. In this example, dimensions of theoutput product 406 and/or materials used to produce theoutput product 406 may be modified to produce anoutput product 406 with a greater height as shown in thesecond view 450. Further, athird view 460 shows anotherdifferent output product 406 within thedisplay window 408 based ondifferent input parameters 404 within theinput window 402. For example, parameters relating to the curvature of the output product may be modified by a user in order to produce anotherdifferent output product 406 as shown in thethird view 460. -
FIG. 5 shows a graphical interface for architecting a robotic building process, according to an example embodiment For example, the graphical interface may be part ofsoftware control 26 as described above with respect toFIG. 1 . As shown, agraphical interface 500 may contain aninput window 502 which allows a user to control aspects of the building process, including nodes related to robot actors, tools, motion paths, and tool operations for use during construction. Thegraphical interface 500 may additionally contain adisplay window 510 which contains a 3D geometric view of the physical workcell, including components such as robot actors, tools, materials, and/or constructed output products. In example embodiments, theinput window 502 may provide a visual programming interface or different type of interface that may allow a user to enter parameters describing a desired output product and/or information about the actors and tools to be used in the building process. Input data collected using theinput window 502 may be used to control geometry and/or other aspects of the physical workcell displayed within thedisplay window 510. - In one example, a user may input parameters to control a building process using an
input window 502 containing a visual programming interface, such as an interface built using a visual programming language, such as a commercial software program known as Grasshopper. The interface may allow a user to include one ormore nodes 504 which may represent components of the building process, such as robot nodes representing different types and/or configurations of robots, tool nodes representing different types and/or configurations of tools, IO nodes representing types of available IO, track nodes representing possible tracks of motion of robot actors, and command nodes for determining motion commands and other types of commands for robot actors. - As shown within
window 502 ofFIG. 5 ,individual nodes 504 may be connected together usingconnectors 506. Aconnector 506 between two nodes may indicate that the output of a first node is to be used as an input to a second node. For instance, a single robot node may receive as inputs information from several different component nodes, such as nodes representing the type of robot, the type of tool used by the robot, a track the robot can travel along, and so on. - In further examples, the
window 502 ofFIG. 5 may contain atimeline 508. Thetimeline 508 may have a cursor representing a current timestamp (e.g., 83 as shown in the figure) which may represent a particular point in time of the manufacturing process. In addition, thetimeline 508 may contain buttons to play through the building process at a particular speed, or fast-forward or rewind through the building process. Thetimeline 508 may be used to control the point in time at which the geometry and/or other aspects of the physical workcell are displayed within thedisplay window 510. Further, thetimeline 508 may be used to indicate a particular point in time either for purposes of simulating the building process or for visualizing within software an actual physical building process taking place within the physical world. - As shown in
FIG. 5 , the user interface may additionally contain adisplay window 510 which may display geometry and/or other aspects of the physical workcell based on inputs from theinput window 502. For example, thedisplay window 510 may include geometry relating to robot actors, tools, building materials, robotic motion paths, and output products, among other things. In one example, thedisplay window 510 may be designed using a commercial 3D modeling software, such as Rhinoceros, as shown withinFIG. 5 . Thedisplay window 510 may display geometry within a particularphysical workcell 512. Thedisplay window 510 may include options to change the perspective of thephysical workcell 512 and/or to zoom in or zoom out a view of thephysical workcell 512. - The
physical workcell 512 may include one ormore robot actors 514. Therobot actors 514 may bedevice actors 42 and/or 44 as described above with respect toFIG. 1 and/orrobotic device 200 as described with respect toFIGS. 2A-2C . Support may be provided for numerous different types of multi-axis robotic systems of different types and/or from different manufacturers. In some examples, one or more of therobot actors 514 may be traditional six-axis robots. In additional examples, other types of robots that may be configured to operate along fewer or more axes may be included for use within thephysical workcell 512 in addition or instead. - In further examples, robot actors may be represented within a software interface as robot nodes, which may be put together from a number of interchangeable component nodes, including robot nodes representing different makes and models of commercial robots, tool nodes representing different types of physical tools that may be used for construction such as grippers or spindles, IO nodes representing
different types 10 available to communicate with a robot actor and track nodes representing different types of axes that a robot can move along. In some examples, individual tools and/or tooling parameters (such as wrist mount offsets or tool center points) can be abstracted into components that can be assembled by a user into compound tools as well. - The
display window 510 may additionally contain one ormore motion paths 516 representing paths of motion ofindividual robot actors 514. Themotion paths 516 may indicate paths to be taken by therobot actors 514 during the building process, such as to pick up materials and attach them to an object under construction. In some examples, themotion paths 516 may further indicate points at which particular input or output actions will occur. For instance, an “x” on amotion path 516 may indicate a point at which arobot actor 514 uses a tool such as a gripper to pick up a particular type of material. In further examples, themotion paths 516 may be synchronized with thetimeline 508 from theinput window 502. Accordingly, in some examples, therobot actors 514 may be made to move along themotion paths 516 to positions at particular points in time based on the timestamp indicated by thetimeline 508. - The
physical workcell 512 may additionally contain one ormore materials 518 to be used during the building process. In this simplified example, thematerials 518 consist of sticks used to construct awall 520.Motion paths 516 may be determined for therobot actor 514 to take in order to move the individual sticks 518 onto thewall 520. In other examples, a variety of different types of materials, including connective materials such as glue, may be used simultaneously by the robot actors to construct more complex output products. - In further examples, the
physical workcell 512 may also contain other components not shown inFIG. 5 that may be used in the building process. For instance, one or more sensors may be included to sense information about the robot actors and/or materials in the physical workcell in order to influence motion paths taken by the robot actors. For example, a torque sensor may be used to determine if a particular piece of material is likely to break under stress. A control system, such asmaster control 10 described above with respect toFIG. 1 , may be used to interface with the robot actors and/or sensors within the physical workcell. - In some examples, the
display window 510 may provide users with multiple 3D views of the physical workcell, and may allow a user to change the orientation and/or zoom of a particular view. In other examples, thedisplay window 510 may present other types of representations of the physical workcell, such as numerical representations, as well or instead. - In some examples, an input window may additionally contain a toolbar containing digital tools to control aspects of the building process.
FIG. 6A shows a toolbar for a graphical interface, according to an example embodiment. Thetoolbar 602 may be equipped \-Vith a variety ofdifferent toolsets 604 that may be used to design or control a building process within an input window of a graphical interface.Toolsets 604 may be provided with digital tools relating to generating robot motion paths, transforming between different planes or axes, describing robot actors, describing physical building tools, sequencing individual robot motions, communicating data input and/or output to and/or from robot actors, mapping between a virtual software environment and a physical workcell, and/or enabling visualization of a building process, for example. -
FIG. 6B shows an organization of digital tools within a tool bar, according to an example embodiment. As shown, the digital tools may be divided into a number of different categories. The digital tools may then be used in combination to design a building process, as shown byFIG. 6C .FIG. 6C is a block diagram of an example workflow, according to an example embodiment. In particular,FIG. 6C shows workflow involving a number of digital tools, which may be accessible, within a toolbar as depicted inFIG. 6i andFIG. 6B or by another means within a graphical interface. As shown, the digital tools may be divided into a number of different categories. One or more digital tools from a number of different categories may be selected by a user to affect particular aspects of the building process, including the robot actors and other components within a physical workcell that may be used in the process. - In one example, a toolbar may include
path tools 608 relating to generating target planes that may be used to determine motion paths of robot actors. In some examples, thepath tools 608 may take asinput geometry 606 describing a desired output product, such as geometry generated by parametlic modeling software, Grasshopper. For instance, theoutput product geometry 606 may be generated based on user input within an input window specifying aspects of the output geometry, including dimensions, density, curvature, materials, and so on. Thepath tools 608 may then determine target planes for robot motion paths based on theoutput product geometry 606. - In some examples, the
output product geometry 606 may include particular splines, surfaces, and/or other geometrical constructions to be included within an output product. Thepath tools 608 may then provide shortcuts for generating target planes relating to aspects of the output product in a format that can be turned into particular robot trajectories that may be used to construct an output product containing the target planes. Motion paths for individual robot actors may then be derived as a function of the target planes in addition to tool definitions and material properties, for example. - In further examples, a toolbar may include transform
tools 610 relating to transformations between different axis frames or offsets, as shown byFIG. 6B andFIG. 6C . For instance, thetransform tools 610 may provide transformations between coordinate frames at the base or joints of a particular robot and a stage containing the output product. In other examples, thetransform tools 610 may additionally allow for transformations between multiple robots operating within different frames of reference as well. As shown mFIG. 6C , transformations may be applied before and/or after determining sequences of motion for individual robot actors. - In further examples, a toolbar may include
stage tools 612 that represent aspects of a physical workcell, such as robot actors, tools, IO, and/or axes. In some examples,stage tools 612 may also provide a mapping between virtual robots in software and physical robot actors within the physical workcell, as shown byFIG. 6B andFIG. 6C . Thestage tools 612 may be used byengine node 624 to send trajectories for robot actors to take based on output fromcommand tools 622 to acontrol system 628. In some examples,stage node 612 may be configured in order to specify the currently available robotic devices and/or tools within a particular physical workcell. Thecontrol system 626 may then command robot actors and/or other components within thephysical world 630 based on information fromstage tools 612. - In some examples,
stage tools 612 may take input from one ormore robot nodes 614 representing attributes of individual robot actors within a physical workcell, as shown byFIG. 6B andFIG. 6C . Arobot node 614 may be used to define attributes of a robot actor, such as traditional six-axis robots or other types of robots. The robot attributes may include link lengths of the robot and/or arm lengths of the robot, offsets of the robot and/or joints of the robot, and/or limits on robot joint movement or maximum torque that a robot joint can handle, for example. - In additional examples,
stage tools 612 may also take input from one ormore tool nodes 616 for defining the attributes of physical building tools and/or a tool rack for holding the tools, as shown byFIG. 6B andFIG. 6C . For example, attributes of building tools such as grippers or spindles may be specified by tool nodes, which may be used to configure an automatic tool changer so that robot actors can easily switch between tools. In some examples, robot actors may switch between tools using a tool rack which stores the tools and facilitates a tool changing process, as described above with respect toFIGS. 3A and 3B . - In further examples, the
tool nodes 616 may include support for compound tooling that may allow component tools to be assembled into compound tools. In particular, individual tooling parameters (e.g., wrist mount offsets or tool center points) may be abstracted into components that may be assembled into compound tools. For instance, multiple tools may be aggregated into one compound tool that has multiple tool orientations and/or center points that may be used depending on which component of the tool may be required at a particular time. For example, a compound tool with an automatic tool changer may be represented by a master interface plus a number of different attachments, such as a spindle, a vacuum array, or a gripper. In another example, a compound tool may include a series of different tools, such as a gripper plus a tensioner plus a roller. Other examples of combining multiple tools and/or combining tools by abstracting tooling into parameters that define tool orientation and/or center points are also possible. - In further examples,
stage tools 612 may also take input from one or more 10nodes 618. TheIO nodes 618 may describe information relating to digital and/or analog input and/or output actions that may be taken by a robot actor, such as the type of action (e.g., gripping a material) and attributes associated with the action (e.g., width of material that can be gripped). In additional examples, the robot attributes may include one ormore axis nodes 620. Theaxis nodes 620 may describe one or more linear and/or rotational axes along which a robot can travel, including limitations on the robot's movements along the axes. - In additional examples, a toolbar may include
command tools 622, as shown byFIGS. 6B and 6C . Thecommand tools 622 may be used to determine robot commands to cause one or more of the robot actors to execute particular operations, which may include point-to-point motions, motions along external axes, and/or input or output events. For example, one ofcommand tools 622 may be used to direct a particular robot motion along one of six degrees of freedom, a particular robot motion along an external axis, or a particular input or output event, such as applying glue to a material in a particular manner. Additionally,command tools 622 may be included for creating step nodes that instruct robot actors to take a particular sequence motion steps and carry out one or more tool actions. In further examples, coordinated sequences of motions may be generated for multiple robot actors working together within a single timeframe. - In some examples, step nodes, or sequences of motions and actions, can be abstracted into reusable subroutines. For instance, a subroutine can be defined by connecting visual building blocks, which may represent particular motion commands or motion parameters. The subroutine could then be used to make one or more robots carry out the same action sequence multiple times within a single building process. In some examples, steps can be synchronized across multiple robots so that multiple robots can, work in a shared environment simultaneously. Example systems may also include an
engine node 624, which may assign each of the steps to particular robotic devices within a stage. - In further examples, users may be provided with functionality to switch between steps within the graphical interface. For instance,
timeline 508 as illustrated and described with respect toFIG. 5 may also includes buttons to skip between steps on the timeline. In some examples, digital bookmarks may be inserted by a user for particular steps. For instance, through the graphical interface, it may be possible to jump from the beginning of a “fetch stick” step to the beginning of a “nail stick” step. These bookmarks steps within the timeline may match the steps authored by the user by inputting motion commands, 10 commands, and/or other commands in a step node. - Additionally, the
engine node 624 may communicate withcontrol system 626. Thecontrol system 626 may be a computing device capable of communicating wirelessly with robot actors and/or other components such as sensors within the physical, workcell in thephysical world 630. In particular, thecontrol system 626 may provide access to real time data streams from all robot actors and devices, which may allow for precise control over the physical workcell at particular points in time. The control system could communicate with some or all of the actors or devices through wired connections or other types of communication channels as well or instead, including previously described network protocols. - In some examples, the control system may additionally contain a physical control interface such as a touchscreen interface that may allow a user to interact with the control system to view live data or modify robot actions in real time. For instance, a stage file containing information about the physical workcell including actors, tools, materials, and environmental setup on the
control system 626 may be accessible via a programming interface. A user who is watching a building process within thephysical world 630 may then make modifications to the process before it is completed. - In additional examples, a toolbar may include data input/
output tools 628 that may allow thecontrol system 626 to send and/or receive data to and/or from the virtual software environment that determines robot motion paths, as shown byFIG. 6B andFIG. 6C . Accordingly, telemetry from thecontrol system 626 may be used to create a live link between the virtual world in software and thephysical world 630. For instance, the data input/output tools 628 may be used to process information from thecontrol system 626 relating to the robot actors within the physical workcell and/or other components in the workcell such as sensors. Based on this information about thephysical world 630, the virtual robots within software may be updated with real-time feedback from the physical world 630 (e.g., motion paths for robot actors may be determined or modified based on real-time sensor data). - Additionally, the data input/
output tools 628 may be used to send data back to thecontrol system 626 so that thecontrol system 626 can effectuate particular input or output actions within thephysical world 630, for example. For instance, thecontrol system 626 may instruct a robot actor how use a tool in the physical world 630 (e.g., how to control a spindle) based on information from one or more digital tools within the software interface. - In further examples,
engine node 624 include visualizer or simulation tools that may allow a user to simulate a building process through a user interface in software, as shown byFIG. 6B andFIG. 6C . In some examples, the visualizer tools may display the building process as geometry drawn on a screen that shows the physical workcell. In other examples, the visualizer tools may display the building process as curves representing particular data values as well or instead. Additionally, in further examples, the visualizer tools may also be used to visualize a building process in software as it is actually occurring within thephysical world 630. In some examples, the visualizer tools may additionally provide a graphical representation of potential conflicts within a particular building process, such as when a robot's motion path extends outside its possible range of motion or when two robot actors may be going to collide based on the currently defined trajectories and/or step sequences. - In further examples, the visualizer component may allow a user to see simulations of the building process in advance and/or as the building takes place. In some examples, the user may use the visualizer component offline to see robotic motion paths as well as input/output events over a series of sequential steps as geometry drawn within a viewing window. In other examples, the user may be able to visualize a simulated playback as numerical data streams relating to the robot actors, materials, and/or other aspects of the physical workcell represented through curves in addition to or instead of visual geometry. In further examples, the user may also be able to see particular data points at individual timesteps, such as robotic joint values, axis values, or input/output values.
- In some example system a user may also be able to use the visualizer component to visualize a building process that is occurring in the physical world in real time. The system may interface with a control system that receives real-time data streams from sensors that may be used to scan the physical workcell, individual robot actors, and/or parts used in construction as an output product is being built. Accordingly, the visualizer's user interfaces may be updated in real time to reflect real world dimensions, properties, and/or positions of objects and actors within the environment
- A
method 700 is provided for causing one or more robot actors to construct an output product, according to an example embodiment. In some examples,method 700 may be carried out by a control system, such asmanufacture control system 100,master control 10, and/orsoftware control 26 as described in reference toFIG. 1 . The control system may communicate with the robot actors using any of the network protocols or communication methods previously described. In additional examples, part or all ofmethod 700 may be executed by any of the graphical interfaces or systems described and illustrated with respect toFIGS. 4A-4B, 5, and 6A-6C . In further examples, part or all ofmethod 700 may be carried out by one or more robotic devices, such asdevice actors system devices 40 as described in reference toFIG. 1 , ordevice actor 200 as illustrated and described in reference toFIGS. 2A-2C . Additionally, while examples with certain numbers and types of system devices may be described, various alternative embodiments may include any number and type of robotic devices as well. - Furthermore, it is noted that the functionality described in connection with the flowcharts described herein can be implemented as special-function and/or configured general-function hardware modules, portions of program code executed by a processor for achieving specific logical functions, determinations, and/or steps described in connection with the flowchart shown in
FIG. 7 . Where used, program code can be stored on any type of computer-readable medium, for example, such as a storage device including a disk or hard drive. - In addition, each block of the flowchart shown in
FIG. 7 may represent circuitry that is wired to perform the specific logical functions in the process. Unless specifically indicated, functions in the flowchart shown inFIG. 7 may be executed out of order from that shown or discussed, including substantially concurrent execution of separately described functions, or even in reverse order in some examples, depending on the functionality involved, so long as the overall functionality of the described method is maintained. - As shown by
block 702 ofFIG. 7 ,method 700 may initially involve causing a graphical interface to be displayed on a display device. In particular, a graphical interface may be provided that allows for configuration of one or more robot actors for use during a building process. In some examples, the graphical interface may include a display window containing a 3D model of a physical workcell, as described and illustrated with respect toFIG. 4A-4B . In further examples, the graphical interface may additionally include an input window containing a configurable node-based interface as described with respect toFIG. 5 . Other types or variations of graphical interfaces could be used as well. For instance, in some examples, the display window and the input window could be combined into a single graphical interface and/or divided into multiple additional sub-windows. In further examples, the graphical interface could be displayed on a PC workstation, a touch-surface device, or on a number of other different types of computing devices. - As shown by
block 704 ofFIG. 7 ,method 700 may additionally involve receiving input data corresponding to one or more interactions with the graphical interface. Users may interact with the graphical interface in a number of different ways. In particular, a user may be provided with one or more menus or toolbars that allow for configuration of aspects of a physical workcell and/or automated building process. In some examples, users may be able to select and position digital nodes, which may be represented as shapes, icons, text, modules, or as other graphical representations on a display. In further examples, the user may additionally be able to connect two or more nodes together (e.g., by using a pointing device or touch-surface interface) in order to indicate a logical relationship between the nodes (e.g., the output of one node may act as input for a second node). In other examples, data may be received through the graphical interface in other ways. For instance, a user may be able to type or otherwise enter numerical input or other types of input relating to certain aspects of a building process as well. In additional examples, user may be able to interact with geometry in a 3D modeling display (e.g., by drawing a line between two positions in a modeling environment). - In further examples, the received input data may indicate at least one motion path and at least one sequence of tool actions to execute at one or more points within the at least one motion path. In particular, a motion path may indicate a path of motion for an end effector of a robot actor to follow within the physical workcell. At one or more points of the motion path, particular tool operations may be specified, such as to pick up an object, place an object, use a drill, or apply glue to a surface. An example motion path and sequence of tool actions is illustrated and described above with respect to
FIG. 5 . - In some examples, a motion path may be defined by a sequence of frames, or planes with six degrees of freedom in Cartesian space. For instance, the sequence of frames may indicate a path for an end effector of a robot actor to follow in order to get from one point within the physical workcell to another point (e.g., to pick up a stick and then move to place the stick on a wall). In some examples, a user may enter a motion path through the graphical interface by drawing a line from one point in a 3D display window to a second point, representing two points within the physical workcell. Motion paths may also be entered in other ways in some examples as well, such as by entering numerical data representing particular Cartesian coordinates within the physical workcell.
- In further examples, a user may be provided with path tools, as described with respect to
FIGS. 6A-6C , in order to assist the user in inputting motion paths through the graphical interface. In particular, a user may first determine a model of a desired output product, as described with respect toFIGS. 4A-4B . Based on parameters entered to describe the output product, mathematical models may be generated in order to represent the geometry of the output product as splines or surfaces. Path tools may be provided that determine target frames on the output product based on the mathematical models. In some examples, the target frames may then be used in order to assist the user in determining motion paths for robot actors to follow. For instance, a target frame may indicate a particular frame (e.g., position and orientation) in which a robot's end effector can be positioned in order to place a stick at a particular point on a stick wall. In some examples, the user may be able to manipulate the motion path in between particular target frames. - In additional examples, the received input data may further indicate node configuration data. As previously described, a user may be able to select and configure a variety of digital nodes in order to configure a manufacturing process. Node configuration data may indicate how to represent particular nodes within a display window. In further examples, node configuration data may be received based on menu selections, commands from a pointing device (e.g., nodes may be drag and dropped), and/or through other means of user interaction.
- As shown by
block 706 ofFIG. 7 ,method 700 may further involve generating, based on the node configuration data, a plurality of digital nodes. In particular, based on input received through the graphical interface, a configuration of digital nodes may be determined and displayed within a display window.FIGS. 8A-8C collectively show an example configuration of digital nodes, according to an example embodiment. The simplified examples shown inFIGS. 8A-8C and elsewhere are meant for demonstration purposes and should not be viewed as limiting. In particular, other types and configurations of nodes besides those specifically illustrated here may be used as well or instead in some examples. - Based on the node configuration data, at least one robot node may be generated. A robot node may include any of the attributes previously described with respect to
FIG. 6A-6C . An example robot node is shown inFIG. 8A . In particular, arobot node 806 may be generated and displayed within aninput window 802. Therobot node 806 may be used to define attributes of a robot actor, such as traditional six-axis robots or other types of robots. For instance,robot node 806 may describe kinematic system characteristics of a particular make and model of robot, including robot attributes such as link lengths of the robot and/or arm lengths of the robot, offsets of the robot and/or joints of the robot, and/or limits on robot joint movement or maximum torque that a robot joint can handle, for example. In further examples,robot node 806 may also include information relating to tool actions or other types of actions that may be taken by a robot actor, such as the type of action (e.g., gripping a material) and attributes associated with the action (e.g., width of material that can be gripped). - In some examples, one or more robot nodes may connected to a
stage node 804, which may describe available robot actors as well as other aspects of a physical workcell, including tools, IO, and/or axes. Using astage node 804, tools may be shared between robot actors and/or swapped on a particular robot actor at different points of a sequence of robot operations. In some examples, a digital robot actor may further include input from one or more component nodes viastage node 804. In one example, a component node may be atool node 808, which may indicate one or more tools (e.g., a spindle or grip gun) which may be operable by the robotic device represented byrobot node 806. In further examples, a component node may be anIO node 810, which may indicate one or more types of IO (e.g., a VFD switch or solanoid trigger) which may be operable by the robotic device represented byrobot node 806. In yet further examples, another component node may be anaxis node 812 that may include one or more linear and/or rotational axes along which the robot can travel, including limitations on the robot's movements along the axes. Other configurations of a robot node and other digital component nodes are also possible. - In some examples, a robot node and/or one or more component nodes may be configurable using a dropdown menu, as shown by
FIG. 8B . In particular, adropdown menu 814 may contain preconfigured options for particular types of robots, axes of motion, IO types, and/or tools (e.g., that coincide with particular types of industrial robots). In some examples, one or more ofrobot node 806,tool node 808,IO node 810, and/oraxis node 812 may selected or configured usingdropdown menu 814. In other examples, custom robots with custom arrangements of robot type, tools, IO, and/or axes may be designed and then saved for later use. - In further examples, a 3D model of a robotic device may be displayed corresponding to a particular digital robot actor. For example, as shown in
FIG. 8C , a digital robot actor may be generated withininput window 802 viastage node 804 corresponding torobot node 806,tool node 808,IO node 810, andaxis node 812. Within aseparate display window 814, a 3D model of arobotic device 816 corresponding to the digital robot actor may be displayed. Parameters ofrobotic device 816 may be determined based on configurations ofrobot node 806,tool node 808,IO node 810, and/ortrack node 812. In particular, the type ofrobotic device 816 may be determined based onrobot node 806. Additionally, aspindle 818 may be shcn.vn as an end-effector-mounted tool on therobotic device 816 based ontool node 808. Accordingly, usingdisplay window 814, a user may be able to visualize how a configured robot actor will look within a physical workcell. - In additional examples, the digital nodes generated based on the node configuration data may include at least one motion command node. A motion command node may determine a sequence of robot motion commands for a particular robot actor in order to cause the robot actor to move along a received motion path. In some examples, a motion command node may take as input a robot node and a motion path. The motion command node may then determine motion commands for the robotic device corresponding to the robot node in order to cause the robotic device (or an end effector of the robotic device) to move along the motion path.
- In some examples, a user may specify a motion path in Cartesian space (e.g., XYZ coordinates in the physical workcell). For instance, a user may draw a line within a display
- window indicating a path to be followed by a robotic device. In further examples, the user may specify the position within the physical workcell for the end effector of the robotic device to move to (e.g., by specifying XYZ coordinates) as well as an angle of rotation with which to use an end-effector-mounted tool, such as a gripper (e.g., by specifying Euler angles). In some examples, a user may be assisted in determining motion paths based on generated mathematical models, as previously described.
- Using the motion path, a motion command node may determine how to command a particular robot, by specifying the joint angles and/or changes in joint angles at each of the robot's joints. For example, a typical industrial robot has six joints: three in the wrist for rotation, and three at the shoulder and elbow for positioning the wrist in Cartesian space. Because the robot has six joints, it can position an end-effector-mounted tool at an arbitrary six degree-of-freedom Cartesian pose (XYZ position and 3 Euler angles), within some limitations. The translation from Cartesian space to joint angles is called inverse kinematics (IK). In some examples, a motion command node may be referred to as an IK command node that solves the inverse kinematics problem in order to determine motion commands for a particular robot. In further examples, the motion command node may take into account limitations and ranges of motion of different types of robotic devices in determining robot motion commands.
- In further examples, a motion command node or IK command node may select from a number of different possible solutions to the inverse kinematics problem. In some examples, an IK command node may select a solution by evaluating a history of previous IK solutions to find solutions that provided desired results. In additional examples, a user may optionally specify particular aspects of a solution, such as whether the robot elbow should be positioned up or down, or the robot wrist should be positioned up or down.
- In yet further examples, an IK command node may allow the user to choose whether determined robot motions will be point-to-point or linear. Point-to-point motions may allow a robot some degree of freedom in moving between interpolation points while maximizing speed and/or smoothness. Accordingly, point-to-point motions may be curvilinear by allowing a robot to rotate its joints in between interpolation points. In contrast, linear motion may be used to force a robot to move in straight lines in between interpolation points. In further examples, a separate parameter may be included to cause a control system to interpolate between motions continuously.
- In additional examples, a user may be able to specify through the graphical interface an interval at which the motion command node determines robot motion commands for a given motion path. By selecting a smaller interval, greater precision in robot movements may be achieved, but at a greater cost of system resources. Accordingly, in some examples, a smaller interval may be chosen for more critical functions within a building process. In certain examples, the interval may be chosen using an interpolation slider, such that the interval changes for one or more motion command nodes based on the position of a graphical marker on a slider
- within the graphical interface. The interval may be selected in other ways in some examples as well.
- In further examples, at least one tool command node may be generated based on the node configuration data. In particular, a tool command node may determine a sequence of tool commands in order to cause a robot actor to operate a tool according to the received sequence of tool actions. For instance, the user may specify through the graphical interface one or more points within a motion path where certain tool operations are to be performed (e.g., pick up a stick at one point and place the stick at a second point). A tool command node may then determine tool commands for a particular robot actor in order to make the robot actor perform those tool operations at the specified locations.
- In some examples, a tool command node may take into account parameters describing the robot actor in order to determine particular commands for the robot actor to follow. For instance, a robot actor may be configured to receive certain types of input signals in order to activate end-effector-mounted tools. A tool command node may determine a sequence of digital or analog signals that will cause a particular robot actor to activate a particular physical tool in the specified manner at specified points within a motion path. In some examples, 10 values describing an aspect of tool operation (e.g., a frequency of a spindle) may be specified within the sequence of tool actions. In such examples, a tool command node may determine tool commands that cause a particular robot actor to operate a tool with the specified IO values. For example, the tool command node may determine analog signals to command the robot actor to operate a spindle with particular specified frequencies at different points within a motion path. In another example, the tool command node may determine digital signals to command a robot actor to open and close a gripper at different points within a motion path. Depending on the types of robot actors, tools, and 10 used within a physical workcell, other types of tool commands or signals to activate physical tools at particular points could be determined by tool command nodes as well.
- As shown by
block 708 ofFIG. 7 ,method 700 may also involve providing instructions for the at least one robot actor to construct the output product In particular, a robot actor may be controlled to move based on the sequence of robot motion commands determined by a motion command node. Additionally, the robot actor may be controlled to execute tool commands at particular points within the sequence of robot motion commands based on the sequence of tool commands determined by a tool command node. In some examples, a step sequencer may be provided in order to sequence together motion commands and tool commands before exporting the commands to a robot actor to execute within the physical workcell. -
FIG. 9 is block diagram of an example node configuration, according to an example embodiment. In particular,FIG. 9 shows the generation of step nodes in order to sequence together robot motion commands and tool commands before exporting the commands to a robot actor.Output geometry 902 may first be determined describing a desired output product. As previously described, a user may then determine motion paths andtool actions 904 for a robot actor to use in construction of output product based onoutput geometry 902. The robot actor, including any relevant parameters of the robotic device, tools, IO, or tracks, may be described usingrobot node 906,tool node 908,IO node 910, andaxis node 912. Each of these component nodes may be input intostage node 914, which may describe all of the available components within a particular physical workcell. - The motion paths and
tool actions 904 may be input into amotion command node 916, along withrobot node 906 andtool node 908. As described previously,motion command node 916 may determine a sequence of robot motion commands in order to cause a robotic device described byrobot node 906 andtool node 908 to move its end-effector-mounted tool through the inputtedmotion paths 904. Additionally, the motion paths andtool actions 904 andIO node 910 may be input into a tool command node, orIO command node 918.IO command node 918 may determine a sequence of IO triggers to cause the robotic device described byrobot node 906 to execute tool operations at certain points along the motion path of the tool based on the inputted motion paths andtool actions 904. In further examples,axis node 912 may be input intoaxis command node 920 in order to determine a sequence of axis commands to cause the robotic device described byrobot node 906 to move along an axis within the physical workcell described byaxis node 912. Note that in some examples, one or more robotic devices may be stationary within a physical workcell. - In additional examples, the output from
motion command node 916,IO command node 918, andaxis command node 920 may combined within astep node 922. In some examples, thestep node 922 may interleave the sequence of robot motion commands and tool commands in order to create a step that defines a step of a building process within the physical workcell. For instance, a step may indicate how the robotic device corresponding torobot node 906 may perform a particular function within a modeling process, such as moving to a pile of sticks, picking up a stick, moving to a stick wall under construction, and placing the stick on the stick wall. - In further examples, a
duplicate step node 928 may be generated within a graphical interface. For example, a user may be able to copystep node 922 in order to cause a robotic device to repeat particular operations or to cause a different robotic device to execute the operations. Accordingly, step nodes such asstep node 922 may function as reusable building blocks within a graphical interface that may allow a user to design and configure more complex sequences of motion, possibly sequencing together a number of different steps before exporting to a robotic device for execution. - In further examples, output from the
step node 922 andduplicate step node 928 may be input intoengine node 924.Engine node 924 may compile multiple steps to be performed by one or more robotic actors, and may additionally provide visualization and/or simulation functionality. In additional examples,stage node 914 may be defined by a user, describing each of the robotic devices, tools, IO, and/or axes available within a particular physical workcell. For instance, a user may configurestage node 914 to indicate the devices and/or components currently available for a building process. Then, theengine node 924 may interface with thestage node 914 in order to determine particular robot actors (including robotic devices, tools, IO, and/or axes) to perform particular steps within the building process. In other examples, thestage node 914 may define other characteristics of a physical workcell as well, such as available sensors, materials for construction, or other hardware components. - Output from
engine node 924 may be input to anexport node 926, which may be used to send instructions to a robotic device within the physical workcell. Communication between a control system and a robotic device may be achieved using any of the methods previously described. In some examples, theexport node 926 may allow a user to select a particular robotic device to cause to execute the determined step operations. In further examples, theexport node 926 may contain a dropdown menu that allows for the selection of a robotic device from one or more robotic devices within a particular physical workcell. - In yet further examples, different steps may be simultaneously executed by multiple robotic devices within a single physical workcell. In some examples, a sync node may be used to connect two or more steps by indicating that commands from each of the step nodes should start at the same time. Accordingly, each of the robotic devices may be prevented from executing commands within one of the step nodes attached to a sync node until all of the robotic devices are available and ready to begin executing commands specified by corresponding step nodes. Robot operations may be synchronized between different robot actors in different ways as well.
- In an alternative embodiment, data may be received through the graphical interface indicating at least one construction path for use by at least one robot actor. In particular, the construction path may include at least one sequence of robot operations, where at least some of the robot operations include use of a tool. For instance, a user may be able to specify both a motion path for a robot actor to follow as well as tool operations occurring along the motion path as a single construct. In one example, the user may be able to specify a construction path by drawing a line through a modeling environment with an “x” at particular points indicating tool operations, as illustrated and described with respect to
FIG. 5 . - Additionally, based on node configuration data, at least one command node may be generated in order to cause a robot actor to perform robot operations within the construction path. In particular, a single type of command node may determine both robot motion commands and tool commands in order to cause a particular robot actor corresponding to a robot node to perform operations within the construction path. For example, the command node may solve an inverse kinematics problem to determine particular robot joint angles or movements and also determine how to command the robot to perform particular tool actions at points within the received construction path. In some examples, the command node may also perform a sequencing function in order to sequence motion commands with tool commands into a single sequence of robot commands.
- In further examples, one or more of the command nodes may be conditional command nodes. A conditional command node may determine robot operations for a particular robot actor to follow based on sensor data received from one or more sensors within a physical workcell. For instance, a conditional command node may contain a conditional statement that will cause a robotic actor to only use a piece in construction if it will fit within a certain area. Further, sensor data received during construction may indicate the shape of a particular piece. Based on the sensor data, the conditional command node may determine whether or not to output robot operations to pick up and use the particular piece during the building process. Other types of conditional command nodes are also possible.
- Other variations on the disclosed systems and methods, including combining or dividing functions of digital nodes, are also possible.
- The present disclosure is not to be limited in terms of the particular embodiments described in this application, which are intended as illustrations of various aspects. Many modifications and variations can be made without departing from its spirit and scope, as will be apparent to those skilled in the art. Functionally equivalent methods and apparatuses within the scope of the disclosure, in addition to those enumerated herein, will be apparent to those skilled in the art from the foregoing descriptions. Such modifications and variations are intended to fall within the scope of the appended claims.
- The above detailed description describes various features and functions of the disclosed systems, devices, and methods with reference to the accompanying figures. In the figures, similar symbols typically identify similar components, unless context dictates otherwise. The example embodiments described herein and in the figures are not meant to be limiting. Other embodiments can be utilized, and other changes can be made, without departing from the spirit or scope of the subject matter presented herein. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations, all of which are explicitly contemplated herein.
- With respect to any or all of the ladder diagrams, scenarios, and flow charts in the figures and as discussed herein, each block and/or communication may represent a processing of information and/or a transmission of information in accordance with example embodiments. Alternative embodiments are included within the scope of these example embodiments. In these alternative embodiments, for example, functions described as blocks, transmissions, communications, requests, responses, and/or messages may be executed out of order from that shown or discussed, including substantially concurrent or in reverse order, depending on the functionality involved. Further, more or fewer blocks and/or functions may be used with any of the ladder diagrams, scenarios, and flow charts discussed herein, and these ladder diagrams, scenarios, and flow charts may be combined with one another, in part or in whole.
- A block that represents a processing of information, such as a block of
method 1100 described above, may correspond to circuitry that can be configured to perform the specific logical functions of a herein-described method or technique. Alternatively or additionally, a block that represents a processing of information may correspond to a module, a segment, or a portion of program code (including related data). The program code may include one or more instructions executable by a processor for implementing specific logical functions or actions in the method or technique. The program code and/or related data may be stored on any type of computer readable medium such as a storage device including a disk or hard drive or other storage medium. - The computer readable medium may also include non-transitory computer readable media such as computer-readable media that stores data for short periods of time like register memory, processor cache, and random access memory (RAM). The computer readable media may also include non-transitory computer readable media that stores program code and/or data for longer periods of time, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. A computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device.
- Moreover, a block that represents one or more information transmissions may correspond to information transmissions between software and/or hardware modules in the same physical device. However, other information transmissions may be between software modules and/or hardware modules in different physical devices.
- The particular arrangements shown in the figures should not be viewed as limiting. It should be understood that other embodiments can include more or less of each element shown in a given figure. Further, some of the illustrated elements can be combined or omitted. Yet further, an example embodiment can include elements that are not illustrated in the figures.
- While various aspects and embodiments have been disclosed herein, other aspects and embodiments will be apparent to those skilled in the art. The various aspects and embodiments disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.
Claims (21)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US17/379,920 US20220009100A1 (en) | 2014-04-01 | 2021-07-19 | Software Interface for Authoring Robotic Manufacturing Process |
Applications Claiming Priority (4)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US14/242,828 US9701018B2 (en) | 2014-04-01 | 2014-04-01 | Software interface for authoring robotic manufacturing process |
US15/616,096 US10486308B2 (en) | 2014-04-01 | 2017-06-07 | Software interface for authoring robotic manufacturing process |
US16/598,655 US11103997B2 (en) | 2014-04-01 | 2019-10-10 | Software interface for authoring robotic manufacturing process |
US17/379,920 US20220009100A1 (en) | 2014-04-01 | 2021-07-19 | Software Interface for Authoring Robotic Manufacturing Process |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/598,655 Continuation US11103997B2 (en) | 2014-04-01 | 2019-10-10 | Software interface for authoring robotic manufacturing process |
Publications (1)
Publication Number | Publication Date |
---|---|
US20220009100A1 true US20220009100A1 (en) | 2022-01-13 |
Family
ID=53015902
Family Applications (4)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/242,828 Active 2035-08-21 US9701018B2 (en) | 2014-04-01 | 2014-04-01 | Software interface for authoring robotic manufacturing process |
US15/616,096 Active 2034-11-22 US10486308B2 (en) | 2014-04-01 | 2017-06-07 | Software interface for authoring robotic manufacturing process |
US16/598,655 Active 2034-06-10 US11103997B2 (en) | 2014-04-01 | 2019-10-10 | Software interface for authoring robotic manufacturing process |
US17/379,920 Abandoned US20220009100A1 (en) | 2014-04-01 | 2021-07-19 | Software Interface for Authoring Robotic Manufacturing Process |
Family Applications Before (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/242,828 Active 2035-08-21 US9701018B2 (en) | 2014-04-01 | 2014-04-01 | Software interface for authoring robotic manufacturing process |
US15/616,096 Active 2034-11-22 US10486308B2 (en) | 2014-04-01 | 2017-06-07 | Software interface for authoring robotic manufacturing process |
US16/598,655 Active 2034-06-10 US11103997B2 (en) | 2014-04-01 | 2019-10-10 | Software interface for authoring robotic manufacturing process |
Country Status (4)
Country | Link |
---|---|
US (4) | US9701018B2 (en) |
EP (1) | EP3126101B1 (en) |
CN (1) | CN107073710B (en) |
WO (1) | WO2015153466A1 (en) |
Families Citing this family (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP6350037B2 (en) * | 2014-06-30 | 2018-07-04 | 株式会社安川電機 | Robot simulator and robot simulator file generation method |
US9492923B2 (en) * | 2014-12-16 | 2016-11-15 | Amazon Technologies, Inc. | Generating robotic grasping instructions for inventory items |
KR101962889B1 (en) * | 2015-07-27 | 2019-03-28 | 한국전자통신연구원 | Robot motion data providing apparatus using a robot to work and method therefor |
WO2017044892A1 (en) | 2015-09-11 | 2017-03-16 | Autodesk, Inc. | Multi-tool manufacturing system |
KR20180015774A (en) * | 2015-09-25 | 2018-02-14 | 두산로보틱스 주식회사 | Method and apparatus for controlling robot |
US10703083B2 (en) | 2017-01-13 | 2020-07-07 | Autodesk, Inc. | Multi-tool scheduling for cooperative manufacturing |
WO2019021044A1 (en) * | 2017-07-26 | 2019-01-31 | Abb Schweiz Ag | Method and system for performing an operation with an industrial robot |
WO2019021045A1 (en) * | 2017-07-26 | 2019-01-31 | Abb Schweiz Ag | Method and system for parmeter based operation of an industrial robot |
JP2019029765A (en) * | 2017-07-27 | 2019-02-21 | オムロン株式会社 | Slave device, industrial network system, and slave device control method |
CN107378954A (en) * | 2017-09-22 | 2017-11-24 | 苏州诺乐智能科技有限公司 | A kind of industrial robot Automated condtrol software |
EP3483755B1 (en) * | 2017-11-09 | 2022-07-13 | Dassault Systèmes | Additive manufacturing of a 3d part |
US10960548B2 (en) | 2017-11-21 | 2021-03-30 | Service Robotics & Technologies, Inc. | Map-based framework for the integration of robots and smart devices |
WO2019139815A1 (en) | 2018-01-12 | 2019-07-18 | Duke University | Apparatus, method and article to facilitate motion planning of an autonomous vehicle in an environment having dynamic objects |
WO2019152769A1 (en) * | 2018-02-05 | 2019-08-08 | Advanced Solutions Life Sciences, Llc | System and workstation for the design, fabrication and assembly of 3-dimensional constructs |
TWI822729B (en) | 2018-02-06 | 2023-11-21 | 美商即時機器人股份有限公司 | Method and apparatus for motion planning of a robot storing a discretized environment on one or more processors and improved operation of same |
CN108415427B (en) * | 2018-02-11 | 2021-06-11 | 昆山艾派科技有限公司 | Path splicing method for multi-step continuous movement of robot |
CN108284445B (en) * | 2018-03-06 | 2018-12-11 | 崔海龙 | Personalization machine people's custom-built system |
WO2019183141A1 (en) | 2018-03-21 | 2019-09-26 | Realtime Robotics, Inc. | Motion planning of a robot for various environments and tasks and improved operation of same |
GB2573544A (en) * | 2018-05-09 | 2019-11-13 | Sony Interactive Entertainment Inc | Apparatus control system and method |
EP3924792A4 (en) * | 2019-02-13 | 2022-09-14 | ABB Schweiz AG | Method and apparatus for managing robot program |
CN109807896B (en) * | 2019-02-28 | 2021-05-04 | 深圳镁伽科技有限公司 | Motion control method and system, control device, and storage medium |
CN109982426B (en) * | 2019-03-21 | 2021-04-13 | 中国科学院合肥物质科学研究院 | Unmanned vehicle sensor data offline synchronization method |
CN113905855B (en) * | 2019-04-17 | 2023-08-25 | 实时机器人有限公司 | User interface, system, method and rules for generating motion planning diagrams |
CN110413110A (en) * | 2019-07-05 | 2019-11-05 | 深圳市工匠社科技有限公司 | The control method and Related product of virtual role |
US11787048B2 (en) | 2019-08-30 | 2023-10-17 | Intrinsic Innovation Llc | Robot planning from process definition graph |
EP3792716B1 (en) * | 2019-09-13 | 2022-11-09 | Siemens Aktiengesellschaft | Numerical control with buffering of load setpoints |
US11254015B2 (en) | 2019-09-24 | 2022-02-22 | Thermo Crs Ltd. | Multi-axis gripper for lab automation robot |
US12011836B2 (en) * | 2020-04-08 | 2024-06-18 | Robim Technologies Inc. Of Canada | Cloud based computer-implemented system and method for computer-assisted planning and simulation of robot motions in construction |
US11511419B2 (en) * | 2020-05-01 | 2022-11-29 | Intrinsic Innovation Llc | Task planning for measurement variances |
US12083683B2 (en) * | 2020-08-20 | 2024-09-10 | Robim Technologies Inc. Of Canada | Cloud based computer-implemented system and method for grouping action items on visual programming panel in robot simulator |
CN112207818A (en) * | 2020-08-28 | 2021-01-12 | 扬州哈工科创机器人研究院有限公司 | Six-axis mechanical arm control method and system |
CN115302507B (en) * | 2022-08-10 | 2024-10-29 | 武汉理工大学 | Intelligent decision method for disassembly process of digital twin-driven industrial robot |
TWI828369B (en) * | 2022-10-14 | 2024-01-01 | 創博股份有限公司 | Method of automatically establishing a robot model |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7652238B2 (en) * | 2004-09-08 | 2010-01-26 | Sick Ag | Method and apparatus for detecting an object through the use of multiple sensors |
US20100191372A1 (en) * | 2009-01-26 | 2010-07-29 | Fanuc Ltd | Production system having cooperating process area between human and robot |
JP5035768B2 (en) * | 2006-04-18 | 2012-09-26 | 独立行政法人産業技術総合研究所 | Safety device for human robot coexistence work |
US9043025B2 (en) * | 2012-08-31 | 2015-05-26 | Rethink Robotics, Inc. | Systems and methods for safe robot operation |
US9452531B2 (en) * | 2014-02-04 | 2016-09-27 | Microsoft Technology Licensing, Llc | Controlling a robot in the presence of a moving object |
Family Cites Families (40)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5511147A (en) | 1994-01-12 | 1996-04-23 | Uti Corporation | Graphical interface for robot |
US5798627A (en) | 1995-01-04 | 1998-08-25 | Gilliland; Malcolm T. | Method for simultaneous operation of robot welders |
US5950006A (en) | 1997-11-05 | 1999-09-07 | Control Technology Corporation | Object-oriented programmable controller |
US6161051A (en) | 1998-05-08 | 2000-12-12 | Rockwell Technologies, Llc | System, method and article of manufacture for utilizing external models for enterprise wide control |
US6718533B1 (en) | 1999-02-26 | 2004-04-06 | Real-Time Innovations, Inc. | Method for building a real-time control system with mode and logical rate |
DE19959330A1 (en) | 1999-12-09 | 2001-06-13 | Kuka Roboter Gmbh | Method and device for controlling a robot |
GB2373347B (en) | 2001-03-07 | 2006-11-22 | Touch Clarity Ltd | Control system to actuate a robotic operating system |
US6612143B1 (en) | 2001-04-13 | 2003-09-02 | Orametrix, Inc. | Robot and method for bending orthodontic archwires and other medical devices |
WO2004033159A1 (en) | 2002-10-11 | 2004-04-22 | Fujitsu Limited | Robot control algorithm construction device, robot control algorithm construction program, robot control device, robot control program, and robot |
DE10305384A1 (en) | 2003-02-11 | 2004-08-26 | Kuka Roboter Gmbh | Method and device for visualizing computer-aided information |
US7890194B2 (en) | 2005-12-13 | 2011-02-15 | Brooks Automation, Inc. | Robotics programming interface |
US20070282480A1 (en) | 2003-11-10 | 2007-12-06 | Pannese Patrick D | Methods and systems for controlling a semiconductor fabrication process |
JP2006187826A (en) | 2005-01-05 | 2006-07-20 | Kawasaki Heavy Ind Ltd | Robot controller |
US7149606B2 (en) | 2005-03-01 | 2006-12-12 | Fanul Robotics America, Inc. | Synchronizing controllers linked by a communications network |
JP4210270B2 (en) | 2005-06-07 | 2009-01-14 | ファナック株式会社 | Robot control apparatus and robot control method |
KR101323597B1 (en) | 2005-09-02 | 2013-11-01 | 니토 로보틱스 인코퍼레이티드 | Multi-function robotic device |
ES2712461T3 (en) | 2006-03-03 | 2019-05-13 | Universal Robots As | Joint for a robot |
US7590680B2 (en) | 2006-06-29 | 2009-09-15 | Microsoft Corporation | Extensible robotic framework and robot modeling |
WO2008008727A2 (en) | 2006-07-10 | 2008-01-17 | Applied Materials, Inc. | Scheduling method for processing equipment |
EP2140316B1 (en) * | 2007-03-29 | 2011-12-28 | iRobot Corporation | Robot operator control unit configuration system and method |
US7684892B2 (en) | 2007-10-30 | 2010-03-23 | Gm Global Technology Operations, Inc. | Process for generating control sequence of operations |
US8082515B2 (en) | 2007-11-30 | 2011-12-20 | National Instruments Corporation | Motion controller that supports standard and custom axes |
JP4465012B2 (en) | 2008-01-31 | 2010-05-19 | ファナック株式会社 | Machining line system that cooperates with the loading / unloading robot |
US8229587B2 (en) | 2008-02-22 | 2012-07-24 | Muratec Automation Co., Ltd. | Semiconductor fabrication facility visualization system with performance optimization |
AT10596U1 (en) | 2008-02-26 | 2009-06-15 | Keba Ag | CONFIGURATION OF MACHINE RUNS |
US8639666B2 (en) | 2008-09-05 | 2014-01-28 | Cast Group Of Companies Inc. | System and method for real-time environment tracking and coordination |
CN102378943A (en) * | 2009-02-03 | 2012-03-14 | 范努克机器人技术美国有限公司 | Method of controlling a robotic tool |
US20100312387A1 (en) | 2009-06-05 | 2010-12-09 | The Boeing Company | Supervision and Control of Heterogeneous Autonomous Operations |
EP2493664B1 (en) | 2009-10-27 | 2019-02-20 | Battelle Memorial Institute | Semi-autonomous multi-use robot system and method of operation |
CN102791214B (en) | 2010-01-08 | 2016-01-20 | 皇家飞利浦电子股份有限公司 | Adopt the visual servo without calibration that real-time speed is optimized |
DE102010012598A1 (en) | 2010-02-26 | 2011-09-01 | Kuka Laboratories Gmbh | Process module library and programming environment for programming a manipulator process |
US8731714B2 (en) | 2010-09-22 | 2014-05-20 | GM Global Technology Operations LLC | Concurrent path planning with one or more humanoid robots |
US10475240B2 (en) * | 2010-11-19 | 2019-11-12 | Fanuc Robotics America Corporation | System, method, and apparatus to display three-dimensional robotic workcell data |
US8660738B2 (en) | 2010-12-14 | 2014-02-25 | Catepillar Inc. | Equipment performance monitoring system and method |
US9160898B2 (en) | 2011-01-25 | 2015-10-13 | Autofuss | System and method for improved video motion control |
JP4775516B1 (en) | 2011-03-14 | 2011-09-21 | オムロン株式会社 | Control device, control method, program, recording medium |
US9010684B2 (en) | 2012-06-27 | 2015-04-21 | The Boeing Company | Automated inspection of soft-tooled hollow structure |
US8833827B2 (en) * | 2012-10-09 | 2014-09-16 | Willow Garage, Inc. | Kinetic and dimensional optimization for a tendon-driven gripper |
KR102000264B1 (en) * | 2013-10-01 | 2019-07-15 | 한국전자통신연구원 | Apparatus for inputting teaching data and apparatus and method for generating teaching command of robot |
US8803844B1 (en) | 2013-10-02 | 2014-08-12 | DigiPuppets LLC | Capacitive finger puppet for use on touchscreen devices |
-
2014
- 2014-04-01 US US14/242,828 patent/US9701018B2/en active Active
-
2015
- 2015-03-30 EP EP15719329.3A patent/EP3126101B1/en active Active
- 2015-03-30 CN CN201580027799.6A patent/CN107073710B/en active Active
- 2015-03-30 WO PCT/US2015/023358 patent/WO2015153466A1/en active Application Filing
-
2017
- 2017-06-07 US US15/616,096 patent/US10486308B2/en active Active
-
2019
- 2019-10-10 US US16/598,655 patent/US11103997B2/en active Active
-
2021
- 2021-07-19 US US17/379,920 patent/US20220009100A1/en not_active Abandoned
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7652238B2 (en) * | 2004-09-08 | 2010-01-26 | Sick Ag | Method and apparatus for detecting an object through the use of multiple sensors |
JP5035768B2 (en) * | 2006-04-18 | 2012-09-26 | 独立行政法人産業技術総合研究所 | Safety device for human robot coexistence work |
US20100191372A1 (en) * | 2009-01-26 | 2010-07-29 | Fanuc Ltd | Production system having cooperating process area between human and robot |
US9043025B2 (en) * | 2012-08-31 | 2015-05-26 | Rethink Robotics, Inc. | Systems and methods for safe robot operation |
US9452531B2 (en) * | 2014-02-04 | 2016-09-27 | Microsoft Technology Licensing, Llc | Controlling a robot in the presence of a moving object |
Non-Patent Citations (2)
Title |
---|
Augustsson, Human and Robot Interaction based on safety zones in a shared work environment, University West, June 2013, 42 pages. * |
Rybski et al., Sensor Fusion for Human Safety in Industrial Workcells, IEEE, October 2012, 8 pages. * |
Also Published As
Publication number | Publication date |
---|---|
WO2015153466A1 (en) | 2015-10-08 |
US20170266810A1 (en) | 2017-09-21 |
EP3126101A1 (en) | 2017-02-08 |
US11103997B2 (en) | 2021-08-31 |
US9701018B2 (en) | 2017-07-11 |
EP3126101B1 (en) | 2022-05-04 |
US20150273685A1 (en) | 2015-10-01 |
US10486308B2 (en) | 2019-11-26 |
CN107073710B (en) | 2020-06-30 |
CN107073710A (en) | 2017-08-18 |
US20200061821A1 (en) | 2020-02-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11103997B2 (en) | Software interface for authoring robotic manufacturing process | |
US11409260B2 (en) | Runtime controller for robotic manufacturing system | |
US9915937B2 (en) | Systems and methods for time-based parallel robotic operation | |
US9623559B2 (en) | Systems and methods for instructing robotic operation | |
US9993924B2 (en) | Closed-loop control system for robotic operation | |
JP7388074B2 (en) | Simulation device, simulation program and simulation method | |
JP7275840B2 (en) | simulation device | |
Iqbal et al. | Intelligent Motion Planning in Human-Robot Collaboration Environments | |
Lucchesi | Collision-Free Cobot Interaction: A machine vision approach for real time human-obstacle avoidance | |
Bischoff et al. | Flexible control of complex kinematic chains |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: BOT & DOLLY, LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LINNELL, JEFFREY;BYRNE, KENDRA;BITTERMAN, MATTHEW;REEL/FRAME:056949/0674 Effective date: 20140702 Owner name: X DEVELOPMENT LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:GOOGLE LLC;REEL/FRAME:056949/0943 Effective date: 20171030 Owner name: GOOGLE LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:BOT & DOLLY, LLC;REEL/FRAME:056949/0842 Effective date: 20140702 |
|
AS | Assignment |
Owner name: INTRINSIC INNOVATION LLC, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:X DEVELOPMENT LLC;REEL/FRAME:057650/0405 Effective date: 20210701 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |