Nothing Special   »   [go: up one dir, main page]

WO2021076333A1 - Intelligent controller and sensor network bus, system and method including smart compliant actuator module - Google Patents

Intelligent controller and sensor network bus, system and method including smart compliant actuator module Download PDF

Info

Publication number
WO2021076333A1
WO2021076333A1 PCT/US2020/053589 US2020053589W WO2021076333A1 WO 2021076333 A1 WO2021076333 A1 WO 2021076333A1 US 2020053589 W US2020053589 W US 2020053589W WO 2021076333 A1 WO2021076333 A1 WO 2021076333A1
Authority
WO
WIPO (PCT)
Prior art keywords
fiber optic
ports
nodes
compliant actuator
optic connector
Prior art date
Application number
PCT/US2020/053589
Other languages
French (fr)
Inventor
Eugene Lee
Original Assignee
Vulcan Technologies Shanghai Co., Ltd.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US16/653,558 external-priority patent/US11269316B2/en
Application filed by Vulcan Technologies Shanghai Co., Ltd. filed Critical Vulcan Technologies Shanghai Co., Ltd.
Priority to CN202080005480.4A priority Critical patent/CN112867997A/en
Publication of WO2021076333A1 publication Critical patent/WO2021076333A1/en

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L12/00Data switching networks
    • H04L12/28Data switching networks characterised by path configuration, e.g. LAN [Local Area Networks] or WAN [Wide Area Networks]
    • H04L12/40Bus networks
    • H04L12/40169Flexible bus arrangements
    • H04L12/40176Flexible bus arrangements involving redundancy
    • H04L12/40182Flexible bus arrangements involving redundancy by using a plurality of communication lines

Definitions

  • the present invention relates to the field of buses. More particularly, the present invention relates to a controller and sensor network bus architecture.
  • a machine automation system for controlling and operating an automated machine.
  • the system includes a controller and sensor bus including a central processing core and a multi-medium transmission intranet for implementing a dynamic burst to broadcast transmission scheme where messages are burst from nodes to the central processing core and broadcast from the central processing core to all of the nodes.
  • a first aspect is directed to a machine automation system for controlling and operating an automated machine.
  • the system comprises a controller and sensor bus including plurality of ports, wherein the bus comprises a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and one or more subnetworks, the central optical fiber transmission networks including a plurality of nodes and one or more gates and the plurality of subnetworks each coupled to a different one of the gates of one of the central transmission networks, the subnetworks including a plurality of subnodes, wherein each of the nodes and the subnodes are associated with one or more of the ports, one or more controllers each coupled with first ones of the ports of the bus and one or more compliant actuator modules each including a first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC), the modules each operably coupled with second ones of the ports via fiber optic cable coupled from the first fiber optic connector to the second ones of the ports, wherein the nodes and the subn
  • each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable.
  • a plurality of the compliant actuator modules are coupled in parallel to an optical splitter and the optical splitter is coupled to one of the second ones of the ports via the fiber optic cable.
  • each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC.
  • the compliant actuator modules are coupled in series such that the first fiber optic connector of a first of the compliant actuator modules is coupled to one of the second ones of the ports via the fiber optic cable and the second fiber optic connector of the first of the compliant actuator modules is coupled to the first fiber optic connector of a second of the compliant actuator modules.
  • each of the compliant actuator modules comprise a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
  • the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, wherein the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor.
  • the first fiber optic connectors are pigtail fiber optic connection points.
  • the second ones of the ports are optical ports coupled with pigtail fiber optic connectors.
  • the automated machine is one of the groups consisting of a robot and a self-driving vehicle.
  • a second aspect is directed to a compliant actuator module for use with a controller and sensor bus.
  • the compliant actuator module comprises one or more compliant actuator motors, one or more magnetic sensors and a control board including a first fiber optic connector, a second fiber optic connector, a system on chip (SoC) and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC, wherein the control board receives sensor data from the magnetic sensors and outputs status messages based on the sensor data via the first fiber optic connector, controls the motors based on control messages input via the first fiber optic connector having a target wavelength, forwards status messages received via the second fiber optic connector out the first fiber optic connector with the optical splitter and forwards the control messages input via the first fiber optic connector having a non-target wavelength.
  • SoC system on chip
  • the module further comprises an image sensor, wherein the control board receives image data from the image sensor and outputs sensor data messages based on the image data via the first fiber optic connector.
  • the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor.
  • the control board comprises a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
  • the first fiber optic connector is a pigtail fiber optic connection point.
  • a third aspect is directed to a method of operating a controller and sensor bus including a plurality of ports for coupling with a plurality of external machine automation devices of a machine automation system, the bus having a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and including a plurality of nodes and one or more gates and a plurality of subnetworks each coupled to a different one of the gates of one of the central optical fiber transmission networks, the subnetworks including a plurality of subnodes.
  • the method comprises coupling one or more controllers with first ones of the ports of the bus, coupling one or more compliant actuator modules with second ones of the ports via fiber optic cable coupled from a first fiber optic connector to the second ones of the ports, the compliant actuator modules each including the first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC) and relaying messages between the controllers and the compliant actuator modules through the bus via the one or more central optical fiber transmission networks with the nodes and the subnodes of the bus, the messages including sensor data from the sensors of the compliant actuator modules and control data from the controllers.
  • SoC system on chip
  • each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable.
  • coupling the compliant actuator modules includes coupling the compliant actuator modules in parallel to an optical splitter and coupling the optical splitter to one of the second ones of the ports via the fiber optic cable.
  • each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC.
  • coupling the compliant actuator modules includes coupling the first fiber optic connector of a first of the compliant actuator modules to one of the second ones of the ports via the fiber optic cable and coupling the second fiber optic connector of the first of the compliant actuator modules to the first fiber optic connector of a second of the compliant actuator modules.
  • each of the compliant actuator modules comprise a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
  • the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, the method further comprises controlling the articulation of the image sensor with the motors and determining the orientation of the image sensor with the magnetic sensors.
  • the first fiber optic connectors are pigtail fiber optic connection points.
  • the second ones of the ports are optical ports coupled with pigtail fiber optic connectors.
  • the automated machine is one of the groups consisting of a robot and a self-driving vehicle. BRIEF DESCRIPTION OF THE DRAWINGS
  • Figure 1 illustrates a machine automation system according to some embodiments.
  • Figure 2 illustrates an intelligent controller and sensor intranet bus according to some embodiments.
  • Figure 3 illustrates a tree topology of an intelligent controller and sensor intranet bus according to some embodiments.
  • Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the system according to some embodiments.
  • Figure 5 illustrates a method of operating a machine automation system including an intelligent controller and sensor intranet bus according to some embodiments.
  • Figure 6 A illustrates an exemplary GEM packet format according to some embodiments.
  • Figure 6B illustrates a detailed view of a GEM packet header format according to some embodiments.
  • Figure 6C illustrates a detailed view of a GEM header format for a node report message according to some embodiments.
  • Figure 6D illustrates a detailed view of a first variant of a GEM header format for a root port bandwidth grant message according to some embodiments.
  • Figure 6E illustrates a detailed view of a second variant of a GEM header format for a root port bandwidth grant message according to some embodiments.
  • Figure 6F illustrates a detailed view of a GEM header format for a control message according to some embodiments.
  • Figure 7 A illustrates a Broadcast-PHY-Frame according to some embodiments.
  • Figure 7B illustrates a Burst-PHY-Frame according to some embodiments.
  • Figure 7C illustrates a gate Burst-PHY-Frame according to some embodiments.
  • Figure 8 illustrates a method of operating the intelligent controller and sensor intranet bus according to some embodiments.
  • FIG. 9 illustrates a smart compliant actuator (SC A) and sensor module according to some embodiments.
  • Figure 10A illustrates a first variant of a control board of SCA and sensor module according to some embodiments.
  • Figure 10B illustrates a second variant of a control board of SCA and sensor module according to some embodiments.
  • Figure IOC illustrates a third variant of a control board of SCA and sensor module according to some embodiments.
  • Figures 11 A and 1 IB illustrate a machine automation system including coupled SCA and sensor modules according to some embodiments.
  • Figure 12 illustrates a method of operating a controller and sensor bus according to some embodiments.
  • Embodiments described herein are directed to a machine automation system, method and device for controlling and operating an automated machine.
  • the system, method and device including a controller and sensor bus including a central processing core and a multi medium transmission intranet for implementing a dynamic burst to broadcast transmission scheme where messages are burst from nodes to the central processing core and broadcast from the central processing core to all of the nodes.
  • the system, method and device provides the advantage of high speed performance despite combining lower speed network medium as well as one unified software image for the full intranet system including all gate, node and root ports enabling simplified software architecture, shorter product development cycle, and easier system level debug, monitoring and trouble shooting remotely.
  • the system, method and device provides a unique intranet system architecture specially defined and optimized for machine automation applications.
  • Figure 1 illustrates a machine automation system 100 according to some embodiments.
  • the system 100 comprises one or more external devices 102 operably coupled together with an intelligent controller and sensor intranet bus 104.
  • the system 100 is able to be a part of an automated device such as a self- driving vehicle, an automated industrial machine or an automated self-controlled robot.
  • the system 100 is able to be a part of other machine automation applications.
  • the devices 102 are able to comprise one or more of sensor devices (e.g. ultrasonic, infrared, camera, light detection and ranging (LIDAR), sound navigation and ranging (SONAR), magnetic, radio detection and ranging (RADAR)), internet devices, motors, actuators, lights, displays (e.g.
  • Each of the devices 102 is able to be operably wired and/or wirelessly coupled with the bus 104 via one or more bus input/output (IO) ports (see Figure 2).
  • IO bus input/output
  • the system 100 comprises a discrete amount of external devices 102 and buses 104, more or less devices 102 and/or buses 104 are contemplated.
  • Figure 2 illustrates the intelligent controller and sensor intranet bus 104 according to some embodiments.
  • the bus 104 comprises an intranet formed by a central core 200 that is coupled with one or more gates 202 and a plurality of edge nodes 204 (each having one or more external IO ports 99) via one or more central transmission networks 206, and coupled with one or more edge sub-nodes 208 (each having one or more external IO ports 99) via one or more sub-networks 210 that extend from the gates 202.
  • the bus 104 forms a network tree topology where the central networks 206 branch from the core 200 (e.g. root ports 230 of the core) to edge nodes 204 and gates 202, and the subnetworks 210 branch from the gates 202 to sub-nodes 208 and/or sub-gates 202'.
  • the core 200 is able to see all of the nodes 204 and sub-nodes 208 (as the gates 202 and sub-gates 202' are transparent to the core 200).
  • one or more of the gates 202 are directly coupled with IO ports 99 without a node (e.g. to couple with external CPU, GPU, AI cores and/or solid state drives (SSD)).
  • the ports 99 are able to be any kind of interface port such as peripheral component interconnect express (PCIe), mobile industry processor interface (MIPI), Ethernet, universal serial bus (USB), general purpose input output (GPIO), universal asynchronous receiver/transmitter (UART), inter-integrated circuit (I 2 C) and/or other types of ports.
  • PCIe peripheral component interconnect express
  • MIPI mobile industry processor interface
  • USB universal serial bus
  • GPIO general purpose input output
  • UART universal asynchronous receiver/transmitter
  • I 2 C inter-integrated circuit
  • the bus 104 comprises a discrete amount of ports 99, cores 200, nodes 204, 208, gates 202, networks 206, 210, other elements and components thereof, more or less ports 99, cores 200, nodes 204, 208, gates 202, networks 206, 210, other elements and/or components there of are contemplated.
  • the central transmission networks 206 are able to comprise connection media that is faster/lower latency than the connection media of the subnetworks 210 coupled to a gate 202 of that central transmission network 206.
  • the subnetworks 210 are able to comprise connection media that is faster/lower latency than the connection media of the subnetworks 210' coupled to a gate 202' of the subnetwork 210 and so on for each iterative subnetwork.
  • This network/subnetwork connection media speed/latency relationship enables the bus 104 to prevent the slowing of the processing of the entire bus 104 despite still including the slower connection media as describe in detail below.
  • one or more of the subnetworks 210, 210' and/or the central networks 206 are able to have the same or other connection media speed/latency relationships.
  • connection media of the central transmission networks 206 comprises optical fiber cables 212 split using optical splitters 214 (e.g 2-to-l splitters) and having optical transceivers 216 to couple to and received data from the nodes 204, 208.
  • connection media of the subnetworks 210 comprises optical connection media (e.g. like the central transmission networks 206, but possibly slower rating), wireless connections (e.g. radio frequency transceivers 218), copper connections (e.g. twisted-pair copper wires 220 optionally split using analog splitters 222( e.g.
  • the bus 104 supports multi-rate traffic transmissions where depending on the latency/speed, connectivity and/or distance requirements of the data/traffic/extemal devices 102, different nodes/networks are able to be used to coupled to the bus 104 while still providing the desired throughput. For example, for high speed, low latency and long-distance requirements the optical connection media of the central network is able to be used by coupling to the nodes 204.
  • the other networks 210 are able to be used depending on cost, speed, connection and/or distance requirements.
  • the central networks 206 are passive optical networks and/or the copper subnetworks 210 are active networks.
  • one or more of the nodes 204 is coupled to a controller area network (CAN) 226 such that the node inputs data from each of the controllers coupled to the controller are network.
  • CAN controller area network
  • one or more of the subnetworks 210 are able to be a CAN coupled with the core 200 via one of the gates 202.
  • the bus 104 is able to utilize a multi-layered addressing scheme where the root ports 230, IO ports 99, nodes 204, 208, 234 and/or gates 202 are able to use node, epoch and GEM identifying addresses for directing messages through the bus 104.
  • each of the root ports 230, nodes 204, 208, 234 and gates 202 are able to be assigned a node identifier (node-ID), with the nodes 204, 208 and gates 202 also being assigned at least one epoch identifier (epoch-ID) and at least one GEM identifier (GEM-ID).
  • the epoch-IDs are able to be used to identify the source/destination of messages in the network 206, 210 (e.g. node/gate devices and their IO ports, embedded CPUs and/or other types of services) while at the same time the GEM-IDs are able to be used to identify the targets of messages (e.g. sets and subsets of the node/gate devices and their IO ports, embedded CPUs and/or other types of services).
  • the epoch-IDs are able to be used for the transmission/routing of messages throughout the network 206, 210 while the GEM-IDs are able to be used by the devices themselves (via the ports 99) to determine whether to capture received/broadcast messages as being targeted to them.
  • the nodes/gates are able to be assigned multiple epoch-IDs and multiple GEM-IDs.
  • SLA service level agreement
  • the node-ID of each of the nodes 204, 208 and gates 202 is able to map to one or a plurality of epoch-IDs which are able to map to one or a plurality of GEM-IDs.
  • a node 204, 208 coupled with two IO ports 99 is able to have a single node-ID, two epoch-IDs (one for each port 99) and ten GEM-IDs (one associated with the first epoch-ID and first port 99 and nine associated with the second epoch-ID and second port 99).
  • the node-IDs and epoch-IDs are unique to each node/gate/port, the GEM-IDs are able to be shared between nodes/gates/ports.
  • ports 99 of the same node 204, 208 or different ports 99 of different nodes 204, 208 are able to both be associated with matching or overlapping sets GEM-IDs.
  • the gates 202 are also able to be assigned one or more virtual node-IDs for the ports 99 directly coupled with the gate 202. Like the regular nodes, these virtual nodes represented by the gates 202 are able to be assigned multiple epoch-IDs and multiple GEM-IDs depending on the SLA profde of the gate 202 (which is able to correspond to the devices coupled to the port(s) 99 of the virtual node/gate).
  • the other nodes 234 and cores 232 are each able to have one or more GEM-IDs along with a global node-ID, but do not need to be assigned epoch-IDs, which are not required because messages to and from these nodes 234 to the core 200 are wholly within the core 200.
  • the number of GEM-IDs assigned to each of the nodes 234 and cores 232 is able to be determined based on the SLA profile for that node 234 or core 232 (which is able to correspond to the devices coupled to the port(s) 99 of the node 234).
  • Each of the core switch 220, root ports 230, nodes 204, 208, 234, and/or gates 202 are able to maintain and update a local SLA table that indicates the mapping between each of the node-IDs, epoch-IDs and GEM-IDs.
  • the bus addressing provides the advantage of using epoch-IDs and/or node-IDs to facilitate simplified burst/broadcast messaging between nodes, gates and the core within the network 100, while at the same time using GEM-IDs facilitate any desired more complex messaging between the devices/IO ports 99 and/or the core themselves.
  • the bus 104 is able to encapsulate all input data and internally generated data (e.g. control, operation and management messages) into a generic encapsulation mode (GEM) for transport across the bus 104 intranet.
  • GEM acts as a unique standardized data and message container for transmitting data between nodes and/or to the core 200 via the bus 104 intranet.
  • the input data is able to be encapsulated into the GEM format at each of the nodes as it enters the bus 104 and is routed through the core 200 (where it is decapsulated for processing and re-encapsulated for transmission) and onto its destination node which decapsulates the data back to the original format for egress to the target external device 102 or other destination.
  • This input data is able to be from various sources (e.g. devices 102,
  • CAN 2236 input via the ports 99 at the nodes 204, 208, 234 or gates 202 and/or the embedded CPU cores 232.
  • GEM packet There are two types of GEM formats: GEM packet and GEM control.
  • the GEM packet format comprises a GEM header plus a GEM payload (e.g. length from 8 bytes to 4 kilobytes).
  • GEM payload e.g. length from 8 bytes to 4 kilobytes.
  • the GEM packet format what is used to encapsulate the input port data, packets and messages at the ingress (e.g. nodes, ports).
  • GEM packet format to carry Ethernet packets from local gate 202 and/or node 204, 208 through bus 104 after GEM encapsulation to far-end gate 202 and/or node 204 (e.g. this is able to be for internet and Wi-Fi interfaces through Ethernet Port or PCIe Ports);
  • GEM packet format to carry sensor data from local gate 202 and/or node 204, transmit through bus 104 after GEM encapsulation to far-end gate 202 and/or node 204 (e.g. CAN bus data, Camera (MIPI) Frame data,
  • GEM packet format to carry jumbo size data and packets and transmit through fragmentation and de-fragmentation scheme, from local node 204, 208 to far-end node 204, 208. This is able to include fragmentation, defragmentation and re-ordering / re-transmission functions;
  • PLOAM physical layer operation, administration and maintenance
  • NMCI node management control interface
  • OAM operations, administration and maintenance
  • GEM packet format to carry CPU/PCIe access CMD/DATA from core 200 and local gate 202 and/or node 204 through bus 104 after GEM encapsulation, to far-end local gate 202 and/or node 204 (e.g. CPU 232 access target device 102 from NODE-to-NODE through PCIe, USB, I2C, UART and GPIO interfaces).
  • GEM packet format for VPN channel application between local-nodes 204, 208 to far nodes 204, 208 through bus 104.
  • the GEM control message format comprises message plus extended message (e.g. length 8 bytes + 8 bytes ).
  • the GEM control message format is able to be used in the bus 104 for internal network management and control purposes, including messages of dynamic bandwidth allocation (DBA) reporting, DBA-Granting, GEM RX- Acknowledge, GEM Flow-Control, GEM Power-Management, GEM-Sniffer, GEM-Remote messages and/or other types of control messages.
  • DBA dynamic bandwidth allocation
  • nodes 204 are responsible for encapsulating/decapsulating data to/from GEM packet and GEM control message format.
  • This scheme is able to expand PCIe interface protocol from point-to-point topology to point-to-multi-point topology and extend the interface distance from short reach to long reach.
  • FIGS 6A-F illustrate an exemplary GEM packet format and GEM header formats according to some embodiments.
  • an GEM packet 600 is able to comprise a header 602 and a corresponding payload 604.
  • the header is able to be a set size (e.g. 8 bytes) and the payload is able to vary in length (e.g. length from 8 bytes to 4 kilobytes) and for control packet the header is able to be, for example, 8 bytes with or without one or more 8 byte extensions.
  • Figure 6B illustrates a detailed view of a GEM packet header format according to some embodiments.
  • the header 602 comprises a GEM type field 606, a payload length indication field 608, an encryption key index field 610 (e.g. AES Key Index), a node/epoch ID field 612, a GEM-ID field 614, a GEM packet type field 616, a transmission sequence identifier field 618, an acknowledgment required field 620, a last fragment indication field 622 and a header error correction/check (HEC) field 622.
  • the fields are able to be omitted and/or one or more additional fields are able to be added.
  • the GEM type field 606 is two bits
  • the payload length indication field 608 is twelve bits
  • the encryption key index field 610 is two bits
  • the node/epoch ID field 612 is twelve bits
  • the GEM-ID field 614 is twelve bits
  • the GEM packet type field 616 is three bits
  • the transmission sequence identifier field 618 is six bits
  • the acknowledgment required field 620 is one bit
  • the last fragment indication field 622 is one bit
  • the header error correction/check (HEC) field 622 is thirteen bits.
  • the GEM type field 606 indicates which type of header 602 (and thus which type of packet) the GEM packet 600 is.
  • the GEM type field is able to indicate that the header 602 is one or more of a packet header, a bandwidth grant message header (e.g. transmitted from a root port 230 to a gate/node), a bandwidth report message header (e.g. transmitted from a gate/node to a root port 230) and/or a control message (e.g. between one or more of the root ports 230, the gates 202 and/or the nodes 204, 208, 234).
  • a bandwidth grant message header e.g. transmitted from a root port 230 to a gate/node
  • a bandwidth report message header e.g. transmitted from a gate/node to a root port 230
  • a control message e.g. between one or more of the root ports 230, the gates 202 and/or the nodes 204, 208, 234.
  • the payload length indication field 608 indicates the length of the payload 604 of the packet 600.
  • the encryption key index field 610 indicates the type of encryption to use on the packet 600. For example, the encryption key index field 610 is able to be used as an index value within an encryption table to identify one or more of: whether to encrypt the packet or not, which key to use to encrypt the packet, and/or which method of encryption to use.
  • the node/epoch ID field 612 is able to identify either the source node or the destination node of the packet 600. For example, for a GEM packet 600 being burst from a node to the core, the field 612 is able to be or represent the node’s epoch-ID to indicate the source of the packet 600. As another example, for a GEM packet 600 being broadcast from a root port 230 to the nodes/gates within its network 206, 210, the field 612 is able to be or represent the destination’s node -ID (including a unicast node-ID, a multicast node -ID and/or a broadcast node-ID).
  • the GEM-ID field 614 is able to be or represent the source node’s data/packet/message identifier for a point to point message, or is able to be or represent the destination node’s GEM-ID (e.g. including CAN message GEM-IDs, sensor data GEM-IDs and/or ethemet packet GEM-IDs) for point to multi-point messages.
  • GEM-ID e.g. including CAN message GEM-IDs, sensor data GEM-IDs and/or ethemet packet GEM-IDs
  • the GEM format provides the advantage of enabling the bus 104 to identify both the immediate source and/or destination nodes via the node/epoch ID field 612 while also enabling the target devices/port/services to be identified using the GEM-ID field 614.
  • the GEM packet type field 616 is able to indicate the type and format of the header of the message encapsulated within the GEM format (e.g. as received from the devices 102 and/or through the ports 99).
  • the field 616 is able to indicate that the message header is a PLOAM message, a node management and control interface (NMCI) message, a CAN command message, sensor data, an ethemet packet, CPU-IO (e.g. PCIe/USB) message and/or a node operation and control report (NOCR) message.
  • NMCI node management and control interface
  • NOCR node operation and control report
  • the acknowledgment required field 620 is able to indicate if an acknowledgment message in response to the message is require and the transmission sequence identifier field 618 is able to identify the transmission sequence number of the packet 600 within a set of packets from the source node and/or an epoch-ID thereof (for a packet being burst from the node to the core 200). In some embodiments, it requires an acknowledgment message from the receiving root port 230 when indicated by the acknowledgment required field 620. For a packet broadcast from the root port 230 to a node/gate, the transmission sequence identifier field 618 is able to identify the transmission sequence number of the unicast/broadcast/multi-cast GEM-ID (e.g.
  • the last fragment indication field 622 is able to indicate if this packet 600 is the last fragment of a series of fragments of a large packet and the header error correction/check (HEC) field 622 is able to be used to check the header 602 for errors.
  • HEC header error correction/check
  • Figure 6C illustrates a detailed view of a GEM header format for a node report message according to some embodiments.
  • the header 602 comprises a GEM type field 606, a report message type field 624, a source epoch-ID field 626, a report total size field 628, a report threshold size field 630, a report sequence number field 632, one or more source node virtual output queue (VOQ) status fields 634 (e.g. CPU-IO, PLOAM, NMCI, CAN, Sensor, Ethernet, or other types), a report priority field 636 and a header error correction/check (HEC) field 622.
  • VOQ source node virtual output queue
  • the fields are able to be omitted and/or one or more additional fields are able to be added.
  • the GEM type field 606 is two bits
  • the report message type field 624 is two bits
  • the source epoch-ID field 626 is twelve bits
  • the report total size field 628 is fourteen bits
  • the report threshold size field 630 is eight bits
  • the report sequence number field 632 is five bits
  • the one or more source node virtual output queue status fields 634 are each one bit (or a single field of six bits)
  • the report priority field 636 is two bits
  • the header error correction/check (HEC) field 622 is thirteen bits.
  • one or more of the fields are able to be larger or smaller.
  • the report message type field 624 indicates which type of report header 602 (and thus which type of report message) the GEM packet 600 is.
  • the report message type field 624 is able to indicate that the header 602 is one or more of an invalid report message, a node report message for itself (e.g. where the epoch-ID of the source of the packet is mapped to the node-ID of the source of the packet), a node report message for another node (e.g. where the epoch-ID of the source of the packet is not mapped to the node-ID of the source of the packet), and/or a dying gasp report message (e.g. a message that needs/requests top priority).
  • the source epoch-ID field 626 is able to be or represent: the source node’s epoch- ID (e.g. for a report for PLOAM and NMCI plus CAN/sensor/ethernet queue flags), the CAN’s epoch-ID (e.g. for a report for the CAN), the epoch-ID of one of the sensors/nodes (e.g. for a report for the sensor), the ethemet epoch-ID (e.g. for a report for ethemet packets) and/or a PCIe/USB epoch-ID (e.g. for a PCIe/USB report message).
  • the source node’s epoch- ID e.g. for a report for PLOAM and NMCI plus CAN/sensor/ethernet queue flags
  • the CAN’s epoch-ID e.g. for a report for the CAN
  • the report total size field 628 is able to indicate the total size of the GEM data within the VOQ (for that epoch-ID and/or Node -ID), whereas the report threshold size field 630 is able to indicate the GEM packet boundary(ies) within the VOQ (e.g. for use when determining the size of burst windows granted for the epoch and/or node).
  • the report sequence number field 632 is able to indicate which number in the sequence that the message is (e.g. if there are a sequence of related report messages in order to determine if one is lost or mis-sequenced).
  • the one or more source node virtual output queuing (VOQ) status fields 634 are each able to indicate a status of the source node with respect to a particular function/type of data (e.g. CPU/IO, PLOAM, NMCI, CAN, sensor, ethernet).
  • the report priority field 636 is able to indicate what priority to give the message (e.g. best efforts, normal bandwidth request priority, CAN message request priority, dying gasp request priority).
  • Figures 6D and E illustrate a detailed view of two variants of a GEM header format for a root port bandwidth grant message according to some embodiments.
  • the header 602 is able to comprise a GEM type field 606, an epoch-ID field 638, a start time field 640, a grant size field 642, a grant flag field 644, a report command field 646, a grant command field 648, a force wake-up indicator (FWI) field 650, a burst profile field 652 and a header error correction/check (HEC) field 622.
  • the fields are able to be omitted and/or one or more additional fields are able to be added.
  • the GEM type field 606 is two bits
  • the epoch-ID field 638 is twelve bits
  • the start time field 640 is fourteen bits
  • the grant size field 642 is fourteen bits
  • the grant flag field 644 is one bit
  • the report command field 646 is three bits
  • the grant command field 648 is two bits
  • the force wake-up indicator field 650 is one bit
  • the burst profile field 652 is two bits
  • the header error correction/check (HEC) field 622 is thirteen bits.
  • one or more of the fields are able to be larger or smaller.
  • the epoch-ID field 638 is able to be or represent the epoch-ID of the node or node-ID that the message is for.
  • the start time field 640 is able to indicate a starting time of the grant window that is being granted to the target node (e.g. epoch of that node) and the grant size field 642 is able to indicate the size/duration of the grant window.
  • the grant flag field 644 is able to indicate whether the window was granted.
  • the report command field 646 is able to indicate what reporting is requested from the node/epoch/port.
  • the report command field 646 is able to indicate one or more of: no node request to send (RTS) status report or force node to report RTS message to port for blackbox and diagnostic test; combined with one or more of: PLOAM and NMCI reporting only forced reporting of CPU- IO messages, CAN messages and sensor data plus PLOAM/NMCI; forced reporting for ethernet packets plus CPU-IO/CAN/sensor and PLOAM/NMCI; and/or forced full report of PLOAM/NMCI/CPU-IO/CAN/sensor/ethernet plus a node operation and control report (NOCR).
  • the grant command field 648 is able to indicate what type of messages/data are granted the burst window.
  • the grant command field 648 is able to indicate one or more of: the window is not for PLOAM and NMCI messages; the grant window is only for PLOAM messages; the grant window is only for NMCI messages; and/or the grant is for PLOAM, NMCI and NOCR messages.
  • the FWI field 650 is to indicate whether to force a sleeping node to wake-up and the burst profde field 652 is able to indicate a burst configuration (e.g. length, pattern and/or other characteristics of the SOB delimiter, EOB delimiter and/or preamble).
  • the header 602 is able to be substantially the same as the header of Figure 6D except without the report command field 646 and the FWI field 650.
  • the grant command field 648 is able to be six bits.
  • the grant command field 648 is able to be larger or smaller.
  • the grant command field 648 is able to indicate a GEM bandwidth grant of different types.
  • the field 648 is able to indicate a bandwidth grant for: all VOQ/CoS (class of service) based on the nodes ’s output scheduling settings, for CAN messages only, for sensor data only, dying gasp messages only and/or for both CAN messages and sensor data. Additionally, the field 648 is able to force power saving for the node-ID where the node replies with an acknowledge message.
  • all VOQ/CoS class of service
  • Figure 6F illustrates a detailed view of a GEM header format for a control message according to some embodiments.
  • the header 602 comprises a GEM type field 606, a control message type field 654, one or more control message fields 656 and a header error correction/check (HEC) field 622.
  • GEM type field 606 is two bits
  • the control message type field 654 is four bits
  • the one or more control message fields together are forty-five bits
  • the header error correction/check (HEC) field 622 is thirteen bits.
  • one or more of the fields are able to be larger or smaller.
  • the control message type field 654 is able to indicate what type of control message the message is (e.g. so the control message fields 656 and their offsets are known for processing). In some embodiments, the control message type field 654 indicates one or more of: a report acknowledgment message; a CAN acknowledgment message; a flow control message; a power saving message; and IO event message (e.g. dying gasp); a run-time status message; and/or a timestamp update (e.g. from port to node).
  • the control message fields 656 are able to include various control message fields based on the type of control message (as indicated in control message type field 654).
  • the GEM format provides the benefit of enabling the bus 104 to encapsulate varying input data and messages of significantly different types of networks (e.g. controller area networks, optical networks, sensor device broadcasting networks, wireless networks, CPU access networks) to one unique format (GEM).
  • This unique format is then able to facilitate high speed standardized processing and transmission of the varied data input in both burst and broadcast messages thereby enabling the efficient operation of the multi network multi-device bus architecture required for modem machine automation applications.
  • the broadcast messages are formatted into a Broadcast-PHY-Frame defined by: Preamble + Start-of-Frame-Delimiter + Frame -Payload, wherein the frame payload includes multiple GEM-Packet data and GEM-Control messages.
  • the Broadcast-PHY-Frame is able be a fixed frame size (e.g. between 25-125ps). Alternatively, greater or smaller frame sizes are able to be used. For example, for central networks 206 and subnetworks 210 with less node devices 204, 208, the frame size is able to be smaller (e.g. 25 ps or 50ps).
  • the Broadcast-PHY-Frame is constructed to carry GEM-Packet and GEM-Control messages from the root ports 230 to the gate 202 and/or nodes 204, 208, 234 through the networks 206, 210 including optical, copper and wireless networks.
  • the burst messages are formatted into a Burst-PHY-Frame defined by: Preamble + Start-of-Frame-Delimiter + Frame Payload + End-of-Frame-Delimiter, wherein the frame payload includes one or more GEM-Packet data and GEM-Control messages.
  • the Burst-PHY-Frame size is able to vary depending on the total Burst-Window size of node/gate granted by root port HDBA and/or gate DBA.
  • the max size of Burst-PHY-Frame (from a gate 202 or a node 204, 208, 234) cannot exceed the max Broadcast-PHY-Frame size (e.g.
  • the Burst-PHY-Frame is constructed to carry GEM-Packet and GEM-Control messages from gates 202 and/or nodes 204, 208, 234 to the root ports 230 and/or gates 202 via the networks 206, 210 including optical, copper and wireless networks.
  • Figure 7A illustrates a Broadcast-PHY-Frame 700 according to some embodiments. As shown in Figure 7A, the Broadcast-PHY-Frame 700 comprises a physical synchronization block for broadcast (PSBbc) 702 and a broadcast framing sublayer frame 704 including a GEM control message 706, one or more GEM packets 600 and a framing sublayer (FS) trailer 708.
  • PSBbc physical synchronization block for broadcast
  • FS framing sublayer
  • each of the GEM packets 600 include a header 602 and a payload 604 as described above.
  • the broadcast FS frame is FEC protected.
  • Figure 7B illustrates a Burst-PHY-Frame 710 according to some embodiments.
  • the Burst- PHY-Frame 710 comprises a physical synchronization block unicast start of burst delimiter (PSBuc_sd) 712, a burst framing sublayer (FS) 714 and a physical synchronization block unicast end of burst delimiter (PSBuc ed) 716.
  • PSBuc_sd physical synchronization block unicast start of burst delimiter
  • FS burst framing sublayer
  • PSBuc ed physical synchronization block unicast end of burst delimiter
  • the PSBuc sd 712 is able to include a preamble 718 and a start of burst (SOB) delimiter 720 and the PSBuc ed 716 is able to include an end of burst (EOB) delimiter 722.
  • the burst FS 714 is able to include a FS header 724, one or more epochs 726 and an FS trailer 708.
  • Each of the epochs 726 are able to include one or more GEM packets 600 having a header 602 and a payload 604 as described above.
  • the burst FS frame is FEC protected.
  • the structure 710 enables a sniffer, analytics engine or other element to monitor the traffic within the bus 104 because it enables the element to determine the end of each burst frame based on the EOB delimiter despite not knowing/accessing the size of the frame.
  • Figure 7C illustrates a gate Burst-PHY-Frame 728 according to some embodiments.
  • the gate Burst-PHY-Frame 728 is able to comprise one or more Burst-PHY-Frames 710 combined together into a single combined burst-PHY-frame having a single preamble 729 and one or more gaps 730.
  • the gates 202 are able to receive burst frames 728 from one or more subnodes 208 as well as one or more IO ports 99 (for which they serve as a virtual node) and combine those frames 728 into a combined gate Burst-PHY-Frame 728 as shown in Figure 7C.
  • the system 100 provides the advantage of more efficient message communication via combined burst frames as well as less overhead per frame by using only a single preamble for the combined frame as a whole instead of a separate preamble for each combined burst frame (whose preamble can be up to 256 bytes each or more).
  • Figure 8 illustrates a method of operating the intelligent controller and sensor intranet bus 103 according to some embodiments.
  • one or more of the nodes 204, 208 input one or more messages from the one or more of the devices 102 coupled to the one or more of the ports 99 at the step 802.
  • the nodes 204, 208 encapsulate the messages into the generic encapsulation mode (GEM) format for transmission to the central processing core 200 at the step 804.
  • GEM generic encapsulation mode
  • the core 200 decapsulates, processes and re-encapsulates the messages back into the GEM format for broadcast to their destination(s) at the step 808.
  • the nodes 204, 208 decapsulate the messages as received from the core 200 from the GEM format to an original format of the input data as received from one of the devices 102 at the step 810.
  • the input messages are input from nodes 234 inside the core 200 they are able to be input and processed by the core 200 (without being encapsulated) and only encapsulated by the core 200 for broadcast if their destination is one or more nodes 204, 208 outside the core 200.
  • the method provides the advantage of enabling the communication of many different types of data (e.g. sensor, controller bus, ethernet, or other types of data), more efficient message communication via combined burst frames, and less overhead per frame by using only a single preamble for the combined frame as a whole instead of a separate preamble for each combined burst frame.
  • data e.g. sensor, controller bus, ethernet, or other types of data
  • the core 200 is able to comprise a core switch 228, one or more root ports 230 (internal ports), a central processing unit 232 and one or more core nodes 234 having IO ports 99 (external ports).
  • the core 200 further comprises a secure memory (e.g. secure digital (SD) memory) node 236 for storing data in a black box memory 238.
  • SD node 236 and/or memory 238 are able to be omitted.
  • the core nodes 234 enable a user to couple a user plug-in module (e.g. CPU core, WIFI LTE/5G, User Application software) directly to the core 200 bypassing the networks 206, 210.
  • a user plug-in module e.g. CPU core, WIFI LTE/5G, User Application software
  • the core switch 228 comprises a forwarding engine element, a queuing buffer manager and a traffic manager.
  • Forwarding engine element is able to comprise a plurality of forwarding engines. For example, it is able to include one engine used for L2/L3/L4 Ethernet header parser, lookup and classification/access control list (ACL) function, including L2 medium access control (MAC) Address learning and forwarding functions, L3 internet protocol (IP) Address to GEM -ID Routing/mapping. Additional, one engine is able to be used for GEM Header message parser, lookup, ACL and forwarding and/or another is able to be used to support DOS attack functions to protect the bus 104 from external internet DOS attack.
  • ACL classification/access control list
  • IP internet protocol
  • the GEM-Queuing-Buffer Manager is able to be a centralized buffering architecture, which employs link-list based buffer and queuing memory methods combining store-N-forward and cut-through forwarding schemes. For latency sensitive GEM-Packet and GEM-Messages, it is able to use a cut-through forwarding scheme and for congestion GEM-Packets it is able to use store-N-forward scheme. Both schemes are able to be dynamically mixed together and dynamically switched between each other depending on the run-time traffic congestion situations.
  • the GEM-Traffic Manager supports GEM-ID and NODE-ID base dual-token policing, single-token rate-limiting and output shaping functions, including related management information base (MIB) counters.
  • MIB management information base
  • GEM-ID base weighted random early detection (WRED) and Tail-Drop functions are able to be supported as well as early traffic congestion detection and indication and feedback mechanisms to notify hybrid dynamic bandwidth allocation mechanisms (HDBA), root ports 230, gates 202 and nodes 204, 208, 234 to slow down traffic transmission in order to avoid traffic congestion from occurring.
  • HDBA hybrid dynamic bandwidth allocation mechanisms
  • the core switch 228 is able to provide the functions of on ingress, the switch 228 receives GEMs from one or more of the root ports 230, local nodes 234, computer 232 and/or other IO ports, processes the GEMs and on egress, forwards and transmits the received GEMs to one or more of the root ports 230, local nodes 234, computer 232 and/or other IO ports.
  • the switch 228 is able to accept GEM-Packets from multiple sources; perform GEM and Ethernet L2/L3/L4 header parsing, L2 MAC lookup and learning, GEM message and 5 -tuple ACL and classification; modify GEM-Header and GEM payload Ethernet header (if necessary); and store and forward GEM-Packet (or cut-through buffer memory) to one or multiple hybrid automatic repeat request (HARQ) functional blocks and the broadcast-MAC of one or more root ports 230.
  • GEM-Packets from multiple sources; perform GEM and Ethernet L2/L3/L4 header parsing, L2 MAC lookup and learning, GEM message and 5 -tuple ACL and classification; modify GEM-Header and GEM payload Ethernet header (if necessary); and store and forward GEM-Packet (or cut-through buffer memory) to one or multiple hybrid automatic repeat request (HARQ) functional blocks and the broadcast-MAC of one or more root ports 230.
  • HARQ hybrid automatic repeat request
  • the switch 228 In performing this processing and/or forwarding function, the switch 228 is able to support hybrid store-and forward and cut-through forwarding schemes in order to reduce propagation latency for latency sensitive GEMs and provide big enough buffering for over burst GEM traffic. Additionally, the switch 228 is able to support instant- flow-control mechanisms within the bus 104, including hybrid dynamic bandwidth allocation and granting to ensure overall quality of service (QoS) across the bus 104. Further, the switch 228 is able to support L2/L3/L4 ACL and classification, L2 MAC address learning and forwarding, L3 IP address to GEM-ID routing/mapping, as well as DOS attack protection. Finally, the switch 228 is able to support QoS scheduling, GEM buffering WRED/Tail dropping, node and/or GEM policing and output shaping functions. Root Ports
  • the root ports 230 are able to comprise a root transmission MAC, a root reception MAC, a security engine (e.g. advanced encryption standard (AES)), a forward error correction (FEC) engine, a hybrid dynamic bandwidth allocation (HDBA) engine, an activation processor (e.g. activation state machine) and a burst-mode SERDES IP. Alternatively, one or more of the above elements are able to be omitted.
  • the transmission MAC of each of the root ports 230 is responsible for accepting GEMs ready for egress from switch 228 and/or HARQ; map and pack the GEMs into a broadcast frame format (e.g.
  • Broadcast PHY-Frame structure Broadcast PHY-Frame structure); and broadcast the GEMs to all of the gates 202 and/or nodes 204 on the central transmission network 206 to which the root port 230 is coupled (e.g. through root SERDES and optical/copper network broadcast domains).
  • the reception MAC of each of the root ports 230 is responsible for receiving GEMs in a burst frame format (e.g. Burst-PHY-Frame structure) from Burst-Mode SERDES and gates 202 and/or nodes 204,
  • the root ports 230 are each able to receive burst traffic from the nodes 204 and/or gates 202 (forwarded from nodes 208 in the subnetwork 210 of the gate 202), convert the burst traffic to the correct format for processing by the switch 228 and then reformat and broadcast output traffic to all of the nodes 204 and nodes 208 (via the gates 202) to destinations as directed by the switch 228.
  • the hybrid dynamic bandwidth allocation (HDBA) engine is responsible for receiving reports about bandwidth usage, traffic congestion and other factors (e.g. NODE-DBA Reports); performing HDBA analysis based on an SLA profile for the node/port/device associated with each report, the DBA-Report data itself and committed information rate (CIRypeak information rate (PIR) feedback; and granting burst windows to each NODE device and assigned port/EPOCH-ID.
  • the HDBA engine inputs data from each of the nodes 204, 208 (of the network 206 associated with the root port 230 and subnetworks 210 thereof) and/or other sources about bandwidth usage/traffic congestion and dynamically allocates burst transmission window start times and/or sizes to each of those nodes 204, 208.
  • the gate 202 that provides access to the nodes 208 is transparent to the HDBA engine.
  • the gate 202 receives the desired data and performs the burst transmission within the assigned windows for each of the nodes 208 of the gate’s 202 subnetwork 210.
  • the HDBA engine is also able issue reporting acknowledgment messages (GEM -Report- ACK message) to nodes 204, 208 to confirm that the report messages (GEM -DBA Reports) were received.
  • the root Activation State-Machine is responsible for performing and completing node 204, 208, 234 device activation and registration through activation processes and procedures by exchanging physical layer operations, administration and maintenance (PLOAM) GEM messages between nodes 204, 208, 234 and the root port 230.
  • the security engine is able to be an AES-128/256 encryption and decryption functional block used for both the reception and transmission MACs. Alternatively, other encryption is able to be used.
  • the forward error correction (FEC) engine is used for controlling errors in data transmission over unreliable or noisy communication channels. In some embodiments, the FEC engine uses Reed Solomon FEC coding schemes of RS(255,216) and RS(225,232) for 10G and 2.5G data rates, respectively.
  • the FEC engine is able to user low-density parity-check (LDPC) schemes and/or other FEC algorithms.
  • the burst-mode SERDES uses fast clock and data recovery (CDR) locking mode to ensure proper burst messages (e.g. burst-PHY-Frames) are received correctly.
  • CDR clock and data recovery
  • the fast locking function of CDR is required in fiber-cut, fast fail-over and protection switch recovery.
  • the root ports 230 receive broadcast data distribution service (DDS) messages from nodes 204, 208 that notify the root port 230 that new nodes/devices have joined and registered to bus 104. Accordingly, the root ports 230 are configured to always listen and accept these data distribution service (DDS) messages from the switch 228 and new node’s 204, 208 declaration of joining the bus 104, and update the Root-Port SLA profde database and settings to reflect the newly added nodes/devices.
  • DDS broadcast data distribution service
  • the edge nodes 204, 208, 234 provide a bridge function within the bus 104 to interface with external devices 102 via the IO ports 99 on one side and connect to bus intranet 104 on the other side.
  • the nodes 204, 208, 234 construct and transmit burst messages (e.g. Burst-PHY-Frames of the data encapsulated as GEMs) through the bus 104 to the other nodes 204, 208 via the root port 230 (of the network 206 of which they are a part or a subnetwork 210 thereof).
  • burst messages e.g. Burst-PHY-Frames of the data encapsulated as GEMs
  • the nodes 204, 208, 234 receive broadcast message (e.g. Broadcast-PHY-Frames of the data encapsulated as GEMs) from other nodes 204, 208 via the root port 230 (of the network 206 of which they are a part or a subnetwork 210 thereof), extract the data from the broadcast messages (e.g. GEMs from RX BC-PHY-Frames), and filter and accept the data that belongs (is addressed to) the node 204, 208.
  • broadcast message e.g. Broadcast-PHY-Frames of the data encapsulated as GEMs
  • the broadcast messages e.g. GEMs from RX BC-PHY-Frames
  • the edge nodes 204, 208 are able to comprise one or more IO ports 99, an encapsulation/decapsulation engine, a HARQ block and a node MAC.
  • Each of the ports 99 is able to be one of a CPU interface (e.g. PCIe, USB and UART), a sensor interface (e.g. MIPI, analog to digital converter (ADC), GPIO), an internet interface (e.g. Ethernet, EtherCAT, and CAN-Bus), and a motor module interface (e.g. pulse width modulation (PWM), I 2 C, ADC and GPIO).
  • a CPU interface e.g. PCIe, USB and UART
  • a sensor interface e.g. MIPI, analog to digital converter (ADC), GPIO
  • an internet interface e.g. Ethernet, EtherCAT, and CAN-Bus
  • PWM pulse width modulation
  • the encapsulation/decapsulation engine accepts input data from the ports 99 and encapsulates received data packets, commands (CMD) and messages received from the internet ports (e.g. Ethernet, Wi-Fi), sensor interfaces, motor module interface and CPU (e.g. PCIe and USB) to the GEM format at the ingress.
  • CMS commands
  • the nodes 204, 208 then are able to output to the encapsulated messages (e.g. GEMs) to the HARQ and/or node transmission MAC (described below).
  • the egress accepts GEM-packets from the node reception MAC (received from the root port 230 and/or another node 204, 208, 234) and decapsulates the GEM back to the original data format (as received from the coupled device 102) for output to the device 102 via one of the ports 99.
  • the HARQ of the nodes 204, 208 perform the hybrid automatic-repeat-request function to ensure that the GEM-Packets are delivered to their destination node or nodes 204, 208, 234 successfully.
  • the HARQ is able to be built-in with a repeat transmit timer, transmit GEM list flag table and receipt acknowledgment checking function (e.g. GEM RX- Acknowledge) to trigger GEM re-transmission when timer time-out occurs without receiving the acknowledgment.
  • the node MAC comprises a transmission MAC (TX MAC), a reception MAC (RX MAC), a security engine (e.g. AES), a forward error correction (FEC) engine, a DBA-Report engine and SERDES IP.
  • TX MAC is responsible for mapping/packing GEMs into a burst structure (e.g. Burst-PHY-Frame structure) and transmitting the burst messages to root ports 230 and/or nodes 204, 208, 234 during the burst window for the node granted by the dynamic burst allocation engine of the root port 230 for that node.
  • the RX MAC is responsible for receiving and terminating broadcast messages (e.g.
  • Broadcast-PHY-Frames from root ports 230 and/or nodes 204, 208, 234, extracting GEMs from the broadcast message format, parsing and accepting GEMs addressed to it (e.g. addressed to one of its ports 99) based on the node’s SLA Profde setting, and subsequently outputting the data to the encapsulation/decapsulation engine.
  • the DBA report engine reports total data packet and message in queues (e.g. EPOCH Queues) to the HDBA engine of the associated root port 230 through the burst reporting (as described above). Additionally, the DBA report engine accepts GEM-Grant messages from the HDBA of the associated root port 230 and/or the DBA of the associated gate 202, and prepares the node transmission MAC to build a burst message (e.g. Burst-PHY-Frame) with the GEMs stored in the queues (e.g. EPOCH Queues).
  • a burst message e.g. Burst-PHY-Frame
  • the node activation processor is responsible for performing and completing the node 204, 208, 234 activation process and procedures between nodes 204, 206, 234 and root ports 230.
  • the security engine is able to be an AES- 128/256 encryption and decryption functional block used for both the reception and transmission MACs. Alternatively, other encryption is able to be used.
  • the FEC engine is used for controlling errors in data transmission over unreliable or noisy communication channels. In some embodiments, the FEC engine uses Reed Solomon FEC coding schemes of RS(255,216) and RS(225,232) for 10G and 2.5G data rates, respectively.
  • the burst-mode SERDES uses fast clock and data recovery (CDR) locking mode to ensure fast fiber-cut, fast fail-over and protection switch recovery.
  • the nodes 204, 206, 234 are able to broadcast a DDS message to entire bus 104 to inform and notice the root ports 230, switch 228, gates 202 and/or other nodes 204, 206, 234 that a new device has joined and registered to bus 104 at that node 204, 208, 234. Further, the nodes 204, 206, 234 are able to listen to DDS messages from the switch 228 and other new the nodes’ 204, 206, 234 declaration of joining the bus 104 and update their global SLA profde database and settings based on the DDS messages.
  • the gates 202 are able to comprise a node MAC (with multiple Virtual node State-Machines and buffering), an adaptive domain bridge (ADB), a root port MAC (with built-in gate DBA functionality/gate DBA), a gate SLA profile database and a burst-mode SERDES.
  • the node MAC comprises one or more of a transmission MAC, reception MAC, security engine (e.g. AES), FEC engine, DBA report functional module, SERDES functional module and/or multiple sets (e.g. one for each node within the subnetwork 210) of virtual node processors, virtual node profdes and settings, and related MIB counters and reporting logics.
  • the transmission MAC receives GEMs from the gate ADB and maps and packs then into their associated virtual node burst structure (e.g. Burst-PHY-Frame structure) based on the gate’s virtual node SLA Profile database settings. Further, the transmission MAC aggregates multiple virtual node burst structures (e.g. Burst-PHY-Frames) into one gate burst structure (e.g. GATE/Turbo Burst-PHY-Frame) and transmits burst message to the root port 230 through the network 206 based on the granted burst window for those nodes 208 received from the HDBA of the root port 230.
  • the node reception MAC receives broadcast messages (e.g.
  • Broadcast-PHY-Frames from the root port 230, extracts GEMs from the messages, parses the headers of the GEMs, determines which messages are for nodes 208 within the subnetwork 210 of the gate 202 based on the GEM -Headers and virtual nodes SLA Profile database settings and outputs those messages to the ADB.
  • the ADB performs a bridging function between the node MAC and the root MAC of the gates 202. Specifically, in the broadcast direction (from the root port 230 to the nodes 208), the ADB receives GEMs from node reception MAC and performs a GEM header lookup, checking and filtering function based on the gate virtual node profile database in order to accept GEMs belonging to nodes 208 of the gate’s 202 subnetwork 210. The ADB is then able to output those GEMs to root port transmission MAC of the gate 202.
  • the ADB receives GEMs from root reception MAC, stores them in their associated virtual node buffer memory, and output them to the virtual node transmission MAC when their burst window start time arrives.
  • the root port MAC of the gates 202 comprise a transmission MAC, a reception MAC, a security engine (e.g. AES), an FEC engine, a gate DBA and burst mode SERDES modules.
  • the transmission MAC is responsible for accepting GEMs from ADB, mapping and packing the GEMs into a broadcast format (e.g. Broadcast-PHY-Frame structure), and outputting the broadcast formatted frames to burst-mode SERDES.
  • the reception MAC is responsible for receiving burst messages (e.g. Burst-PHY-Frames) from burst-mode SERDES (e.g.
  • the DBA of the gate 202 is an extension HDBA of the root ports 230.
  • the gate DBA grants and allocates node burst windows based on the gate DBA SLA profile settings (which is a subset of the root HDBA).
  • the gate SLA profile database includes a list of node identifiers belonging to this gate 202 (e.g.
  • the burst mode SERDES accepts broadcast messages (e.g. Broadcast-PHY-Frames) from the root transmission MAC and transmits to nodes 208 in the subnetwork 210 in the broadcast transmission direction. In reception direction, the burst mode SERDES receives burst messages (e.g. Burst-PHY-Frames) from nodes 208 through the subnetwork 210 and outputs them to the root reception MAC for message/ffame termination and GEM extraction.
  • broadcast messages e.g. Broadcast-PHY-Frames
  • burst messages e.g. Burst-PHY-Frames
  • gates 202 The main function of gates 202 is to extend the central transmission network 206 of one of the root ports 230 by bridging to one or more subnetworks 210 (and the nodes 208 therein ) through adaptive bridging.
  • the gates 202 are able to burst messages from the nodes 208 and/or other gates 202' within their subnetwork 210 to the root port 230 of the network 206 they are in as if the burst traffic were coming from nodes within the central transmission network 206.
  • the gates 202 are able to broadcast messages received from other nodes 204, 208, 234, the switch 228 and/or root port 230 to the nodes 208 and/or other gates 202' within their subnetwork 210 they are in as if the nodes 208 and/or other gates 202' were within the central transmission network 206.
  • the gates 202 are able to extend the central transmission networks 206 to additional nodes 208 and/or different types of subnetworks 210 while maintaining a burst/broadcast communication method within the central transmission networks 206.
  • the burst window granting mechanism from node 208 to gate 202 to root 230 is able to comprise the following steps.
  • the DBA of the gate 202 is a subset of the HDBA of the root port 230 (of the network 206 that the gate 202 is a part of) and therefore is transparent to the root port 230 and nodes 208.
  • the gate 202 uses the message header (e.g. GEM-Header) to lookup gate SLA profde database for GEM forwarding information.
  • the gate 202 uses the header data to determine if the grant message is for any of the nodes 208 within its subnetwork 210 as indicated in the gate SLA profile database. If the grant message is not for any of the nodes 208 of its subnetwork 210 the gate 202 drops the grant message, otherwise, the gate stores the message in its virtual node database, updates the database and broadcasts a new window grant message (e.g. GEM- Grant message) to all the nodes/gates in its subnetwork 210 that is directed to the node 208 to which the original grant message was directed.
  • a new window grant message e.g. GEM- Grant message
  • the node 208 provides a burst message to the gate 202 and the gate 202 formats and/or otherwise prepares the message for bursting to the root port 230 at the burst window start indicated in the received window grant message for that node 208.
  • gate 202 is able to adjust the grant window indicated in this new grant message to be at least a predetermined amount of time before the grant window indicated in the original grant message.
  • this amount of time provides the gate 202 time to receive and format the burst data from the node 208 before bursting the data from the gate 202 to the root port 230 at the time indicated by the original window grant message.
  • the gate 202 is able to aggregate the messages from multiple different nodes (e.g. multiple Burst-PHY-frames) into a single bigger burst message (e.g. GATE Burst-PHY-Frame).
  • root port 230 and gates 202 are able to maintain a group-membership list table and be aware of the virtual nodes 208 that each of the gates 230 below to as a group.
  • a node 208 issues a report message (e.g. GEM-Report) to HDBA of the root port 230
  • the gate 203 is able to intercept the report message, modify it to include the GEMs data temporarily stored in gate’s 202 virtual node buffer memory if there is any, and issue a new report message to HDBA of the root port 230.
  • the gates 202 are able to combine reporting messages from the nodes in their subnetworks 210 in order to make the reporting more efficient.
  • HDBA of the root ports 230 are issuing a grant message (e.g. GEM-Grant message) to nodes 208 that are in a subnetwork 210, because they are aware of all of the nodes 208 that are in that subnetwork 210 (e.g. via the virtual node database), the HDBA of the root ports 230 are able to ensure that the grant windows for nodes 208 that belong to the same gate 202 and/or subnetwork 210 are in sequence/continuous order so that the gate 202 is able to combine and/or burst all the virtual node’s burst messages (e.g. burst-PHY-Frames) without each having a preamble except for the first one.
  • This provides the benefit of reducing preamble overhead and increasing the burst bandwidth efficiency (especially for small bursts of GEM-Control messages).
  • the gates 202 receive burst messages (e.g. burst-PHY-frames) from burst-mode SERDES and far-end nodes 208, extracts the GEMs from the messages in the root reception MAC of the gate 202, stores the GEMs in their associated virtual NODE buffer memory and waits for the virtual node burst window grant to come in from the root port 230 for those virtual nodes 208. Then, the gates 202 are able to map and pack the stored GEMs for that node 208 and other nodes 208 back into the burst message format thereby aggregating multiple burst messages together into one bigger burst message in the node transmission MAC of the gates 202. Finally, the gates 202 are able to transmit this bigger burst message to the SERDES and to the root port 230 through the network 206 based on granted burst windows (e.g. the multiple consecutive virtual node burst windows of that gate 202).
  • burst messages e.g. burst-PH
  • the gates 202 are able to extend central networks 206 to the subnetworks 210 while being transparent to both the root port 230 for their network 206 and the nodes 208 in their subnetwork 210.
  • the gates 202 are able to act like virtual nodes and receive broadcast messages (e.g. Broadcast-PHY-Frames) from the root ports 230, extract the GEMs from the messages, drop any GEMs that are not directed to one of the nodes 208/gates 202' in their subnetwork 210 (e.g. as indicated by the message headers and the gate SLA profile database).
  • broadcast messages e.g. Broadcast-PHY-Frames
  • the gates 202 are able to use store-N-forward and/or cut-through schemes to pack and map the GEMs back into the root port broadcast message structure (e.g. Broadcast-PHY-Frame structure) in a root transmission MAC of the gate 202 and broadcast the new broadcast message to all the nodes 208 and/or gates 202' in its subnetwork 210.
  • the root port broadcast message structure e.g. Broadcast-PHY-Frame structure
  • the bus 104 operates using a burst/broadcast communication scheme wherein all data messages from the nodes 204, 208, 234 (and gates 202) are funneled to the core 200 using a burst transmission method where transmission windows that are dynamically adjustable in size (by the core 200) are granted to the nodes 204, 208, 234 such that they (or a gate 202 on their behalf) are able transmit their data messages as a “burst” within the granted window.
  • the gate 202 (acting as a root port of that network 210) receives the bursted message from the node 208 through the subnetwork 210 and then subsequently bursts the message through the central network 206 to the core 200 (as if the node 208 was a part of the central network 206). In doing this burst communication, the gate 202 is able to aggregate burst messages from multiple nodes 208 within the subnetwork 210 thereby increasing efficiency and reducing the effects of the subnetwork’s 210 possibly increased latency relative to the central network 206.
  • this is able to be repeated for gates 202' within subnetworks 210 that provide a gate way to sub-subnetworks 210' and so on to support any number of “chained/gated” networks.
  • the gate 202 is able to be transparent to the core 200 and nodes 208 in this process such that messages do not need to be addressed to the gate 202.
  • the core 200 receives these messages (from one or more root ports 230 coupling the core 200 to each of the central networks 206), processes them (including modifying and/or determining their target destination), and broadcasts them (and any messages originating in the core 200) onto whichever central transmission network 206 the target node 204, 208, 234 (or gate 202 representing the target node 208) for that message is located.
  • the gate 202 bridging to that subnetwork 210 is able to receive/intercept the message from the core an rebroadcast the message to all of the node 208 (and/or gates 202') on the subnetwork 210.
  • Any broadcast messages for target nodes 204 not on the subnetwork 210 are able to be discarded by the gate 202 in order to increase efficiency. Again, this process is transparent and able to be repeated by gates 202' within subnetworks 210 and so on for any number of chained networks to broadcast the messages through the networks.
  • all the nodes 204, 208, 234 (and gates 202) on each of the networks 206 (and subnetworks 210 coupled thereto) receive all of the messages from the core 200 broadcast on that network 206 and merely need to look for which messages are directed to them while discarding the others.
  • the nodes 204, 208, 234 when the nodes 204, 208, 234 receive data from one or more external devices 102 through one or more of their IO ports 99, they store the data in a GEM-ID queue buffer memory and burst a report message (e.g. GEM-Report) to the root port 230 of the central network 206 that they are in (either directly or through one or more gates 202 if they are in a subnetwork 210 of the central network 206) and wait to be granted a burst window to transmit the input data.
  • a report message e.g. GEM-Report
  • the gates 202 are able to collect and aggregate report messages from a plurality of the nodes 208 (and or gates 202') in their subnetwork 210 into a single bigger report message that the gate 202 is able to more efficiently burst to the root port 230 during the burst window for those ports 208.
  • the nodes 204, 208, 234 are able to encapsulate the input data into the GEM format (fragmenting GEMs exceeding a predefined size into smaller GEMs), encrypt GEMs with the security key of the node 204, 208, 234, update the HARQ table, map and pack the GEMs into a burst format (e.g. Burst- PHY-Frame format) and perform encoding (e.g. FEC RS(255,216) encoding). Subsequently, upon grant and arrival of the burst window for each of the nodes, the nodes burst the GEMs including the input data to the associated root port 230.
  • a burst format e.g. Burst- PHY-Frame format
  • FEC RS(255,216) encoding e.g. FEC RS(255,216) encoding
  • the HDBA of the root ports 230 receive all of the report messages from the nodes 204, 208 (and/or gates 202) and perform a DBA analysis for each of the nodes 204, 208 based on the SLA profile database, latency sensitive level, traffic congestion feedback, committed information rate (CIR)/peak information rate (PIR) feedback and/or other factors to determine grant window burst size and start-time for each of the nodes 204, 208.
  • CIR committed information rate
  • PIR peak information rate
  • the broadcast messages from the root ports 230 are the same size, whereas the burst windows from the nodes 204, 208 to the root ports 230 are able to vary in size as dynamically assigned by the HDBA.
  • the gates 202 upon receipt of the broadcast grant messages targeting nodes 208 within their subnetwork 210 (or a subnetwork thereof), broadcast new grant messages to all of the nodes 208 with the subnetwork 210. Specifically, these new grant messages are able to specifying burst windows that occur before the time indicated by the original/root port grant window.
  • the gates 202 are able to make up for inefficiencies and/or slower aspects of the subnetworks 210 such that they do not slow down the efficiency of the central transmission networks 206.
  • the root ports 230 Upon receipt of the burst messages including the GEMs (including the input data from the external devices 102), the root ports 230 are able to perform decoding (e.g. FEC RS(255,216) decoding) and error correction on the burst messages to decode and correct any transmission errors. The root ports 230 are then able to extract the GEMs from the burst messages (e.g. the transmission frame format), decrypt the extracted GEMs (e.g. with AES-128/256 and a source-node security key), bypass the GEM fragmentation block and pass GEMs to the switch 228.
  • decoding e.g. FEC RS(255,216) decoding
  • error correction e.g. FEC RS(255,216) decoding
  • the root ports 230 are then able to extract the GEMs from the burst messages (e.g. the transmission frame format), decrypt the extracted GEMs (e.g. with AES-128/256 and a source-
  • the switch 228 For each of the GEMs, the switch 228 is then able to perform a GEM-Header lookup, parse and classify Ethernet L2/L3 address and headers, process GEM forward flow-chart and determine GEM forwarding destination info, store the GEM in (cut-through) buffer-memory, and output the GEM to HARQ and to the destination root port 230 (e.g. the root port 230 whose network 206 or subnetwork 210 thereof includes the destination node 204, 208) based on the SLA database QoS output scheduler.
  • the destination root port 230 e.g. the root port 230 whose network 206 or subnetwork 210 thereof includes the destination node 204, 208 based on the SLA database QoS output scheduler.
  • the root ports 230 receive the GEMs, perform GEM encryption (e.g. AES-128/256 encryption) with target node’s (or broadcast GEM’s) security key, pack and map GEMs into a broadcast message structure (e.g. Broadcast-Frame structure), encode the message (e.g. FEC RS(255,216) encoding), and finally broadcast the broadcast messages to all of the nodes 204, 208 in that root port’s network 206 and subnetworks 210 thereof. If the node 208 is within a subnetwork 210, the gate 202 to that subnetwork receives the broadcast message and broadcasts the message to all of the nodes 208 within the subnetwork 210.
  • GEM encryption e.g. AES-128/256 encryption
  • target node’s or broadcast GEM’s
  • pack and map GEMs into a broadcast message structure e.g. Broadcast-Frame structure
  • encode the message e.g. FEC RS(255,216) encoding
  • the gates 202 filter out any broadcast messages that are not targeted to nodes 208 within its subnetwork 210 (or a subnetwork thereof) and only broadcasts the broadcast messages that do target one of those nodes 208.
  • the gates 202 are able to rebroadcast all of the broadcast messages to the nodes 208 within its subnetwork 210 without determining if the messages relate to one of those nodes 208. All the nodes 204, 208 monitor the received broadcast messages, processing those intended for the node 204, 208 and discarding the others. Specifically, for the non-discarded messages, the nodes 204, 208 decode and error correct the messages (e.g.
  • the bus 104 and system 100 provides the benefit of being able to combine multiple different networks having varying input data, varying processing speeds and data constraints while still maintaining low latency and high throughput needed for machine automation systems. This is a unique intranet system architecture and specially defined and optimized for such machine automation applications.
  • Figure 4 illustrates a block diagram of an exemplary computing device 400 configured to implement the system 100 according to some embodiments.
  • the external devices 102 are able to include some or all of the features of the device 400 described below.
  • a hardware structure suitable for implementing the computing device 400 includes a network interface 402, a memory 404, a processor 406, I/O device(s) 408 (e.g. reader), a bus 410 and a storage device 412.
  • a hardware structure suitable for implementing the computing device 400 includes a network interface 402, a memory 404, a processor 406, I/O device(s) 408 (e.g. reader), a bus 410 and a storage device 412.
  • I/O device(s) 408 e.g. reader
  • bus 410 e.g. reader
  • the storage device 412 is able to include a hard drive, CDROM, CDRW, DVD, DVDRW, flash memory card or any other storage device.
  • the computing device 400 is able to include one or more network interfaces 402.
  • An example of a network interface includes a network card connected to an Ethernet or other type of LAN.
  • FIG. 408 are able to include one or more of the following: keyboard, mouse, monitor, display, printer, modem, touchscreen, button interface and other devices.
  • the operating software/applications 430 or function(s)/module(s) thereof are likely to be stored in the storage device 412 and memory 404 and processed as applications are typically processed. More or fewer components shown in Figure 4 are able to be included in the computing device 400.
  • machine automation system hardware 420 is included.
  • the computing device 400 in Figure 4 includes applications 430 and hardware 420 for the system 100, the system 100 is able to be implemented on a computing device in hardware, firmware, software or any combination thereof.
  • Figure 5 illustrates a method of operating a machine automation system 100 including an intelligent controller and sensor intranet bus 104 according to some embodiments.
  • the nodes 204, 208 receive input data from a plurality of the external devices 102 via one or more ports 99 of the bus 104 at the step 502.
  • the nodes 204, 208 burst the input data as burst messages to the core 200 in variable size burst windows at the step 504.
  • the HDBA of the root ports 230 dynamically adjusts the burst window start time and size of the variable burst window and assign the adjusted window the corresponding node 204, 208 in a broadcast grant window message based on data traffic parameters reported from that one of the nodes 204, 208.
  • the gates 202 aggregate two or more burst messages including input data and/or traffic reporting received from the nodes 208 into single larger burst reporting or input data message for bursting to the core 200. In such embodiments, the gates 202 are able to omit portions of the received burst messages (e.g. preambles) in order to enhance the efficiency of the bus 104. In some embodiments, upon receiving the broadcast window grant messages from the core 200, the gates 202 adjust the original time of the burst window to an earlier time and broadcast the adjusted broadcast window grant messages to the nodes 208.
  • the nodes 208 burst their data to the gates 202 before the window granted by the root port 230 such that the gates 202 are able to combine multiple burst messages together and burst them in the later original time window.
  • the core 200 processes and broadcasts the input data as broadcast messages to each of the nodes 204, 208 within the central network 206 and subnetworks 210 required to reach the target node 204, 208 of the message at the step 506.
  • the target node 204, 208 converts data of the broadcast message into a format accepted by the device 102 coupled to the node 204, 208 and outputs the data to the device 102 at the step 508.
  • the method provides the advantage of enabling the bus 104 to maintain high speed despite the use of lower speed network mediums.
  • the devices 102 are able to be device modules.
  • Figure 9 illustrates a smart compliant actuator (SCA) and sensor module 900 according to some embodiments.
  • the SCA and sensor module 900 is able to be one or more of the devices 102 of the machine automation system 100 described herein.
  • the smart compliant actuator (SCA) and sensor module 900 allows deviations from its own equilibrium position, depending on the applied external force, wherein the equilibrium position of the compliant actuator is defined as the position of the actuator where the actuator generates zero force or zero torque.
  • the SCA and sensor module is able to comprise one or more motors 902, one or more sensors 904 and/or a control board 906 (for controlling the motors 902 and/or sensors 904) coupled together via a device network 908.
  • this type of module 900 is able to perform high-bandwidth and/or low-latency required machine automation tasks (e.g. coupled with one or more controller devices 102 via the bus 104).
  • the motors 902 are able to include actuator motors to control the actuation of the module 900 (e.g. movement of a robot arm) and the sensors 904 are able to include image and/or magnetic sensors to input image data and/or detect the position of the module 900 (e.g. a current position of the robot arm, a position of the image sensor, sensed images from the front of a self-driving car, or other sensed data).
  • Figures 10A-C illustrate variants of the control board 906, 906', 906" according to some embodiments.
  • the control board 906 for a multi-connection mode module 900 is able to comprise a node system on chip (SoC) 1002, a transimpedance amplifier (TIA) and/or laser driver (LD) 1004, a bidirectional optical subassembly (BOS A) 1006, a power regulator 1008, a motor driver 1010, a compliant actuator motor and power control connector 1012, a motor control signal transceiver 1014, one or more sensors 1016, a optical splitter 1018, an input power connector 1020, one or more output power connectors 1022, a first fiber optic connector 1024 and one or more second fiber optic connectors 1026 all operatively coupled together.
  • SoC system on chip
  • TIA transimpedance amplifier
  • LD laser driver
  • BOS A bidirectional optical subassembly
  • a power regulator 1008 a motor driver 1010, a compliant
  • the BOS A 1006, splitter 1018 and fiber optic connectors 1024, 1026 are coupled together via fiber optic cable.
  • one or more of the above elements are able to be omitted, their quantity increased or decreased and/or other elements are able to be added.
  • the control board 906 is able to be a flexible printed circuit board.
  • the BOS A 1006 is able to comprise a transmitter optical sub-assembly (TOSA), a receiver optical sub-assembly (ROSA) and a wave division multiplexing (WDM) filter so that it can use bidirectional technology to support two wavelengths on each fiber.
  • the BOSA 1006 is a hybrid silicon photonics BOSA.
  • the motor driver 1010 is able to be a pre-driver, gate driver or other type of driver.
  • the compliant actuator motor and power control connector 1012 is able to transmit control and/or power signals to the motors 902.
  • the motor control signal transceiver 1014 is able to receive motor control signals and/or transmit motor, sensor and/or other data to one or more controller devices 102 via the bus 104.
  • the sensors 1016 are able to comprise magnetic sensors and/or other types of sensors.
  • the sensors 1016 are able to sense a position and/or orientation of the module 900 and provide the positional data as feedback to the SoC 1002 and/or a controller device 102 coupled with the module 900 via the bus 104.
  • the optical splitter 1018 is able to be built-in to the control board 906.
  • the input power connector 1020 receives power for the control board 906.
  • the output power connectors 1022 are configured to supply, transfer and/or forward power to one or more other boards/modules 900.
  • the first fiber optic connector 1024 is coupled with the fiber optic splitter 1018 which splits the cable into two or more cables.
  • One cable couples with the BOSA 1006 for transmitting signals to and from the other elements of the board 906 and the remainder each couple with a different one of the one or more second fiber optic connectors 1026.
  • the first fiber optic connector 1024 and/or second fiber optic connectors 1026 are able to be a pigtail fiber optic connection points and/or connectors 1024.
  • the pigtail fiber optical connection point and/or connector is able to comprise a single, short, usually tight-buffered, optical fiber that has an optical connector pre-installed on one end and a length of exposed fiber at the other end.
  • the end of the pigtail is able to be stripped and fusion spliced to a single fiber of a multi-fiber trunk.
  • other types of optical connection points and/or connectors 1024 are able to be used.
  • the motor driver 1010 is able to receive pulse width modulated (PWM) control signals generated by the SoC 1002 (and/or the controller devices 102 via the SoC 1002) for controlling the torque, speed and/or other operations of the motors 902 of the SCA module 900 (via the compliant actuator motor and power control connector 1012). Additionally, the sensors 1016, the sensors 904 and/or the driver 1010 are able to provide motor and/or sensor status feedback to the SoC 1002 such that the SoC 1002 (and/or the controller devices 102 via the SoC 1002) are able to adjust the control signals based on the feedback in order to control the operation of the motors 902 and/or sensors 904.
  • PWM pulse width modulated
  • the driver 1010 is able to provide motor current sensor feedback comprising phase-A current values, phase-B current values and phase-C current values, wherein an internal analog to digital converter (ADC) of the SoC 1002 converts the values to digital values and the SoC 1002 (and/or the controller devices 102 via the SoC 1002) adjusts the PWM control signals transmitted to the driver 1010 based on the motor current sensor feedback received from the driver 1010 thereby adjusting the speed, torque and/or other characteristics of the motors 902.
  • ADC analog to digital converter
  • the first fiber optic connector 1024 enables the board/module 900 to couple to the bus 104 via an optical fiber cable
  • the splitter 1018 and the second fiber optic connectors 1026 enable the board/module 900 to couple to one or more additional boards/modules 900 via additional optical fiber cable (e.g. for receiving control signals from and/or sending data signals to one or more controller devices 102 coupled to other ports 99 of the bus 104.
  • the boards/modules 900 are able to couple to ports 99 of the bus 104 as a serial cascade wherein only a single port 99 is able to couple to a plurality of boards/modules 900.
  • one board 906 is optically coupled to the port 99 from the first fiber optic connector 1024 (via a fiber optic cable) and each subsequent board 906 has its first fiber optic connector 1024 coupled to the second fiber optic connector 1026 of another one of the boards 906 (all via fiber optic cables).
  • the boards 906 include a plurality of second fiber optic connectors 1026, the cascade is able to branch into a tree structure where single boards/modules 900 are optically coupled to a plurality of other boards/modules 900.
  • the boards/modules 900 are able to share power in the same manner in which they are optically coupled via the input power connector 1020 of one or more of the module 900 receiving power from a power source and one or more of the other modules 900 receiving power by coupling their input power connector 1020 to the output power connector 1022 of another one of the modules 900.
  • the control board 906' for a single-connection mode module 900 is able to not include the one or more second fiber optic connectors 1026 and/or the one or more output power connectors 1022.
  • the control board 906" for a single-connection mode image sensor module 900 is able to further comprise one or more compliant actuator motors 1028 along with one or more image or other types of sensors 1030 (e.g. cameras, LIDAR, magnetic, ultrasound, infrared, radio frequency).
  • the motors 1028 are able to control the movement of the sensors 1030 while the sensors 1016 detect the position and/or orientation of the motors 1028 and/or sensors 1030.
  • the control board 906" is able to be a multi connection mode image sensor module 900 further comprising the one or more second fiber optic connectors 1026 and/or the one or more output power connectors 1022.
  • these single-connection mode modules 900 and/or boards 906' and 906" are able to couple to the cascades or trees formed by the multi-connection mode modules 900 and/or couple in parallel to the bus 104.
  • the system 100 is able to comprise one or more external optical splitters 1102, wherein one or more of the boards/modules 906, 906', 906" configured into serial cascades, trees and/or in parallel are able to be further parallelized and/or serialized in the coupling to the bus 104 using the external optical splitter 1102.
  • an optical splitter 1102 is used to coupled to a single port 99, the output of a cascade of modules 900, one or more individual modules 900 and another splitter 1102.
  • the splitters 1102 are 1 to 4 splitters, they are able to be any ratio 1 to N as desired.
  • the modules 906, 906', 906" are shown as being coupled to the bus 104, it is understood that any combination of other devices 102 are also able to be coupled to the bus 104 along with the modules.
  • one or more controller devices 102 are able to be coupled to the bus 104 for receiving data and issuing commands to the modules.
  • the modules 900 provide the benefit of enabling super high throughput and data bandwidth and can support up to lOx to lOOx of bandwidth and long distance compared to other modules.
  • the ability to utilize optical communication along with serial cascading coupling allows the modules 900 to provide fast data transmission speed and super low latency without being disrupted by electromagnetic interference (EMI).
  • EMI electromagnetic interference
  • modules 900 are particularly advantages in the field of robotics, industrial automation and self-driving vehicles due to its ability to handle their high bandwidth and low latency demands for sensor data.
  • Figure 12 illustrates a method of operating a controller and sensor bus including a plurality of ports for coupling with a plurality of external machine automation devices of a machine automation system according to some embodiments.
  • one or more controller devices 102 are coupled to one or more of the ports 99 of the bus 104 at the step 1202.
  • the first fiber optic connector 1024 of one or more SCA and sensor modules 900 are coupled with one or more of the ports 99 at the step 1204.
  • Messages are relayed between the controllers 104 and the SCA and sensor modules 900 through the bus 104 via the one or more central transmission networks 206 at the step 1206.
  • the control boards 906 adjust operation of the SCA and sensor modules 900 based on the messages received from the controller devices 102 at the step 1208.
  • each of the SCA and sensor modules 900 is directly coupled in parallel to one of the ports 99 via the fiber optic cable.
  • coupling the SCA and sensor modules 900 includes coupling the SCA and sensor modules 900 in parallel to an optical splitter 1102 and coupling the optical splitter 1102 to the ports 99 via the fiber optic cable.
  • coupling the SCA and sensor modules 900 includes coupling the first fiber optic connector 1024 of a first of the SCA and sensor modules 900 to one of the ports 99 via the fiber optic cable and coupling the second fiber optic connector 1026 of the first of the SCA and sensor modules 900 to the first fiber optic connector 1024 of a second of the SCA and sensor modules 900.
  • the system 100 and machine automation controller and sensor bus 104 implementing a dynamic burst to broadcast transmission network has numerous advantages. Specifically, it provides the benefit of a simple cable system and connection; the elimination of significant EMI impacts due to the user of optical fiber cable; guaranteed low latency for node-to-node communication; high throughput bandwidth from node to node transmission (10, 25, 100 or greater Gbps); can extend and reach up to 20km from node to node devices; low power consumption due to passive-optical-network architecture; industry grade QoS without traffic congestion due to centralized DBA scheduling mechanism; built-in HARQ mechanism to guarantee node-to-node and GEM transmission successful; and one unified software image for full intranet system including all gate, node and root ports enabling simplified software architecture, shorter product development cycle, and easier system level debug, monitoring and trouble shooting remotely.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Small-Scale Networks (AREA)

Abstract

A machine automation system for controlling and operating an automated machine. The system includes a controller and sensor bus including a central processing core and a multi-medium transmission intranet for implementing a dynamic burst to broadcast transmission scheme where messages are burst from nodes to the central processing core and broadcast from the central processing core to all of the nodes.

Description

INTELLIGENT CONTROLLER AND SENSOR NETWORK BUS, SYSTEM AND METHOD INCLUDING SMART COMPLIANT ACTUATOR MODULE
RELATED APPLICATIONS
This application is a continuation-in-part of the co-pending United States Patent Application No. 16/572,358, fded September 16, 2019, entitled "INTELLIGENT CONTROLLER AND SENSOR NETWORK BUS, SYSTEM AND METHOD INCLUDING GENERIC ENCAPSULATION MODE," which is a continuation-in-part of United States Patent Application No. 16/529,682, fded August 1, 2019, entitled "INTELLIGENT CONTROLLER AND SENSOR NETWORK BUS, SYSTEM AND METHOD," both of which are hereby incorporated by reference.
FIELD OF THE INVENTION
The present invention relates to the field of buses. More particularly, the present invention relates to a controller and sensor network bus architecture.
BACKGROUND OF THE INVENTION
The field of machine automation is expanding rapidly with the development of self driving cars, intelligent robots and factory automation. However, due to their varied and high-speed needs, there is no bus or network architecture that is able to efficient handle all of the demands of these emerging technologies. Instead, the current networks latency is high, bandwidth is low, cabling is complex, with large electromagnetic interference (EMI), high cost, unsecured data and complex system integration. For example, networks do not have enough speed and throughput to carry sensor data like camera and light detection and ranging (LIDAR) data across the network to CPU Cores. Further, existing cable systems are complex, short-reach, and cannot deal with EMI without expensive shielding due to the use of copper cabling systems. There is no all-in-one “Controller and Sensor Network” system Bus solution that can support and carry internet L2/L3 Ethernet packets, Motor & Motion control messages, sensor data and CPU-CMD across a system from edge node to edge nodes. SUMMARY OF THE INVENTION
A machine automation system for controlling and operating an automated machine. The system includes a controller and sensor bus including a central processing core and a multi-medium transmission intranet for implementing a dynamic burst to broadcast transmission scheme where messages are burst from nodes to the central processing core and broadcast from the central processing core to all of the nodes.
A first aspect is directed to a machine automation system for controlling and operating an automated machine. The system comprises a controller and sensor bus including plurality of ports, wherein the bus comprises a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and one or more subnetworks, the central optical fiber transmission networks including a plurality of nodes and one or more gates and the plurality of subnetworks each coupled to a different one of the gates of one of the central transmission networks, the subnetworks including a plurality of subnodes, wherein each of the nodes and the subnodes are associated with one or more of the ports, one or more controllers each coupled with first ones of the ports of the bus and one or more compliant actuator modules each including a first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC), the modules each operably coupled with second ones of the ports via fiber optic cable coupled from the first fiber optic connector to the second ones of the ports, wherein the nodes and the subnodes of the bus relay messages including sensor data from the sensors of the compliant actuator modules and control data from the controllers between the controllers and the compliant actuator modules through the bus via the one or more central optical fiber transmission networks.
In some embodiments, each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable. In some embodiments, a plurality of the compliant actuator modules are coupled in parallel to an optical splitter and the optical splitter is coupled to one of the second ones of the ports via the fiber optic cable. In some embodiments, each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC. In some embodiments, the compliant actuator modules are coupled in series such that the first fiber optic connector of a first of the compliant actuator modules is coupled to one of the second ones of the ports via the fiber optic cable and the second fiber optic connector of the first of the compliant actuator modules is coupled to the first fiber optic connector of a second of the compliant actuator modules. In some embodiments, each of the compliant actuator modules comprise a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver. In some embodiments, the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, wherein the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor. In some embodiments, the first fiber optic connectors are pigtail fiber optic connection points. In some embodiments, the second ones of the ports are optical ports coupled with pigtail fiber optic connectors. In some embodiments, the automated machine is one of the groups consisting of a robot and a self-driving vehicle.
A second aspect is directed to a compliant actuator module for use with a controller and sensor bus. The compliant actuator module comprises one or more compliant actuator motors, one or more magnetic sensors and a control board including a first fiber optic connector, a second fiber optic connector, a system on chip (SoC) and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC, wherein the control board receives sensor data from the magnetic sensors and outputs status messages based on the sensor data via the first fiber optic connector, controls the motors based on control messages input via the first fiber optic connector having a target wavelength, forwards status messages received via the second fiber optic connector out the first fiber optic connector with the optical splitter and forwards the control messages input via the first fiber optic connector having a non-target wavelength. In some embodiments, the module further comprises an image sensor, wherein the control board receives image data from the image sensor and outputs sensor data messages based on the image data via the first fiber optic connector. In some embodiments, the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor. In some embodiments, the control board comprises a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver. In some embodiments, the first fiber optic connector is a pigtail fiber optic connection point.
A third aspect is directed to a method of operating a controller and sensor bus including a plurality of ports for coupling with a plurality of external machine automation devices of a machine automation system, the bus having a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and including a plurality of nodes and one or more gates and a plurality of subnetworks each coupled to a different one of the gates of one of the central optical fiber transmission networks, the subnetworks including a plurality of subnodes. The method comprises coupling one or more controllers with first ones of the ports of the bus, coupling one or more compliant actuator modules with second ones of the ports via fiber optic cable coupled from a first fiber optic connector to the second ones of the ports, the compliant actuator modules each including the first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC) and relaying messages between the controllers and the compliant actuator modules through the bus via the one or more central optical fiber transmission networks with the nodes and the subnodes of the bus, the messages including sensor data from the sensors of the compliant actuator modules and control data from the controllers.
In some embodiments, each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable. In some embodiments, coupling the compliant actuator modules includes coupling the compliant actuator modules in parallel to an optical splitter and coupling the optical splitter to one of the second ones of the ports via the fiber optic cable. In some embodiments, each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC. In some embodiments, coupling the compliant actuator modules includes coupling the first fiber optic connector of a first of the compliant actuator modules to one of the second ones of the ports via the fiber optic cable and coupling the second fiber optic connector of the first of the compliant actuator modules to the first fiber optic connector of a second of the compliant actuator modules. In some embodiments, each of the compliant actuator modules comprise a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver. In some embodiments, the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, the method further comprises controlling the articulation of the image sensor with the motors and determining the orientation of the image sensor with the magnetic sensors. In some embodiments, the first fiber optic connectors are pigtail fiber optic connection points. In some embodiments, the second ones of the ports are optical ports coupled with pigtail fiber optic connectors. In some embodiments, the automated machine is one of the groups consisting of a robot and a self-driving vehicle. BRIEF DESCRIPTION OF THE DRAWINGS
Figure 1 illustrates a machine automation system according to some embodiments.
Figure 2 illustrates an intelligent controller and sensor intranet bus according to some embodiments.
Figure 3 illustrates a tree topology of an intelligent controller and sensor intranet bus according to some embodiments.
Figure 4 illustrates a block diagram of an exemplary computing device configured to implement the system according to some embodiments.
Figure 5 illustrates a method of operating a machine automation system including an intelligent controller and sensor intranet bus according to some embodiments.
Figure 6 A illustrates an exemplary GEM packet format according to some embodiments.
Figure 6B illustrates a detailed view of a GEM packet header format according to some embodiments.
Figure 6C illustrates a detailed view of a GEM header format for a node report message according to some embodiments.
Figure 6D illustrates a detailed view of a first variant of a GEM header format for a root port bandwidth grant message according to some embodiments.
Figure 6E illustrates a detailed view of a second variant of a GEM header format for a root port bandwidth grant message according to some embodiments.
Figure 6F illustrates a detailed view of a GEM header format for a control message according to some embodiments.
Figure 7 A illustrates a Broadcast-PHY-Frame according to some embodiments.
Figure 7B illustrates a Burst-PHY-Frame according to some embodiments.
Figure 7C illustrates a gate Burst-PHY-Frame according to some embodiments.
Figure 8 illustrates a method of operating the intelligent controller and sensor intranet bus according to some embodiments.
Figure 9 illustrates a smart compliant actuator (SC A) and sensor module according to some embodiments.
Figure 10A illustrates a first variant of a control board of SCA and sensor module according to some embodiments.
Figure 10B illustrates a second variant of a control board of SCA and sensor module according to some embodiments.
Figure IOC illustrates a third variant of a control board of SCA and sensor module according to some embodiments. Figures 11 A and 1 IB illustrate a machine automation system including coupled SCA and sensor modules according to some embodiments.
Figure 12 illustrates a method of operating a controller and sensor bus according to some embodiments.
DETAILED DESCRIPTION OF THE INVENTION
Embodiments described herein are directed to a machine automation system, method and device for controlling and operating an automated machine. The system, method and device including a controller and sensor bus including a central processing core and a multi medium transmission intranet for implementing a dynamic burst to broadcast transmission scheme where messages are burst from nodes to the central processing core and broadcast from the central processing core to all of the nodes. As a result, the system, method and device provides the advantage of high speed performance despite combining lower speed network medium as well as one unified software image for the full intranet system including all gate, node and root ports enabling simplified software architecture, shorter product development cycle, and easier system level debug, monitoring and trouble shooting remotely. In particular, the system, method and device provides a unique intranet system architecture specially defined and optimized for machine automation applications.
Figure 1 illustrates a machine automation system 100 according to some embodiments. As shown in Figure 1, the system 100 comprises one or more external devices 102 operably coupled together with an intelligent controller and sensor intranet bus 104. In some embodiments, the system 100 is able to be a part of an automated device such as a self- driving vehicle, an automated industrial machine or an automated self-controlled robot. Alternatively, the system 100 is able to be a part of other machine automation applications. The devices 102 are able to comprise one or more of sensor devices (e.g. ultrasonic, infrared, camera, light detection and ranging (LIDAR), sound navigation and ranging (SONAR), magnetic, radio detection and ranging (RADAR)), internet devices, motors, actuators, lights, displays (e.g. screens, user interfaces), speakers, a graphics processing units, central processing units, memories (e.g. solid state drives, hard disk drives), controllers/microcontrollers or a combination thereof. Each of the devices 102 is able to be operably wired and/or wirelessly coupled with the bus 104 via one or more bus input/output (IO) ports (see Figure 2). Although as shown in Figure 1, the system 100 comprises a discrete amount of external devices 102 and buses 104, more or less devices 102 and/or buses 104 are contemplated. Figure 2 illustrates the intelligent controller and sensor intranet bus 104 according to some embodiments. As shown in Figure 2, the bus 104 comprises an intranet formed by a central core 200 that is coupled with one or more gates 202 and a plurality of edge nodes 204 (each having one or more external IO ports 99) via one or more central transmission networks 206, and coupled with one or more edge sub-nodes 208 (each having one or more external IO ports 99) via one or more sub-networks 210 that extend from the gates 202. As a result, as shown in Figure 3, the bus 104 forms a network tree topology where the central networks 206 branch from the core 200 (e.g. root ports 230 of the core) to edge nodes 204 and gates 202, and the subnetworks 210 branch from the gates 202 to sub-nodes 208 and/or sub-gates 202'.
In this way, the core 200 is able to see all of the nodes 204 and sub-nodes 208 (as the gates 202 and sub-gates 202' are transparent to the core 200). In some embodiments, one or more of the gates 202 are directly coupled with IO ports 99 without a node (e.g. to couple with external CPU, GPU, AI cores and/or solid state drives (SSD)).
The ports 99 are able to be any kind of interface port such as peripheral component interconnect express (PCIe), mobile industry processor interface (MIPI), Ethernet, universal serial bus (USB), general purpose input output (GPIO), universal asynchronous receiver/transmitter (UART), inter-integrated circuit (I2C) and/or other types of ports. Although as shown in Figure 2, the bus 104 comprises a discrete amount of ports 99, cores 200, nodes 204, 208, gates 202, networks 206, 210, other elements and components thereof, more or less ports 99, cores 200, nodes 204, 208, gates 202, networks 206, 210, other elements and/or components there of are contemplated.
The central transmission networks 206 are able to comprise connection media that is faster/lower latency than the connection media of the subnetworks 210 coupled to a gate 202 of that central transmission network 206. Similarly, the subnetworks 210 are able to comprise connection media that is faster/lower latency than the connection media of the subnetworks 210' coupled to a gate 202' of the subnetwork 210 and so on for each iterative subnetwork. This network/subnetwork connection media speed/latency relationship enables the bus 104 to prevent the slowing of the processing of the entire bus 104 despite still including the slower connection media as describe in detail below. Alternatively, one or more of the subnetworks 210, 210' and/or the central networks 206 are able to have the same or other connection media speed/latency relationships.
In some embodiments, the connection media of the central transmission networks 206 comprises optical fiber cables 212 split using optical splitters 214 (e.g 2-to-l splitters) and having optical transceivers 216 to couple to and received data from the nodes 204, 208. In some embodiments, the connection media of the subnetworks 210 comprises optical connection media (e.g. like the central transmission networks 206, but possibly slower rating), wireless connections (e.g. radio frequency transceivers 218), copper connections (e.g. twisted-pair copper wires 220 optionally split using analog splitters 222( e.g. fan outs/multiplexers) and having serializer/deserializers (SERDES) 224 to couple to and received data from the nodes 204, 208), and/or combinations thereof (e.g. hybrid optical fiber, copper and/or wireless connection media). As a result, the bus 104 supports multi-rate traffic transmissions where depending on the latency/speed, connectivity and/or distance requirements of the data/traffic/extemal devices 102, different nodes/networks are able to be used to coupled to the bus 104 while still providing the desired throughput. For example, for high speed, low latency and long-distance requirements the optical connection media of the central network is able to be used by coupling to the nodes 204. Otherwise, the other networks 210 are able to be used depending on cost, speed, connection and/or distance requirements. In some embodiments, the central networks 206 are passive optical networks and/or the copper subnetworks 210 are active networks. In some embodiments as shown in Figure 2, one or more of the nodes 204 is coupled to a controller area network (CAN) 226 such that the node inputs data from each of the controllers coupled to the controller are network. Alternatively, as shown in Figure 3, one or more of the subnetworks 210 are able to be a CAN coupled with the core 200 via one of the gates 202.
Multi-layer Bus Addressing
The bus 104 is able to utilize a multi-layered addressing scheme where the root ports 230, IO ports 99, nodes 204, 208, 234 and/or gates 202 are able to use node, epoch and GEM identifying addresses for directing messages through the bus 104. In particular, each of the root ports 230, nodes 204, 208, 234 and gates 202 are able to be assigned a node identifier (node-ID), with the nodes 204, 208 and gates 202 also being assigned at least one epoch identifier (epoch-ID) and at least one GEM identifier (GEM-ID). The epoch-IDs are able to be used to identify the source/destination of messages in the network 206, 210 (e.g. node/gate devices and their IO ports, embedded CPUs and/or other types of services) while at the same time the GEM-IDs are able to be used to identify the targets of messages (e.g. sets and subsets of the node/gate devices and their IO ports, embedded CPUs and/or other types of services). As a result, the epoch-IDs are able to be used for the transmission/routing of messages throughout the network 206, 210 while the GEM-IDs are able to be used by the devices themselves (via the ports 99) to determine whether to capture received/broadcast messages as being targeted to them. Depending on the service level agreement (SLA) profile of the node/gate (which is able to correspond to the devices coupled to the port(s) 99 of the node/gate), the nodes/gates are able to be assigned multiple epoch-IDs and multiple GEM-IDs. As a result, the node-ID of each of the nodes 204, 208 and gates 202 is able to map to one or a plurality of epoch-IDs which are able to map to one or a plurality of GEM-IDs. For example, a node 204, 208 coupled with two IO ports 99 is able to have a single node-ID, two epoch-IDs (one for each port 99) and ten GEM-IDs (one associated with the first epoch-ID and first port 99 and nine associated with the second epoch-ID and second port 99). Further, although the node-IDs and epoch-IDs are unique to each node/gate/port, the GEM-IDs are able to be shared between nodes/gates/ports. For example, ports 99 of the same node 204, 208 or different ports 99 of different nodes 204, 208 are able to both be associated with matching or overlapping sets GEM-IDs.
The gates 202 are also able to be assigned one or more virtual node-IDs for the ports 99 directly coupled with the gate 202. Like the regular nodes, these virtual nodes represented by the gates 202 are able to be assigned multiple epoch-IDs and multiple GEM-IDs depending on the SLA profde of the gate 202 (which is able to correspond to the devices coupled to the port(s) 99 of the virtual node/gate).
The other nodes 234 and cores 232 (that are directly coupled to the core 200 such as IO devices and embedded CPU cores) are each able to have one or more GEM-IDs along with a global node-ID, but do not need to be assigned epoch-IDs, which are not required because messages to and from these nodes 234 to the core 200 are wholly within the core 200. Like nodes 204, 208, the number of GEM-IDs assigned to each of the nodes 234 and cores 232 is able to be determined based on the SLA profile for that node 234 or core 232 (which is able to correspond to the devices coupled to the port(s) 99 of the node 234). Each of the core switch 220, root ports 230, nodes 204, 208, 234, and/or gates 202 are able to maintain and update a local SLA table that indicates the mapping between each of the node-IDs, epoch-IDs and GEM-IDs. As a result, the bus addressing provides the advantage of using epoch-IDs and/or node-IDs to facilitate simplified burst/broadcast messaging between nodes, gates and the core within the network 100, while at the same time using GEM-IDs facilitate any desired more complex messaging between the devices/IO ports 99 and/or the core themselves.
Generic Encapsulation Mode
The bus 104 is able to encapsulate all input data and internally generated data (e.g. control, operation and management messages) into a generic encapsulation mode (GEM) for transport across the bus 104 intranet. Thus, the GEM acts as a unique standardized data and message container for transmitting data between nodes and/or to the core 200 via the bus 104 intranet. As a result, the input data is able to be encapsulated into the GEM format at each of the nodes as it enters the bus 104 and is routed through the core 200 (where it is decapsulated for processing and re-encapsulated for transmission) and onto its destination node which decapsulates the data back to the original format for egress to the target external device 102 or other destination. This input data is able to be from various sources (e.g. devices 102,
CAN 226) input via the ports 99 at the nodes 204, 208, 234 or gates 202 and/or the embedded CPU cores 232.
There are two types of GEM formats: GEM packet and GEM control. The GEM packet format comprises a GEM header plus a GEM payload (e.g. length from 8 bytes to 4 kilobytes). Typically, the GEM packet format what is used to encapsulate the input port data, packets and messages at the ingress (e.g. nodes, ports). The following are some of the IO port data, packet and message examples that are able utilize the GEM packet format:
Use GEM packet format to carry Ethernet packets from local gate 202 and/or node 204, 208 through bus 104 after GEM encapsulation to far-end gate 202 and/or node 204 (e.g. this is able to be for internet and Wi-Fi interfaces through Ethernet Port or PCIe Ports);
Use GEM packet format to carry sensor data from local gate 202 and/or node 204, transmit through bus 104 after GEM encapsulation to far-end gate 202 and/or node 204 (e.g. CAN bus data, Camera (MIPI) Frame data,
Lidar (Ethernet) data, Magnetic Encoder data (ADC) and other type of Sensors data;
Use GEM packet format to carry jumbo size data and packets and transmit through fragmentation and de-fragmentation scheme, from local node 204, 208 to far-end node 204, 208. This is able to include fragmentation, defragmentation and re-ordering / re-transmission functions;
Use GEM packet format to carry the network control, operation and management messages between core 200 and nodes 204, 208 (and/or gates), including physical layer operation, administration and maintenance (PLOAM), node management control interface (NMCI) and operations, administration and maintenance (OAM) messages;
Use GEM packet format to carry CPU/PCIe access CMD/DATA from core 200 and local gate 202 and/or node 204 through bus 104 after GEM encapsulation, to far-end local gate 202 and/or node 204 (e.g. CPU 232 access target device 102 from NODE-to-NODE through PCIe, USB, I2C, UART and GPIO interfaces).
Finally, use GEM packet format for VPN channel application between local-nodes 204, 208 to far nodes 204, 208 through bus 104.
The GEM control message format comprises message plus extended message (e.g. length 8 bytes + 8 bytes ...). The GEM control message format is able to be used in the bus 104 for internal network management and control purposes, including messages of dynamic bandwidth allocation (DBA) reporting, DBA-Granting, GEM RX- Acknowledge, GEM Flow-Control, GEM Power-Management, GEM-Sniffer, GEM-Remote messages and/or other types of control messages. As described above, nodes 204 are responsible for encapsulating/decapsulating data to/from GEM packet and GEM control message format.
This scheme is able to expand PCIe interface protocol from point-to-point topology to point-to-multi-point topology and extend the interface distance from short reach to long reach.
Figures 6A-F illustrate an exemplary GEM packet format and GEM header formats according to some embodiments. As shown in Figure 6A, an GEM packet 600 is able to comprise a header 602 and a corresponding payload 604. As described above, for message packets the header is able to be a set size (e.g. 8 bytes) and the payload is able to vary in length (e.g. length from 8 bytes to 4 kilobytes) and for control packet the header is able to be, for example, 8 bytes with or without one or more 8 byte extensions.
Figure 6B illustrates a detailed view of a GEM packet header format according to some embodiments. As shown in Figure 6B, the header 602 comprises a GEM type field 606, a payload length indication field 608, an encryption key index field 610 (e.g. AES Key Index), a node/epoch ID field 612, a GEM-ID field 614, a GEM packet type field 616, a transmission sequence identifier field 618, an acknowledgment required field 620, a last fragment indication field 622 and a header error correction/check (HEC) field 622. Alternatively, one or more of the fields are able to be omitted and/or one or more additional fields are able to be added. In some embodiments, the GEM type field 606 is two bits, the payload length indication field 608 is twelve bits, the encryption key index field 610 is two bits, the node/epoch ID field 612 is twelve bits, the GEM-ID field 614 is twelve bits, the GEM packet type field 616 is three bits, the transmission sequence identifier field 618 is six bits, the acknowledgment required field 620 is one bit, the last fragment indication field 622 is one bit and the header error correction/check (HEC) field 622 is thirteen bits.
Alternatively, one or more of the fields are able to be larger or smaller. The GEM type field 606 indicates which type of header 602 (and thus which type of packet) the GEM packet 600 is. For example, the GEM type field is able to indicate that the header 602 is one or more of a packet header, a bandwidth grant message header (e.g. transmitted from a root port 230 to a gate/node), a bandwidth report message header (e.g. transmitted from a gate/node to a root port 230) and/or a control message (e.g. between one or more of the root ports 230, the gates 202 and/or the nodes 204, 208, 234). The payload length indication field 608 indicates the length of the payload 604 of the packet 600. The encryption key index field 610 indicates the type of encryption to use on the packet 600. For example, the encryption key index field 610 is able to be used as an index value within an encryption table to identify one or more of: whether to encrypt the packet or not, which key to use to encrypt the packet, and/or which method of encryption to use.
The node/epoch ID field 612 is able to identify either the source node or the destination node of the packet 600. For example, for a GEM packet 600 being burst from a node to the core, the field 612 is able to be or represent the node’s epoch-ID to indicate the source of the packet 600. As another example, for a GEM packet 600 being broadcast from a root port 230 to the nodes/gates within its network 206, 210, the field 612 is able to be or represent the destination’s node -ID (including a unicast node-ID, a multicast node -ID and/or a broadcast node-ID). The GEM-ID field 614 is able to be or represent the source node’s data/packet/message identifier for a point to point message, or is able to be or represent the destination node’s GEM-ID (e.g. including CAN message GEM-IDs, sensor data GEM-IDs and/or ethemet packet GEM-IDs) for point to multi-point messages. As a result, the GEM format provides the advantage of enabling the bus 104 to identify both the immediate source and/or destination nodes via the node/epoch ID field 612 while also enabling the target devices/port/services to be identified using the GEM-ID field 614.
The GEM packet type field 616 is able to indicate the type and format of the header of the message encapsulated within the GEM format (e.g. as received from the devices 102 and/or through the ports 99). For example, the field 616 is able to indicate that the message header is a PLOAM message, a node management and control interface (NMCI) message, a CAN command message, sensor data, an ethemet packet, CPU-IO (e.g. PCIe/USB) message and/or a node operation and control report (NOCR) message. The acknowledgment required field 620 is able to indicate if an acknowledgment message in response to the message is require and the transmission sequence identifier field 618 is able to identify the transmission sequence number of the packet 600 within a set of packets from the source node and/or an epoch-ID thereof (for a packet being burst from the node to the core 200). In some embodiments, it requires an acknowledgment message from the receiving root port 230 when indicated by the acknowledgment required field 620. For a packet broadcast from the root port 230 to a node/gate, the transmission sequence identifier field 618 is able to identify the transmission sequence number of the unicast/broadcast/multi-cast GEM-ID (e.g. CAN Message GEM-ID, sensor Data GEM-ID, Ethernet Packet GEM-ID and CPU/PCIe/USB Data-Message GEM-ID). In some embodiments, it requires acknowledge from receiving root port 230 and/or node when indicated by the acknowledgment required field 620. The last fragment indication field 622 is able to indicate if this packet 600 is the last fragment of a series of fragments of a large packet and the header error correction/check (HEC) field 622 is able to be used to check the header 602 for errors.
Figure 6C illustrates a detailed view of a GEM header format for a node report message according to some embodiments. As shown in Figure 6C, the header 602 comprises a GEM type field 606, a report message type field 624, a source epoch-ID field 626, a report total size field 628, a report threshold size field 630, a report sequence number field 632, one or more source node virtual output queue (VOQ) status fields 634 (e.g. CPU-IO, PLOAM, NMCI, CAN, Sensor, Ethernet, or other types), a report priority field 636 and a header error correction/check (HEC) field 622. Alternatively, one or more of the fields are able to be omitted and/or one or more additional fields are able to be added. In some embodiments, the GEM type field 606 is two bits, the report message type field 624 is two bits, the source epoch-ID field 626 is twelve bits, the report total size field 628 is fourteen bits, the report threshold size field 630 is eight bits, the report sequence number field 632 is five bits, the one or more source node virtual output queue status fields 634 are each one bit (or a single field of six bits), the report priority field 636 is two bits and the header error correction/check (HEC) field 622 is thirteen bits. Alternatively, one or more of the fields are able to be larger or smaller.
The report message type field 624 indicates which type of report header 602 (and thus which type of report message) the GEM packet 600 is. For example, the report message type field 624 is able to indicate that the header 602 is one or more of an invalid report message, a node report message for itself (e.g. where the epoch-ID of the source of the packet is mapped to the node-ID of the source of the packet), a node report message for another node (e.g. where the epoch-ID of the source of the packet is not mapped to the node-ID of the source of the packet), and/or a dying gasp report message (e.g. a message that needs/requests top priority). The source epoch-ID field 626 is able to be or represent: the source node’s epoch- ID (e.g. for a report for PLOAM and NMCI plus CAN/sensor/ethernet queue flags), the CAN’s epoch-ID (e.g. for a report for the CAN), the epoch-ID of one of the sensors/nodes (e.g. for a report for the sensor), the ethemet epoch-ID (e.g. for a report for ethemet packets) and/or a PCIe/USB epoch-ID (e.g. for a PCIe/USB report message). The report total size field 628 is able to indicate the total size of the GEM data within the VOQ (for that epoch-ID and/or Node -ID), whereas the report threshold size field 630 is able to indicate the GEM packet boundary(ies) within the VOQ (e.g. for use when determining the size of burst windows granted for the epoch and/or node).
The report sequence number field 632 is able to indicate which number in the sequence that the message is (e.g. if there are a sequence of related report messages in order to determine if one is lost or mis-sequenced). The one or more source node virtual output queuing (VOQ) status fields 634 are each able to indicate a status of the source node with respect to a particular function/type of data (e.g. CPU/IO, PLOAM, NMCI, CAN, sensor, ethernet). The report priority field 636 is able to indicate what priority to give the message (e.g. best efforts, normal bandwidth request priority, CAN message request priority, dying gasp request priority).
Figures 6D and E illustrate a detailed view of two variants of a GEM header format for a root port bandwidth grant message according to some embodiments. As shown in Figure 6D, for a node grant message where the node-ID is the same as the epoch-ID, the header 602 is able to comprise a GEM type field 606, an epoch-ID field 638, a start time field 640, a grant size field 642, a grant flag field 644, a report command field 646, a grant command field 648, a force wake-up indicator (FWI) field 650, a burst profile field 652 and a header error correction/check (HEC) field 622. Alternatively, one or more of the fields are able to be omitted and/or one or more additional fields are able to be added. In some embodiments, the GEM type field 606 is two bits, the epoch-ID field 638 is twelve bits, the start time field 640 is fourteen bits, the grant size field 642 is fourteen bits, the grant flag field 644 is one bit, the report command field 646 is three bits, the grant command field 648 is two bits, the force wake-up indicator field 650 is one bit, the burst profile field 652 is two bits and the header error correction/check (HEC) field 622 is thirteen bits. Alternatively, one or more of the fields are able to be larger or smaller.
The epoch-ID field 638 is able to be or represent the epoch-ID of the node or node-ID that the message is for. The start time field 640 is able to indicate a starting time of the grant window that is being granted to the target node (e.g. epoch of that node) and the grant size field 642 is able to indicate the size/duration of the grant window. The grant flag field 644 is able to indicate whether the window was granted. The report command field 646 is able to indicate what reporting is requested from the node/epoch/port. For example, the report command field 646 is able to indicate one or more of: no node request to send (RTS) status report or force node to report RTS message to port for blackbox and diagnostic test; combined with one or more of: PLOAM and NMCI reporting only forced reporting of CPU- IO messages, CAN messages and sensor data plus PLOAM/NMCI; forced reporting for ethernet packets plus CPU-IO/CAN/sensor and PLOAM/NMCI; and/or forced full report of PLOAM/NMCI/CPU-IO/CAN/sensor/ethernet plus a node operation and control report (NOCR). The grant command field 648 is able to indicate what type of messages/data are granted the burst window. For example, the grant command field 648 is able to indicate one or more of: the window is not for PLOAM and NMCI messages; the grant window is only for PLOAM messages; the grant window is only for NMCI messages; and/or the grant is for PLOAM, NMCI and NOCR messages. The FWI field 650 is to indicate whether to force a sleeping node to wake-up and the burst profde field 652 is able to indicate a burst configuration (e.g. length, pattern and/or other characteristics of the SOB delimiter, EOB delimiter and/or preamble).
As shown in Figure 6E, for a GEM grant message where the node-ID is not the same as the epoch-ID, the header 602 is able to be substantially the same as the header of Figure 6D except without the report command field 646 and the FWI field 650. Further, unlike in Figure 6D, the grant command field 648 is able to be six bits. Alternatively, the grant command field 648 is able to be larger or smaller. Also unlike in Figure 6D, the grant command field 648 is able to indicate a GEM bandwidth grant of different types. For example, the field 648 is able to indicate a bandwidth grant for: all VOQ/CoS (class of service) based on the nodes ’s output scheduling settings, for CAN messages only, for sensor data only, dying gasp messages only and/or for both CAN messages and sensor data. Additionally, the field 648 is able to force power saving for the node-ID where the node replies with an acknowledge message.
Figure 6F illustrates a detailed view of a GEM header format for a control message according to some embodiments. As shown in Figure 6F, the header 602 comprises a GEM type field 606, a control message type field 654, one or more control message fields 656 and a header error correction/check (HEC) field 622. Alternatively, one or more of the fields are able to be omitted and/or one or more additional fields are able to be added. In some embodiments, the GEM type field 606 is two bits, the control message type field 654 is four bits, the one or more control message fields together are forty-five bits and the header error correction/check (HEC) field 622 is thirteen bits. Alternatively, one or more of the fields are able to be larger or smaller.
The control message type field 654 is able to indicate what type of control message the message is (e.g. so the control message fields 656 and their offsets are known for processing). In some embodiments, the control message type field 654 indicates one or more of: a report acknowledgment message; a CAN acknowledgment message; a flow control message; a power saving message; and IO event message (e.g. dying gasp); a run-time status message; and/or a timestamp update (e.g. from port to node). The control message fields 656 are able to include various control message fields based on the type of control message (as indicated in control message type field 654).
Accordingly, the GEM format provides the benefit of enabling the bus 104 to encapsulate varying input data and messages of significantly different types of networks (e.g. controller area networks, optical networks, sensor device broadcasting networks, wireless networks, CPU access networks) to one unique format (GEM). This unique format is then able to facilitate high speed standardized processing and transmission of the varied data input in both burst and broadcast messages thereby enabling the efficient operation of the multi network multi-device bus architecture required for modem machine automation applications.
Burst/Broadcast Frame Format
In some embodiments, the broadcast messages are formatted into a Broadcast-PHY-Frame defined by: Preamble + Start-of-Frame-Delimiter + Frame -Payload, wherein the frame payload includes multiple GEM-Packet data and GEM-Control messages. The Broadcast-PHY-Frame is able be a fixed frame size (e.g. between 25-125ps). Alternatively, greater or smaller frame sizes are able to be used. For example, for central networks 206 and subnetworks 210 with less node devices 204, 208, the frame size is able to be smaller (e.g. 25 ps or 50ps). In some embodiments, the Broadcast-PHY-Frame is constructed to carry GEM-Packet and GEM-Control messages from the root ports 230 to the gate 202 and/or nodes 204, 208, 234 through the networks 206, 210 including optical, copper and wireless networks.
In some embodiments, the burst messages are formatted into a Burst-PHY-Frame defined by: Preamble + Start-of-Frame-Delimiter + Frame Payload + End-of-Frame-Delimiter, wherein the frame payload includes one or more GEM-Packet data and GEM-Control messages. The Burst-PHY-Frame size is able to vary depending on the total Burst-Window size of node/gate granted by root port HDBA and/or gate DBA. In some embodiments, the max size of Burst-PHY-Frame (from a gate 202 or a node 204, 208, 234) cannot exceed the max Broadcast-PHY-Frame size (e.g. between 25-125ps). In some embodiments, the Burst-PHY-Frame is constructed to carry GEM-Packet and GEM-Control messages from gates 202 and/or nodes 204, 208, 234 to the root ports 230 and/or gates 202 via the networks 206, 210 including optical, copper and wireless networks. Figure 7A illustrates a Broadcast-PHY-Frame 700 according to some embodiments. As shown in Figure 7A, the Broadcast-PHY-Frame 700 comprises a physical synchronization block for broadcast (PSBbc) 702 and a broadcast framing sublayer frame 704 including a GEM control message 706, one or more GEM packets 600 and a framing sublayer (FS) trailer 708. Each of the GEM packets 600 include a header 602 and a payload 604 as described above. In some embodiments, the broadcast FS frame is FEC protected. Figure 7B illustrates a Burst-PHY-Frame 710 according to some embodiments. As shown in Figure 7B, the Burst- PHY-Frame 710 comprises a physical synchronization block unicast start of burst delimiter (PSBuc_sd) 712, a burst framing sublayer (FS) 714 and a physical synchronization block unicast end of burst delimiter (PSBuc ed) 716. The PSBuc sd 712 is able to include a preamble 718 and a start of burst (SOB) delimiter 720 and the PSBuc ed 716 is able to include an end of burst (EOB) delimiter 722. The burst FS 714 is able to include a FS header 724, one or more epochs 726 and an FS trailer 708. Each of the epochs 726 are able to include one or more GEM packets 600 having a header 602 and a payload 604 as described above. In some embodiments, the burst FS frame is FEC protected. In particular, by including an EOB delimiter (in addition to the SOB delimiter and a size of the frame), the structure 710 enables a sniffer, analytics engine or other element to monitor the traffic within the bus 104 because it enables the element to determine the end of each burst frame based on the EOB delimiter despite not knowing/accessing the size of the frame.
Figure 7C illustrates a gate Burst-PHY-Frame 728 according to some embodiments. As shown in Figure 7C, the gate Burst-PHY-Frame 728 is able to comprise one or more Burst-PHY-Frames 710 combined together into a single combined burst-PHY-frame having a single preamble 729 and one or more gaps 730. In particular, as described in detail below, the gates 202 are able to receive burst frames 728 from one or more subnodes 208 as well as one or more IO ports 99 (for which they serve as a virtual node) and combine those frames 728 into a combined gate Burst-PHY-Frame 728 as shown in Figure 7C. As a result, the system 100 provides the advantage of more efficient message communication via combined burst frames as well as less overhead per frame by using only a single preamble for the combined frame as a whole instead of a separate preamble for each combined burst frame (whose preamble can be up to 256 bytes each or more).
Figure 8 illustrates a method of operating the intelligent controller and sensor intranet bus 103 according to some embodiments. As shown in Figure 8, one or more of the nodes 204, 208 input one or more messages from the one or more of the devices 102 coupled to the one or more of the ports 99 at the step 802. The nodes 204, 208 encapsulate the messages into the generic encapsulation mode (GEM) format for transmission to the central processing core 200 at the step 804. If the destination(s) of the input messages is a node 234 inside the core 200, the core decapsulates, processes and transmits the messages to their destination(s) without re-encapsulation at the step 806. Otherwise, if the destination(s) of the input messages is one or more other nodes 204, 208 (outside the core 200), the core 200 decapsulates, processes and re-encapsulates the messages back into the GEM format for broadcast to their destination(s) at the step 808. The nodes 204, 208 decapsulate the messages as received from the core 200 from the GEM format to an original format of the input data as received from one of the devices 102 at the step 810. Alternatively, if the input messages are input from nodes 234 inside the core 200 they are able to be input and processed by the core 200 (without being encapsulated) and only encapsulated by the core 200 for broadcast if their destination is one or more nodes 204, 208 outside the core 200. As a result, the method provides the advantage of enabling the communication of many different types of data (e.g. sensor, controller bus, ethernet, or other types of data), more efficient message communication via combined burst frames, and less overhead per frame by using only a single preamble for the combined frame as a whole instead of a separate preamble for each combined burst frame.
Core
The core 200 is able to comprise a core switch 228, one or more root ports 230 (internal ports), a central processing unit 232 and one or more core nodes 234 having IO ports 99 (external ports). In some embodiments, the core 200 further comprises a secure memory (e.g. secure digital (SD) memory) node 236 for storing data in a black box memory 238. Alternatively, the SD node 236 and/or memory 238 are able to be omitted. The core nodes 234 enable a user to couple a user plug-in module (e.g. CPU core, WIFI LTE/5G, User Application software) directly to the core 200 bypassing the networks 206, 210.
The core switch 228 comprises a forwarding engine element, a queuing buffer manager and a traffic manager. Forwarding engine element is able to comprise a plurality of forwarding engines. For example, it is able to include one engine used for L2/L3/L4 Ethernet header parser, lookup and classification/access control list (ACL) function, including L2 medium access control (MAC) Address learning and forwarding functions, L3 internet protocol (IP) Address to GEM -ID Routing/mapping. Additional, one engine is able to be used for GEM Header message parser, lookup, ACL and forwarding and/or another is able to be used to support DOS attack functions to protect the bus 104 from external internet DOS attack. The GEM-Queuing-Buffer Manager is able to be a centralized buffering architecture, which employs link-list based buffer and queuing memory methods combining store-N-forward and cut-through forwarding schemes. For latency sensitive GEM-Packet and GEM-Messages, it is able to use a cut-through forwarding scheme and for congestion GEM-Packets it is able to use store-N-forward scheme. Both schemes are able to be dynamically mixed together and dynamically switched between each other depending on the run-time traffic congestion situations. The GEM-Traffic Manager supports GEM-ID and NODE-ID base dual-token policing, single-token rate-limiting and output shaping functions, including related management information base (MIB) counters. GEM-ID base weighted random early detection (WRED) and Tail-Drop functions are able to be supported as well as early traffic congestion detection and indication and feedback mechanisms to notify hybrid dynamic bandwidth allocation mechanisms (HDBA), root ports 230, gates 202 and nodes 204, 208, 234 to slow down traffic transmission in order to avoid traffic congestion from occurring.
As a result, the core switch 228 is able to provide the functions of on ingress, the switch 228 receives GEMs from one or more of the root ports 230, local nodes 234, computer 232 and/or other IO ports, processes the GEMs and on egress, forwards and transmits the received GEMs to one or more of the root ports 230, local nodes 234, computer 232 and/or other IO ports. In other words, the switch 228 is able to accept GEM-Packets from multiple sources; perform GEM and Ethernet L2/L3/L4 header parsing, L2 MAC lookup and learning, GEM message and 5 -tuple ACL and classification; modify GEM-Header and GEM payload Ethernet header (if necessary); and store and forward GEM-Packet (or cut-through buffer memory) to one or multiple hybrid automatic repeat request (HARQ) functional blocks and the broadcast-MAC of one or more root ports 230.
In performing this processing and/or forwarding function, the switch 228 is able to support hybrid store-and forward and cut-through forwarding schemes in order to reduce propagation latency for latency sensitive GEMs and provide big enough buffering for over burst GEM traffic. Additionally, the switch 228 is able to support instant- flow-control mechanisms within the bus 104, including hybrid dynamic bandwidth allocation and granting to ensure overall quality of service (QoS) across the bus 104. Further, the switch 228 is able to support L2/L3/L4 ACL and classification, L2 MAC address learning and forwarding, L3 IP address to GEM-ID routing/mapping, as well as DOS attack protection. Finally, the switch 228 is able to support QoS scheduling, GEM buffering WRED/Tail dropping, node and/or GEM policing and output shaping functions. Root Ports
The root ports 230 are able to comprise a root transmission MAC, a root reception MAC, a security engine (e.g. advanced encryption standard (AES)), a forward error correction (FEC) engine, a hybrid dynamic bandwidth allocation (HDBA) engine, an activation processor (e.g. activation state machine) and a burst-mode SERDES IP. Alternatively, one or more of the above elements are able to be omitted. The transmission MAC of each of the root ports 230 is responsible for accepting GEMs ready for egress from switch 228 and/or HARQ; map and pack the GEMs into a broadcast frame format (e.g. Broadcast PHY-Frame structure); and broadcast the GEMs to all of the gates 202 and/or nodes 204 on the central transmission network 206 to which the root port 230 is coupled (e.g. through root SERDES and optical/copper network broadcast domains). Conversely, the reception MAC of each of the root ports 230 is responsible for receiving GEMs in a burst frame format (e.g. Burst-PHY-Frame structure) from Burst-Mode SERDES and gates 202 and/or nodes 204,
208; extracting the GEMs from burst frame format; parsing the GEM-header of the GEMs; and accepting the GEMs addressed to it (e.g. based on the GEM -Header and system service level agreement (SLA) profile settings), then outputting the GEMs/data to the switch 228 for further processing and forwarding. In other words, the root ports 230 are each able to receive burst traffic from the nodes 204 and/or gates 202 (forwarded from nodes 208 in the subnetwork 210 of the gate 202), convert the burst traffic to the correct format for processing by the switch 228 and then reformat and broadcast output traffic to all of the nodes 204 and nodes 208 (via the gates 202) to destinations as directed by the switch 228.
The hybrid dynamic bandwidth allocation (HDBA) engine is responsible for receiving reports about bandwidth usage, traffic congestion and other factors (e.g. NODE-DBA Reports); performing HDBA analysis based on an SLA profile for the node/port/device associated with each report, the DBA-Report data itself and committed information rate (CIRypeak information rate (PIR) feedback; and granting burst windows to each NODE device and assigned port/EPOCH-ID. In other words, the HDBA engine inputs data from each of the nodes 204, 208 (of the network 206 associated with the root port 230 and subnetworks 210 thereof) and/or other sources about bandwidth usage/traffic congestion and dynamically allocates burst transmission window start times and/or sizes to each of those nodes 204, 208. In performing this allocation for the nodes 208 within the subnetworks 210, the gate 202 that provides access to the nodes 208 is transparent to the HDBA engine. As a result, as described in detail below, the gate 202 receives the desired data and performs the burst transmission within the assigned windows for each of the nodes 208 of the gate’s 202 subnetwork 210. The HDBA engine is also able issue reporting acknowledgment messages (GEM -Report- ACK message) to nodes 204, 208 to confirm that the report messages (GEM -DBA Reports) were received.
The root Activation State-Machine is responsible for performing and completing node 204, 208, 234 device activation and registration through activation processes and procedures by exchanging physical layer operations, administration and maintenance (PLOAM) GEM messages between nodes 204, 208, 234 and the root port 230. The security engine is able to be an AES-128/256 encryption and decryption functional block used for both the reception and transmission MACs. Alternatively, other encryption is able to be used. The forward error correction (FEC) engine is used for controlling errors in data transmission over unreliable or noisy communication channels. In some embodiments, the FEC engine uses Reed Solomon FEC coding schemes of RS(255,216) and RS(225,232) for 10G and 2.5G data rates, respectively. Alternatively, the FEC engine is able to user low-density parity-check (LDPC) schemes and/or other FEC algorithms. The burst-mode SERDES uses fast clock and data recovery (CDR) locking mode to ensure proper burst messages (e.g. burst-PHY-Frames) are received correctly. In some embodiments, the fast locking function of CDR is required in fiber-cut, fast fail-over and protection switch recovery.
Finally, after a registration process, the root ports 230 receive broadcast data distribution service (DDS) messages from nodes 204, 208 that notify the root port 230 that new nodes/devices have joined and registered to bus 104. Accordingly, the root ports 230 are configured to always listen and accept these data distribution service (DDS) messages from the switch 228 and new node’s 204, 208 declaration of joining the bus 104, and update the Root-Port SLA profde database and settings to reflect the newly added nodes/devices.
Nodes
The edge nodes 204, 208, 234 provide a bridge function within the bus 104 to interface with external devices 102 via the IO ports 99 on one side and connect to bus intranet 104 on the other side. In order to provide data from the devices 102 coupled to the ports 99 of the nodes 204, 28, the nodes 204, 208, 234 construct and transmit burst messages (e.g. Burst-PHY-Frames of the data encapsulated as GEMs) through the bus 104 to the other nodes 204, 208 via the root port 230 (of the network 206 of which they are a part or a subnetwork 210 thereof). Further, in order to provide data to the devices 102 coupled to the ports 99 of the nodes 204, 28, the nodes 204, 208, 234 receive broadcast message (e.g. Broadcast-PHY-Frames of the data encapsulated as GEMs) from other nodes 204, 208 via the root port 230 (of the network 206 of which they are a part or a subnetwork 210 thereof), extract the data from the broadcast messages (e.g. GEMs from RX BC-PHY-Frames), and filter and accept the data that belongs (is addressed to) the node 204, 208.
To perform these and other functions, the edge nodes 204, 208 are able to comprise one or more IO ports 99, an encapsulation/decapsulation engine, a HARQ block and a node MAC. Each of the ports 99 is able to be one of a CPU interface (e.g. PCIe, USB and UART), a sensor interface (e.g. MIPI, analog to digital converter (ADC), GPIO), an internet interface (e.g. Ethernet, EtherCAT, and CAN-Bus), and a motor module interface (e.g. pulse width modulation (PWM), I2C, ADC and GPIO). The encapsulation/decapsulation engine accepts input data from the ports 99 and encapsulates received data packets, commands (CMD) and messages received from the internet ports (e.g. Ethernet, Wi-Fi), sensor interfaces, motor module interface and CPU (e.g. PCIe and USB) to the GEM format at the ingress. The nodes 204, 208 then are able to output to the encapsulated messages (e.g. GEMs) to the HARQ and/or node transmission MAC (described below). At the egress, it accepts GEM-packets from the node reception MAC (received from the root port 230 and/or another node 204, 208, 234) and decapsulates the GEM back to the original data format (as received from the coupled device 102) for output to the device 102 via one of the ports 99. Like in the root ports 230, the HARQ of the nodes 204, 208 perform the hybrid automatic-repeat-request function to ensure that the GEM-Packets are delivered to their destination node or nodes 204, 208, 234 successfully. Specifically, the HARQ is able to be built-in with a repeat transmit timer, transmit GEM list flag table and receipt acknowledgment checking function (e.g. GEM RX- Acknowledge) to trigger GEM re-transmission when timer time-out occurs without receiving the acknowledgment.
The node MAC comprises a transmission MAC (TX MAC), a reception MAC (RX MAC), a security engine (e.g. AES), a forward error correction (FEC) engine, a DBA-Report engine and SERDES IP. The TX MAC is responsible for mapping/packing GEMs into a burst structure (e.g. Burst-PHY-Frame structure) and transmitting the burst messages to root ports 230 and/or nodes 204, 208, 234 during the burst window for the node granted by the dynamic burst allocation engine of the root port 230 for that node. The RX MAC is responsible for receiving and terminating broadcast messages (e.g. Broadcast-PHY-Frames) from root ports 230 and/or nodes 204, 208, 234, extracting GEMs from the broadcast message format, parsing and accepting GEMs addressed to it (e.g. addressed to one of its ports 99) based on the node’s SLA Profde setting, and subsequently outputting the data to the encapsulation/decapsulation engine.
The DBA report engine reports total data packet and message in queues (e.g. EPOCH Queues) to the HDBA engine of the associated root port 230 through the burst reporting (as described above). Additionally, the DBA report engine accepts GEM-Grant messages from the HDBA of the associated root port 230 and/or the DBA of the associated gate 202, and prepares the node transmission MAC to build a burst message (e.g. Burst-PHY-Frame) with the GEMs stored in the queues (e.g. EPOCH Queues).
The node activation processor is responsible for performing and completing the node 204, 208, 234 activation process and procedures between nodes 204, 206, 234 and root ports 230. The security engine is able to be an AES- 128/256 encryption and decryption functional block used for both the reception and transmission MACs. Alternatively, other encryption is able to be used. The FEC engine is used for controlling errors in data transmission over unreliable or noisy communication channels. In some embodiments, the FEC engine uses Reed Solomon FEC coding schemes of RS(255,216) and RS(225,232) for 10G and 2.5G data rates, respectively. The burst-mode SERDES uses fast clock and data recovery (CDR) locking mode to ensure fast fiber-cut, fast fail-over and protection switch recovery.
Finally, after activation processing (e.g. after the registration process is complete), the nodes 204, 206, 234 are able to broadcast a DDS message to entire bus 104 to inform and notice the root ports 230, switch 228, gates 202 and/or other nodes 204, 206, 234 that a new device has joined and registered to bus 104 at that node 204, 208, 234. Further, the nodes 204, 206, 234 are able to listen to DDS messages from the switch 228 and other new the nodes’ 204, 206, 234 declaration of joining the bus 104 and update their global SLA profde database and settings based on the DDS messages.
Gates
The gates 202 are able to comprise a node MAC (with multiple Virtual node State-Machines and buffering), an adaptive domain bridge (ADB), a root port MAC (with built-in gate DBA functionality/gate DBA), a gate SLA profile database and a burst-mode SERDES. The node MAC comprises one or more of a transmission MAC, reception MAC, security engine (e.g. AES), FEC engine, DBA report functional module, SERDES functional module and/or multiple sets (e.g. one for each node within the subnetwork 210) of virtual node processors, virtual node profdes and settings, and related MIB counters and reporting logics. The transmission MAC receives GEMs from the gate ADB and maps and packs then into their associated virtual node burst structure (e.g. Burst-PHY-Frame structure) based on the gate’s virtual node SLA Profile database settings. Further, the transmission MAC aggregates multiple virtual node burst structures (e.g. Burst-PHY-Frames) into one gate burst structure (e.g. GATE/Turbo Burst-PHY-Frame) and transmits burst message to the root port 230 through the network 206 based on the granted burst window for those nodes 208 received from the HDBA of the root port 230. The node reception MAC receives broadcast messages (e.g. Broadcast-PHY-Frames) from the root port 230, extracts GEMs from the messages, parses the headers of the GEMs, determines which messages are for nodes 208 within the subnetwork 210 of the gate 202 based on the GEM -Headers and virtual nodes SLA Profile database settings and outputs those messages to the ADB.
The ADB performs a bridging function between the node MAC and the root MAC of the gates 202. Specifically, in the broadcast direction (from the root port 230 to the nodes 208), the ADB receives GEMs from node reception MAC and performs a GEM header lookup, checking and filtering function based on the gate virtual node profile database in order to accept GEMs belonging to nodes 208 of the gate’s 202 subnetwork 210. The ADB is then able to output those GEMs to root port transmission MAC of the gate 202. In the burst direction (from the nodes 208 to the root port 230), the ADB receives GEMs from root reception MAC, stores them in their associated virtual node buffer memory, and output them to the virtual node transmission MAC when their burst window start time arrives.
The root port MAC of the gates 202 comprise a transmission MAC, a reception MAC, a security engine (e.g. AES), an FEC engine, a gate DBA and burst mode SERDES modules. The transmission MAC is responsible for accepting GEMs from ADB, mapping and packing the GEMs into a broadcast format (e.g. Broadcast-PHY-Frame structure), and outputting the broadcast formatted frames to burst-mode SERDES. The reception MAC is responsible for receiving burst messages (e.g. Burst-PHY-Frames) from burst-mode SERDES (e.g. a far end node), extracting the GEMs from the messages, parsing and accept only GEMs targeted for nodes 208 within the gate’s 202 subnetwork 210 (as indicated based on the parsed GEM headers and the SLA Profile settings), and then outputting the GEMs to the ADB of the gate 202. The DBA of the gate 202 is an extension HDBA of the root ports 230. The gate DBA grants and allocates node burst windows based on the gate DBA SLA profile settings (which is a subset of the root HDBA). The gate SLA profile database includes a list of node identifiers belonging to this gate 202 (e.g. located within the subnetwork 210 of the gate 202), an SLA profile table of node identifiers for a gate DBA function and GEM forwarding information. The burst mode SERDES accepts broadcast messages (e.g. Broadcast-PHY-Frames) from the root transmission MAC and transmits to nodes 208 in the subnetwork 210 in the broadcast transmission direction. In reception direction, the burst mode SERDES receives burst messages (e.g. Burst-PHY-Frames) from nodes 208 through the subnetwork 210 and outputs them to the root reception MAC for message/ffame termination and GEM extraction. The main function of gates 202 is to extend the central transmission network 206 of one of the root ports 230 by bridging to one or more subnetworks 210 (and the nodes 208 therein ) through adaptive bridging. In particular, the gates 202 are able to burst messages from the nodes 208 and/or other gates 202' within their subnetwork 210 to the root port 230 of the network 206 they are in as if the burst traffic were coming from nodes within the central transmission network 206. Similarly, the gates 202 are able to broadcast messages received from other nodes 204, 208, 234, the switch 228 and/or root port 230 to the nodes 208 and/or other gates 202' within their subnetwork 210 they are in as if the nodes 208 and/or other gates 202' were within the central transmission network 206. As a result, the gates 202 are able to extend the central transmission networks 206 to additional nodes 208 and/or different types of subnetworks 210 while maintaining a burst/broadcast communication method within the central transmission networks 206.
In more detail, in the transmission Burst direction (e.g. from the nodes/gates to the root ports/switch/core), the burst window granting mechanism from node 208 to gate 202 to root 230 is able to comprise the following steps. First, the DBA of the gate 202 is a subset of the HDBA of the root port 230 (of the network 206 that the gate 202 is a part of) and therefore is transparent to the root port 230 and nodes 208. Second, when the gate 202 receives a burst window grant message (e.g. GEM-Grant message) broadcast from its root port 230, it uses the message header (e.g. GEM-Header) to lookup gate SLA profde database for GEM forwarding information. In other words, it uses the header data to determine if the grant message is for any of the nodes 208 within its subnetwork 210 as indicated in the gate SLA profile database. If the grant message is not for any of the nodes 208 of its subnetwork 210 the gate 202 drops the grant message, otherwise, the gate stores the message in its virtual node database, updates the database and broadcasts a new window grant message (e.g. GEM- Grant message) to all the nodes/gates in its subnetwork 210 that is directed to the node 208 to which the original grant message was directed. In response, the node 208 provides a burst message to the gate 202 and the gate 202 formats and/or otherwise prepares the message for bursting to the root port 230 at the burst window start indicated in the received window grant message for that node 208.
Third, in order to get best throughput bandwidth, high burst bandwidth efficiency and/or low transmission latency, gate 202 is able to adjust the grant window indicated in this new grant message to be at least a predetermined amount of time before the grant window indicated in the original grant message. In particular, this amount of time provides the gate 202 time to receive and format the burst data from the node 208 before bursting the data from the gate 202 to the root port 230 at the time indicated by the original window grant message. Indeed, by doing this for multiple nodes 208 at the same time, the gate 202 is able to aggregate the messages from multiple different nodes (e.g. multiple Burst-PHY-frames) into a single bigger burst message (e.g. GATE Burst-PHY-Frame).
Fourth, due to the protocols between gate traffic DBA reporting and the root port 230 window granting, root port 230 and gates 202 are able to maintain a group-membership list table and be aware of the virtual nodes 208 that each of the gates 230 below to as a group. Thus, when a node 208 issues a report message (e.g. GEM-Report) to HDBA of the root port 230, the gate 203 is able to intercept the report message, modify it to include the GEMs data temporarily stored in gate’s 202 virtual node buffer memory if there is any, and issue a new report message to HDBA of the root port 230. In other words, the gates 202 are able to combine reporting messages from the nodes in their subnetworks 210 in order to make the reporting more efficient.
Additionally, when HDBA of the root ports 230 are issuing a grant message (e.g. GEM-Grant message) to nodes 208 that are in a subnetwork 210, because they are aware of all of the nodes 208 that are in that subnetwork 210 (e.g. via the virtual node database), the HDBA of the root ports 230 are able to ensure that the grant windows for nodes 208 that belong to the same gate 202 and/or subnetwork 210 are in sequence/continuous order so that the gate 202 is able to combine and/or burst all the virtual node’s burst messages (e.g. burst-PHY-Frames) without each having a preamble except for the first one. This provides the benefit of reducing preamble overhead and increasing the burst bandwidth efficiency (especially for small bursts of GEM-Control messages).
In other words, for the data-path, the gates 202 receive burst messages (e.g. burst-PHY-frames) from burst-mode SERDES and far-end nodes 208, extracts the GEMs from the messages in the root reception MAC of the gate 202, stores the GEMs in their associated virtual NODE buffer memory and waits for the virtual node burst window grant to come in from the root port 230 for those virtual nodes 208. Then, the gates 202 are able to map and pack the stored GEMs for that node 208 and other nodes 208 back into the burst message format thereby aggregating multiple burst messages together into one bigger burst message in the node transmission MAC of the gates 202. Finally, the gates 202 are able to transmit this bigger burst message to the SERDES and to the root port 230 through the network 206 based on granted burst windows (e.g. the multiple consecutive virtual node burst windows of that gate 202).
Now looking to the broadcast direction (e.g. from the root ports/switch/core to the nodes/gates), again the gates 202 are able to extend central networks 206 to the subnetworks 210 while being transparent to both the root port 230 for their network 206 and the nodes 208 in their subnetwork 210. In order to effectuate this, the gates 202 are able to act like virtual nodes and receive broadcast messages (e.g. Broadcast-PHY-Frames) from the root ports 230, extract the GEMs from the messages, drop any GEMs that are not directed to one of the nodes 208/gates 202' in their subnetwork 210 (e.g. as indicated by the message headers and the gate SLA profile database). Otherwise, the gates 202 are able to use store-N-forward and/or cut-through schemes to pack and map the GEMs back into the root port broadcast message structure (e.g. Broadcast-PHY-Frame structure) in a root transmission MAC of the gate 202 and broadcast the new broadcast message to all the nodes 208 and/or gates 202' in its subnetwork 210.
Data Transmission Operation
In operation, the bus 104 operates using a burst/broadcast communication scheme wherein all data messages from the nodes 204, 208, 234 (and gates 202) are funneled to the core 200 using a burst transmission method where transmission windows that are dynamically adjustable in size (by the core 200) are granted to the nodes 204, 208, 234 such that they (or a gate 202 on their behalf) are able transmit their data messages as a “burst” within the granted window. If the transmitting node is in a subnetwork 210, the gate 202 (acting as a root port of that network 210) receives the bursted message from the node 208 through the subnetwork 210 and then subsequently bursts the message through the central network 206 to the core 200 (as if the node 208 was a part of the central network 206). In doing this burst communication, the gate 202 is able to aggregate burst messages from multiple nodes 208 within the subnetwork 210 thereby increasing efficiency and reducing the effects of the subnetwork’s 210 possibly increased latency relative to the central network 206. Indeed, this is able to be repeated for gates 202' within subnetworks 210 that provide a gate way to sub-subnetworks 210' and so on to support any number of “chained/gated” networks. Further, the gate 202 is able to be transparent to the core 200 and nodes 208 in this process such that messages do not need to be addressed to the gate 202.
The core 200 receives these messages (from one or more root ports 230 coupling the core 200 to each of the central networks 206), processes them (including modifying and/or determining their target destination), and broadcasts them (and any messages originating in the core 200) onto whichever central transmission network 206 the target node 204, 208, 234 (or gate 202 representing the target node 208) for that message is located. Like the burst communication above, if the target node 208 is within the subnetwork 210, the gate 202 bridging to that subnetwork 210 is able to receive/intercept the message from the core an rebroadcast the message to all of the node 208 (and/or gates 202') on the subnetwork 210. Any broadcast messages for target nodes 204 not on the subnetwork 210 (or a subnetwork thereof) are able to be discarded by the gate 202 in order to increase efficiency. Again, this process is transparent and able to be repeated by gates 202' within subnetworks 210 and so on for any number of chained networks to broadcast the messages through the networks. As a result, all the nodes 204, 208, 234 (and gates 202) on each of the networks 206 (and subnetworks 210 coupled thereto) receive all of the messages from the core 200 broadcast on that network 206 and merely need to look for which messages are directed to them while discarding the others.
In more detail, when the nodes 204, 208, 234 receive data from one or more external devices 102 through one or more of their IO ports 99, they store the data in a GEM-ID queue buffer memory and burst a report message (e.g. GEM-Report) to the root port 230 of the central network 206 that they are in (either directly or through one or more gates 202 if they are in a subnetwork 210 of the central network 206) and wait to be granted a burst window to transmit the input data. As described above, the gates 202 are able to collect and aggregate report messages from a plurality of the nodes 208 (and or gates 202') in their subnetwork 210 into a single bigger report message that the gate 202 is able to more efficiently burst to the root port 230 during the burst window for those ports 208.
At the same time, the nodes 204, 208, 234 are able to encapsulate the input data into the GEM format (fragmenting GEMs exceeding a predefined size into smaller GEMs), encrypt GEMs with the security key of the node 204, 208, 234, update the HARQ table, map and pack the GEMs into a burst format (e.g. Burst- PHY-Frame format) and perform encoding (e.g. FEC RS(255,216) encoding). Subsequently, upon grant and arrival of the burst window for each of the nodes, the nodes burst the GEMs including the input data to the associated root port 230.
The HDBA of the root ports 230 receive all of the report messages from the nodes 204, 208 (and/or gates 202) and perform a DBA analysis for each of the nodes 204, 208 based on the SLA profile database, latency sensitive level, traffic congestion feedback, committed information rate (CIR)/peak information rate (PIR) feedback and/or other factors to determine grant window burst size and start-time for each of the nodes 204, 208. Once the granted burst windows have been determined for one or more of the nodes 204, 208, the root port 230 broadcasts the windows to each of the nodes in a broadcast grant message (e.g. GEM-Grant) to all of the nodes 204, 208 in the associated central network 206 and/or any subnetworks 210 (via the gates 202). As described above, the broadcast messages from the root ports 230 are the same size, whereas the burst windows from the nodes 204, 208 to the root ports 230 are able to vary in size as dynamically assigned by the HDBA. The gates 202, upon receipt of the broadcast grant messages targeting nodes 208 within their subnetwork 210 (or a subnetwork thereof), broadcast new grant messages to all of the nodes 208 with the subnetwork 210. Specifically, these new grant messages are able to specifying burst windows that occur before the time indicated by the original/root port grant window. This is to ensure the gates 202 to receive (e.g. be “bursted”) the input data/GEMs from the port 208 before the original/root port grant window, thereby giving the gates 202 time to aggregate the data/GEMs from multiple nodes 208 and/or ports 99 into single larger messages for burst to the root port 230 when the original/root port grant window arrives. As a result, the gates 202 are able to make up for inefficiencies and/or slower aspects of the subnetworks 210 such that they do not slow down the efficiency of the central transmission networks 206.
Upon receipt of the burst messages including the GEMs (including the input data from the external devices 102), the root ports 230 are able to perform decoding (e.g. FEC RS(255,216) decoding) and error correction on the burst messages to decode and correct any transmission errors. The root ports 230 are then able to extract the GEMs from the burst messages (e.g. the transmission frame format), decrypt the extracted GEMs (e.g. with AES-128/256 and a source-node security key), bypass the GEM fragmentation block and pass GEMs to the switch 228. For each of the GEMs, the switch 228 is then able to perform a GEM-Header lookup, parse and classify Ethernet L2/L3 address and headers, process GEM forward flow-chart and determine GEM forwarding destination info, store the GEM in (cut-through) buffer-memory, and output the GEM to HARQ and to the destination root port 230 (e.g. the root port 230 whose network 206 or subnetwork 210 thereof includes the destination node 204, 208) based on the SLA database QoS output scheduler.
The root ports 230 receive the GEMs, perform GEM encryption (e.g. AES-128/256 encryption) with target node’s (or broadcast GEM’s) security key, pack and map GEMs into a broadcast message structure (e.g. Broadcast-Frame structure), encode the message (e.g. FEC RS(255,216) encoding), and finally broadcast the broadcast messages to all of the nodes 204, 208 in that root port’s network 206 and subnetworks 210 thereof. If the node 208 is within a subnetwork 210, the gate 202 to that subnetwork receives the broadcast message and broadcasts the message to all of the nodes 208 within the subnetwork 210. In some embodiments, the gates 202 filter out any broadcast messages that are not targeted to nodes 208 within its subnetwork 210 (or a subnetwork thereof) and only broadcasts the broadcast messages that do target one of those nodes 208. Alternatively, the gates 202 are able to rebroadcast all of the broadcast messages to the nodes 208 within its subnetwork 210 without determining if the messages relate to one of those nodes 208. All the nodes 204, 208 monitor the received broadcast messages, processing those intended for the node 204, 208 and discarding the others. Specifically, for the non-discarded messages, the nodes 204, 208 decode and error correct the messages (e.g. FEC RS(255,216) decoding), extract the GEMs from the broadcast message format (e.g. BC-PHY-Frame), decrypt the extracted GEM (e.g. with AES-128/256 and the destination node’s security key), decapsulate the data from the GEM format back to original IO-Port data format, and output the data through the designated IO port 99 to the external device 102. As a result, the bus 104 and system 100 provides the benefit of being able to combine multiple different networks having varying input data, varying processing speeds and data constraints while still maintaining low latency and high throughput needed for machine automation systems. This is a unique intranet system architecture and specially defined and optimized for such machine automation applications.
Figure 4 illustrates a block diagram of an exemplary computing device 400 configured to implement the system 100 according to some embodiments. In addition to the features described above, the external devices 102 are able to include some or all of the features of the device 400 described below. In general, a hardware structure suitable for implementing the computing device 400 includes a network interface 402, a memory 404, a processor 406, I/O device(s) 408 (e.g. reader), a bus 410 and a storage device 412. Alternatively, one or more of the illustrated components are able to be removed or substituted for other components well known in the art. The choice of processor is not critical as long as a suitable processor with sufficient speed is chosen. The memory 404 is able to be any conventional computer memory known in the art. The storage device 412 is able to include a hard drive, CDROM, CDRW, DVD, DVDRW, flash memory card or any other storage device. The computing device 400 is able to include one or more network interfaces 402. An example of a network interface includes a network card connected to an Ethernet or other type of LAN. The I/O device(s)
408 are able to include one or more of the following: keyboard, mouse, monitor, display, printer, modem, touchscreen, button interface and other devices. The operating software/applications 430 or function(s)/module(s) thereof are likely to be stored in the storage device 412 and memory 404 and processed as applications are typically processed. More or fewer components shown in Figure 4 are able to be included in the computing device 400. In some embodiments, machine automation system hardware 420 is included. Although the computing device 400 in Figure 4 includes applications 430 and hardware 420 for the system 100, the system 100 is able to be implemented on a computing device in hardware, firmware, software or any combination thereof. Figure 5 illustrates a method of operating a machine automation system 100 including an intelligent controller and sensor intranet bus 104 according to some embodiments. As shown in Figure 5, the nodes 204, 208 receive input data from a plurality of the external devices 102 via one or more ports 99 of the bus 104 at the step 502. The nodes 204, 208 burst the input data as burst messages to the core 200 in variable size burst windows at the step 504. In some embodiments, for each of the nodes 204, 208, the HDBA of the root ports 230 dynamically adjusts the burst window start time and size of the variable burst window and assign the adjusted window the corresponding node 204, 208 in a broadcast grant window message based on data traffic parameters reported from that one of the nodes 204, 208. In some embodiments, the gates 202 aggregate two or more burst messages including input data and/or traffic reporting received from the nodes 208 into single larger burst reporting or input data message for bursting to the core 200. In such embodiments, the gates 202 are able to omit portions of the received burst messages (e.g. preambles) in order to enhance the efficiency of the bus 104. In some embodiments, upon receiving the broadcast window grant messages from the core 200, the gates 202 adjust the original time of the burst window to an earlier time and broadcast the adjusted broadcast window grant messages to the nodes 208. As a result, the nodes 208 burst their data to the gates 202 before the window granted by the root port 230 such that the gates 202 are able to combine multiple burst messages together and burst them in the later original time window. The core 200 processes and broadcasts the input data as broadcast messages to each of the nodes 204, 208 within the central network 206 and subnetworks 210 required to reach the target node 204, 208 of the message at the step 506. The target node 204, 208 converts data of the broadcast message into a format accepted by the device 102 coupled to the node 204, 208 and outputs the data to the device 102 at the step 508. As a result, the method provides the advantage of enabling the bus 104 to maintain high speed despite the use of lower speed network mediums.
Device Modules
In some embodiments, the devices 102 are able to be device modules. Figure 9 illustrates a smart compliant actuator (SCA) and sensor module 900 according to some embodiments. The SCA and sensor module 900 is able to be one or more of the devices 102 of the machine automation system 100 described herein. In some embodiments, the smart compliant actuator (SCA) and sensor module 900 allows deviations from its own equilibrium position, depending on the applied external force, wherein the equilibrium position of the compliant actuator is defined as the position of the actuator where the actuator generates zero force or zero torque. As shown in Figure 9, the SCA and sensor module is able to comprise one or more motors 902, one or more sensors 904 and/or a control board 906 (for controlling the motors 902 and/or sensors 904) coupled together via a device network 908. In particular this type of module 900 is able to perform high-bandwidth and/or low-latency required machine automation tasks (e.g. coupled with one or more controller devices 102 via the bus 104). The motors 902 are able to include actuator motors to control the actuation of the module 900 (e.g. movement of a robot arm) and the sensors 904 are able to include image and/or magnetic sensors to input image data and/or detect the position of the module 900 (e.g. a current position of the robot arm, a position of the image sensor, sensed images from the front of a self-driving car, or other sensed data).
Figures 10A-C illustrate variants of the control board 906, 906', 906" according to some embodiments. As shown in Figure 10A, the control board 906 for a multi-connection mode module 900 is able to comprise a node system on chip (SoC) 1002, a transimpedance amplifier (TIA) and/or laser driver (LD) 1004, a bidirectional optical subassembly (BOS A) 1006, a power regulator 1008, a motor driver 1010, a compliant actuator motor and power control connector 1012, a motor control signal transceiver 1014, one or more sensors 1016, a optical splitter 1018, an input power connector 1020, one or more output power connectors 1022, a first fiber optic connector 1024 and one or more second fiber optic connectors 1026 all operatively coupled together. In particular, the BOS A 1006, splitter 1018 and fiber optic connectors 1024, 1026 are coupled together via fiber optic cable. Alternatively, one or more of the above elements are able to be omitted, their quantity increased or decreased and/or other elements are able to be added.
The control board 906 is able to be a flexible printed circuit board. The BOS A 1006 is able to comprise a transmitter optical sub-assembly (TOSA), a receiver optical sub-assembly (ROSA) and a wave division multiplexing (WDM) filter so that it can use bidirectional technology to support two wavelengths on each fiber. In some embodiments, the BOSA 1006 is a hybrid silicon photonics BOSA. The motor driver 1010 is able to be a pre-driver, gate driver or other type of driver. The compliant actuator motor and power control connector 1012 is able to transmit control and/or power signals to the motors 902.
The motor control signal transceiver 1014 is able to receive motor control signals and/or transmit motor, sensor and/or other data to one or more controller devices 102 via the bus 104. The sensors 1016 are able to comprise magnetic sensors and/or other types of sensors. For example, the sensors 1016 are able to sense a position and/or orientation of the module 900 and provide the positional data as feedback to the SoC 1002 and/or a controller device 102 coupled with the module 900 via the bus 104. The optical splitter 1018 is able to be built-in to the control board 906. The input power connector 1020 receives power for the control board 906. The output power connectors 1022 are configured to supply, transfer and/or forward power to one or more other boards/modules 900.
The first fiber optic connector 1024 is coupled with the fiber optic splitter 1018 which splits the cable into two or more cables. One cable couples with the BOSA 1006 for transmitting signals to and from the other elements of the board 906 and the remainder each couple with a different one of the one or more second fiber optic connectors 1026. The first fiber optic connector 1024 and/or second fiber optic connectors 1026 are able to be a pigtail fiber optic connection points and/or connectors 1024. Specifically, the pigtail fiber optical connection point and/or connector is able to comprise a single, short, usually tight-buffered, optical fiber that has an optical connector pre-installed on one end and a length of exposed fiber at the other end. The end of the pigtail is able to be stripped and fusion spliced to a single fiber of a multi-fiber trunk. Alternatively, other types of optical connection points and/or connectors 1024 are able to be used.
In operation within the control boards 906, 906', 906", the motor driver 1010 is able to receive pulse width modulated (PWM) control signals generated by the SoC 1002 (and/or the controller devices 102 via the SoC 1002) for controlling the torque, speed and/or other operations of the motors 902 of the SCA module 900 (via the compliant actuator motor and power control connector 1012). Additionally, the sensors 1016, the sensors 904 and/or the driver 1010 are able to provide motor and/or sensor status feedback to the SoC 1002 such that the SoC 1002 (and/or the controller devices 102 via the SoC 1002) are able to adjust the control signals based on the feedback in order to control the operation of the motors 902 and/or sensors 904. For example, the driver 1010 is able to provide motor current sensor feedback comprising phase-A current values, phase-B current values and phase-C current values, wherein an internal analog to digital converter (ADC) of the SoC 1002 converts the values to digital values and the SoC 1002 (and/or the controller devices 102 via the SoC 1002) adjusts the PWM control signals transmitted to the driver 1010 based on the motor current sensor feedback received from the driver 1010 thereby adjusting the speed, torque and/or other characteristics of the motors 902.
In operation within the system 100, the first fiber optic connector 1024 enables the board/module 900 to couple to the bus 104 via an optical fiber cable, while the splitter 1018 and the second fiber optic connectors 1026 enable the board/module 900 to couple to one or more additional boards/modules 900 via additional optical fiber cable (e.g. for receiving control signals from and/or sending data signals to one or more controller devices 102 coupled to other ports 99 of the bus 104. As a result, as shown in Figure 11A, the boards/modules 900 are able to couple to ports 99 of the bus 104 as a serial cascade wherein only a single port 99 is able to couple to a plurality of boards/modules 900. Specifically, as shown in Figure 11 A, one board 906 is optically coupled to the port 99 from the first fiber optic connector 1024 (via a fiber optic cable) and each subsequent board 906 has its first fiber optic connector 1024 coupled to the second fiber optic connector 1026 of another one of the boards 906 (all via fiber optic cables). Indeed, as shown in Figure 11 A, if the boards 906 include a plurality of second fiber optic connectors 1026, the cascade is able to branch into a tree structure where single boards/modules 900 are optically coupled to a plurality of other boards/modules 900. At the same time, the boards/modules 900 are able to share power in the same manner in which they are optically coupled via the input power connector 1020 of one or more of the module 900 receiving power from a power source and one or more of the other modules 900 receiving power by coupling their input power connector 1020 to the output power connector 1022 of another one of the modules 900.
Alternatively, as shown in Figure 1 IB, the control board 906' for a single-connection mode module 900 is able to not include the one or more second fiber optic connectors 1026 and/or the one or more output power connectors 1022. In some embodiments, as shown in Figure IOC, the control board 906" for a single-connection mode image sensor module 900 is able to further comprise one or more compliant actuator motors 1028 along with one or more image or other types of sensors 1030 (e.g. cameras, LIDAR, magnetic, ultrasound, infrared, radio frequency). In such embodiments, the motors 1028 are able to control the movement of the sensors 1030 while the sensors 1016 detect the position and/or orientation of the motors 1028 and/or sensors 1030. Alternatively, the control board 906" is able to be a multi connection mode image sensor module 900 further comprising the one or more second fiber optic connectors 1026 and/or the one or more output power connectors 1022.
As shown in Figure 11 A, these single-connection mode modules 900 and/or boards 906' and 906" are able to couple to the cascades or trees formed by the multi-connection mode modules 900 and/or couple in parallel to the bus 104. Additionally, as shown in Figure 1 IB, the system 100 is able to comprise one or more external optical splitters 1102, wherein one or more of the boards/modules 906, 906', 906" configured into serial cascades, trees and/or in parallel are able to be further parallelized and/or serialized in the coupling to the bus 104 using the external optical splitter 1102. For example, as shown in Figure 1 IB, an optical splitter 1102 is used to coupled to a single port 99, the output of a cascade of modules 900, one or more individual modules 900 and another splitter 1102. Although as shown in Figure 1 IB, the splitters 1102 are 1 to 4 splitters, they are able to be any ratio 1 to N as desired. Also although as shown in Figures 11 A and 1 IB, only the modules 906, 906', 906" are shown as being coupled to the bus 104, it is understood that any combination of other devices 102 are also able to be coupled to the bus 104 along with the modules. For example, one or more controller devices 102 are able to be coupled to the bus 104 for receiving data and issuing commands to the modules.
As a result, the modules 900 provide the benefit of enabling super high throughput and data bandwidth and can support up to lOx to lOOx of bandwidth and long distance compared to other modules. In particular, the ability to utilize optical communication along with serial cascading coupling allows the modules 900 to provide fast data transmission speed and super low latency without being disrupted by electromagnetic interference (EMI).
Further, the modules 900 are particularly advantages in the field of robotics, industrial automation and self-driving vehicles due to its ability to handle their high bandwidth and low latency demands for sensor data.
Figure 12 illustrates a method of operating a controller and sensor bus including a plurality of ports for coupling with a plurality of external machine automation devices of a machine automation system according to some embodiments. As shown in Figure 12, one or more controller devices 102 are coupled to one or more of the ports 99 of the bus 104 at the step 1202. The first fiber optic connector 1024 of one or more SCA and sensor modules 900 are coupled with one or more of the ports 99 at the step 1204. Messages are relayed between the controllers 104 and the SCA and sensor modules 900 through the bus 104 via the one or more central transmission networks 206 at the step 1206. The control boards 906 adjust operation of the SCA and sensor modules 900 based on the messages received from the controller devices 102 at the step 1208. In some embodiments, each of the SCA and sensor modules 900 is directly coupled in parallel to one of the ports 99 via the fiber optic cable. In some embodiments, coupling the SCA and sensor modules 900 includes coupling the SCA and sensor modules 900 in parallel to an optical splitter 1102 and coupling the optical splitter 1102 to the ports 99 via the fiber optic cable. In some embodiments, coupling the SCA and sensor modules 900 includes coupling the first fiber optic connector 1024 of a first of the SCA and sensor modules 900 to one of the ports 99 via the fiber optic cable and coupling the second fiber optic connector 1026 of the first of the SCA and sensor modules 900 to the first fiber optic connector 1024 of a second of the SCA and sensor modules 900.
The system 100 and machine automation controller and sensor bus 104 implementing a dynamic burst to broadcast transmission network has numerous advantages. Specifically, it provides the benefit of a simple cable system and connection; the elimination of significant EMI impacts due to the user of optical fiber cable; guaranteed low latency for node-to-node communication; high throughput bandwidth from node to node transmission (10, 25, 100 or greater Gbps); can extend and reach up to 20km from node to node devices; low power consumption due to passive-optical-network architecture; industry grade QoS without traffic congestion due to centralized DBA scheduling mechanism; built-in HARQ mechanism to guarantee node-to-node and GEM transmission successful; and one unified software image for full intranet system including all gate, node and root ports enabling simplified software architecture, shorter product development cycle, and easier system level debug, monitoring and trouble shooting remotely.
The present invention has been described in terms of specific embodiments incorporating details to facilitate the understanding of principles of construction and operation of the invention. Such reference herein to specific embodiments and details thereof is not intended to limit the scope of the claims appended hereto. It will be readily apparent to one skilled in the art that other various modifications may be made in the embodiment chosen for illustration without departing from the spirit and scope of the invention as defined by the claims. For example, although as described herein the bus is described as operating within a machine automation system, it is understood that the bus is able to operate with other types of systems and devices thereof for facilitating the communication between the devices.

Claims

C L A I M S What is claimed is:
1. A machine automation system for controlling and operating an automated machine, the system comprising: a controller and sensor bus including plurality of ports, wherein the bus comprises a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and one or more subnetworks, the central optical fiber transmission networks including a plurality of nodes and one or more gates and the plurality of subnetworks each coupled to a different one of the gates of one of the central transmission networks, the subnetworks including a plurality of subnodes, wherein each of the nodes and the subnodes are associated with one or more of the ports; one or more controllers each coupled with first ones of the ports of the bus; and one or more compliant actuator modules each including a first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC), the modules each operably coupled with second ones of the ports via fiber optic cable coupled from the first fiber optic connector to the second ones of the ports; wherein the nodes and the subnodes of the bus relay messages including sensor data from the sensors of the compliant actuator modules and control data from the controllers between the controllers and the compliant actuator modules through the bus via the one or more central optical fiber transmission networks.
2. The system of claim 1, wherein each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable.
3. The system of claim 1, wherein a plurality of the compliant actuator modules are coupled in parallel to an optical splitter and the optical splitter is coupled to one of the second ones of the ports via the fiber optic cable.
4. The system of claim 1, wherein each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC.
5. The system of claim 4, wherein the compliant actuator modules are coupled in series such that the first fiber optic connector of a first of the compliant actuator modules is coupled to one of the second ones of the ports via the fiber optic cable and the second fiber optic connector of the first of the compliant actuator modules is coupled to the first fiber optic connector of a second of the compliant actuator modules.
6. The system of claim 5, wherein each of the compliant actuator modules comprise a bi directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
7. The system of claim 6, wherein the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, wherein the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor .
8. The system of claim 1, wherein the first fiber optic connectors are pigtail fiber optic connection points.
9. The system of claim 1, wherein the second ones of the ports are optical ports coupled with pigtail fiber optic connectors.
10. The system of claim 1, wherein the automated machine is one of the groups consisting of a robot and a self-driving vehicle.
11. A compliant actuator module for use with a controller and sensor bus, the compliant actuator module comprising: one or more compliant actuator motors; one or more magnetic sensors; and a control board including a first fiber optic connector, a second fiber optic connector, a system on chip (SoC) and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC, wherein the control board: receives sensor data from the magnetic sensors and outputs status messages based on the sensor data via the first fiber optic connector; controls the motors based on control messages input via the first fiber optic connector having a target wavelength; forwards status messages received via the second fiber optic connector out the first fiber optic connector with the optical splitter; and forwards the control messages input via the first fiber optic connector having a non target wavelength.
12. The compliant actuator module of claim 11, further comprising an image sensor, wherein the control board receives image data from the image sensor and outputs sensor data messages based on the image data via the first fiber optic connector.
13. The compliant actuator module of claim 12, wherein the motors control the articulation of the image sensor and the magnetic sensors determine the orientation of the image sensor.
14. The compliant actuator module of claim 13, wherein the control board comprises a bi directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
15. The compliant actuator module of claim 14, wherein the first fiber optic connector is a pigtail fiber optic connection point.
16. A method of operating a controller and sensor bus including a plurality of ports for coupling with a plurality of external machine automation devices of a machine automation system, the bus having a central processing core and a multi-medium transmission intranet including one or more central optical fiber transmission networks directly coupled to the core and including a plurality of nodes and one or more gates and a plurality of subnetworks each coupled to a different one of the gates of one of the central optical fiber transmission networks, the subnetworks including a plurality of subnodes, the method comprising: coupling one or more controllers with first ones of the ports of the bus; coupling one or more compliant actuator modules with second ones of the ports via fiber optic cable coupled from a first fiber optic connector to the second ones of the ports, the compliant actuator modules each including the first fiber optic connector, one or more motors, one or more sensors and a system on chip (SoC); and relaying messages between the controllers and the compliant actuator modules through the bus via the one or more central optical fiber transmission networks with the nodes and the subnodes of the bus, the messages including sensor data from the sensors of the compliant actuator modules and control data from the controllers.
17. The method of claim 16, wherein each of the compliant actuator modules is directly coupled in parallel to one of the second ones of the ports via the fiber optic cable.
18. The method of claim 16, wherein coupling the compliant actuator modules includes coupling the compliant actuator modules in parallel to an optical splitter and coupling the optical splitter to one of the second ones of the ports via the fiber optic cable.
19. The method of claim 16, wherein each of the compliant actuator modules comprise a second fiber optic connector and an optical splitter, wherein the optical splitter is coupled with the first fiber optic connector, the second fiber optic connector and the SoC.
20. The method of claim 19, wherein coupling the compliant actuator modules includes coupling the first fiber optic connector of a first of the compliant actuator modules to one of the second ones of the ports via the fiber optic cable and coupling the second fiber optic connector of the first of the compliant actuator modules to the first fiber optic connector of a second of the compliant actuator modules.
21. The method of claim 20, wherein each of the compliant actuator modules comprise a bi-directional optical sub-assembly (BOSA), a transimpedance amplifier, a laser driver and a motor driver.
22. The method of claim 21, wherein the sensors of one or more of the compliant actuator modules include an image sensor and magnetic sensors, further comprising controlling the articulation of the image sensor with the motors and determining the orientation of the image sensor with the magnetic sensors.
23. The method of claim 16, wherein the first fiber optic connectors are pigtail fiber optic connection points.
24. The method of claim 16, wherein the second ones of the ports are optical ports coupled with pigtail fiber optic connectors.
25. The method of claim 16, wherein the automated machine is one of the groups consisting of a robot and a self-driving vehicle.
PCT/US2020/053589 2019-10-15 2020-09-30 Intelligent controller and sensor network bus, system and method including smart compliant actuator module WO2021076333A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202080005480.4A CN112867997A (en) 2019-10-15 2020-09-30 Intelligent controller including intelligent flexible actuator module, and sensor network bus, system and method

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US16/653,558 US11269316B2 (en) 2019-08-01 2019-10-15 Intelligent controller and sensor network bus, system and method including smart compliant actuator module
US16/653,558 2019-10-15

Publications (1)

Publication Number Publication Date
WO2021076333A1 true WO2021076333A1 (en) 2021-04-22

Family

ID=75538357

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2020/053589 WO2021076333A1 (en) 2019-10-15 2020-09-30 Intelligent controller and sensor network bus, system and method including smart compliant actuator module

Country Status (2)

Country Link
CN (1) CN112867997A (en)
WO (1) WO2021076333A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11809163B2 (en) 2019-08-01 2023-11-07 Vulcan Technologies Shanghai Co., Ltd. Intelligent controller and sensor network bus, system and method including a message retransmission mechanism

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6091527A (en) * 1997-03-05 2000-07-18 Nokia Mobile Phones Limited Communications device having an optical bus, and a method for controlling its operation
US8380036B2 (en) * 2009-01-20 2013-02-19 Adc Telecommunications, Inc. Splitter module with connectorized pigtail manager
US9071352B2 (en) * 2005-09-02 2015-06-30 Georgios Margaritis Free space optics alignment method and apparatus
US9674432B2 (en) * 2013-06-28 2017-06-06 Olympus Corporation Image capturing apparatus
US20190007234A1 (en) * 2015-08-31 2019-01-03 National University Corporation Nagoya University Relay device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6091527A (en) * 1997-03-05 2000-07-18 Nokia Mobile Phones Limited Communications device having an optical bus, and a method for controlling its operation
US9071352B2 (en) * 2005-09-02 2015-06-30 Georgios Margaritis Free space optics alignment method and apparatus
US8380036B2 (en) * 2009-01-20 2013-02-19 Adc Telecommunications, Inc. Splitter module with connectorized pigtail manager
US9674432B2 (en) * 2013-06-28 2017-06-06 Olympus Corporation Image capturing apparatus
US20190007234A1 (en) * 2015-08-31 2019-01-03 National University Corporation Nagoya University Relay device

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
KRAUS ET AL.: "Approach for an Optical Network Design for Autonomous Vehicles", ICTON 2019, 19 September 2019 (2019-09-19), XP033617910, Retrieved from the Internet <URL:https://ieeexplore.ieee.org/document/8840176> [retrieved on 20201127], DOI: 10.1109/ICTON.2019.8840176 *
KRAUS ET AL.: "Replacement of the Controller Area Network (CAN) Protocol for Future Automotive Bus System Solutions by Substitution via Optical Networks", ICTON 2016, 25 August 2016 (2016-08-25), XP032949859, Retrieved from the Internet <URL:https://ieeexplore.ieee.org/document/7550335> [retrieved on 20201127], DOI: 10.11Q9/ICTON.2016.7550335 *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11809163B2 (en) 2019-08-01 2023-11-07 Vulcan Technologies Shanghai Co., Ltd. Intelligent controller and sensor network bus, system and method including a message retransmission mechanism

Also Published As

Publication number Publication date
CN112867997A (en) 2021-05-28

Similar Documents

Publication Publication Date Title
US11269316B2 (en) Intelligent controller and sensor network bus, system and method including smart compliant actuator module
US10841230B1 (en) Intelligent controller and sensor network bus, system and method
US11269795B2 (en) Intelligent controller and sensor network bus, system and method including a link media expansion and conversion mechanism
US11689386B2 (en) Intelligent controller and sensor network bus, system and method for controlling and operating an automated machine including a failover mechanism for multi-core architectures
US11089140B2 (en) Intelligent controller and sensor network bus, system and method including generic encapsulation mode
US11086810B2 (en) Intelligent controller and sensor network bus, system and method including multi-layer platform security architecture
US11156987B2 (en) Intelligent controller and sensor network bus, system and method including a message retransmission mechanism
US11263157B2 (en) Intelligent controller and sensor network bus, system and method including a dynamic bandwidth allocation mechanism
US11258538B2 (en) Intelligent controller and sensor network bus, system and method including an error avoidance and correction mechanism
US9609065B2 (en) Bridge for implementing a converged network protocol to facilitate communication between different communication protocol networks
US11169591B2 (en) Ethernet PHY-MAC communication with in-band wake-up/sleep commands
US11809163B2 (en) Intelligent controller and sensor network bus, system and method including a message retransmission mechanism
EP2003823A1 (en) Autonegotiation over an interface for which no autonegotiation standard exists
CN114270328B (en) Intelligent controller and sensor network bus and system and method including multi-layered platform security architecture
JP2022547143A (en) DATA TRANSMISSION CONTROL METHOD, DEVICE, AND STORAGE MEDIUM
GB2532054A (en) NC-SI port controller
US7145877B2 (en) Apparatus and method for distance extension of fibre-channel over transport
WO2021076333A1 (en) Intelligent controller and sensor network bus, system and method including smart compliant actuator module
CN113890680A (en) Transmission method applied to fiber channel avionics network DDS
WO2021055205A1 (en) Intelligent controller and sensor network bus, system and method including generic encapsulation mode
WO2022086723A1 (en) Intelligent controller and sensor network bus, system and method including a link media expansion and conversion mechanism
WO2022076730A1 (en) Intelligent controller and sensor network bus, system and method including a dynamic bandwidth allocation mechanism
CN114208258B (en) Intelligent controller and sensor network bus and system and method including message retransmission mechanism
WO2017020292A1 (en) Optical access device and optical access system
CN108614786B (en) Channel management circuit based on message service type

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20877739

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20877739

Country of ref document: EP

Kind code of ref document: A1