US20140108532A1 - System and method for supporting guaranteed multi-point delivery in a distributed data grid - Google Patents
System and method for supporting guaranteed multi-point delivery in a distributed data grid Download PDFInfo
- Publication number
- US20140108532A1 US20140108532A1 US13/671,369 US201213671369A US2014108532A1 US 20140108532 A1 US20140108532 A1 US 20140108532A1 US 201213671369 A US201213671369 A US 201213671369A US 2014108532 A1 US2014108532 A1 US 2014108532A1
- Authority
- US
- United States
- Prior art keywords
- node
- nodes
- data grid
- list
- incoming message
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L43/00—Arrangements for monitoring or testing data switching networks
- H04L43/08—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
- H04L43/0805—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
- H04L43/0811—Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking connectivity
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/202—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
- G06F11/2041—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant with more than one idle spare processing component
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/2097—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements maintaining the standby controller/processing unit updated
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/10—File systems; File servers
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/20—Information retrieval; Database structures therefor; File system structures therefor of structured data, e.g. relational data
- G06F16/23—Updating
- G06F16/2365—Ensuring data consistency and integrity
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L67/00—Network arrangements or protocols for supporting network services or applications
- H04L67/01—Protocols
- H04L67/10—Protocols in which an application is distributed across nodes in the network
- H04L67/1095—Replication or mirroring of data, e.g. scheduling or transport for data synchronisation between network nodes
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F11/00—Error detection; Error correction; Monitoring
- G06F11/07—Responding to the occurrence of a fault, e.g. fault tolerance
- G06F11/16—Error detection or correction of the data by redundancy in hardware
- G06F11/20—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
- G06F11/202—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
- G06F11/2048—Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant where the redundant components share neither address space nor persistent storage
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2201/00—Indexing scheme relating to error detection, to error correction, and to monitoring
- G06F2201/82—Solving problems relating to consistency
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04L—TRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
- H04L41/00—Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
- H04L41/06—Management of faults, events, alarms or notifications
- H04L41/0654—Management of faults, events, alarms or notifications using network fault recovery
- H04L41/0668—Management of faults, events, alarms or notifications using network fault recovery by dynamic selection of recovery network elements, e.g. replacement by the most appropriate element after failure
Definitions
- the present invention is generally related to computer systems, and is particularly related to a distributed data grid.
- Modern computing systems particularly those employed by larger organizations and enterprises, continue to increase in size and complexity.
- areas such as Internet applications
- millions of users should be able to simultaneously access that application, which effectively leads to an exponential increase in the amount of content generated and consumed by users, and transactions involving that content.
- Such activity also results in a corresponding increase in the number of transaction calls to databases and metadata stores, which have a limited capacity to accommodate that demand.
- a messaging facility in the distributed data grid can receive an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid.
- the messaging facility can deliver the incoming message to the plurality of nodes according to an order in a list.
- a node in the plurality of nodes operates to skip a next node in the list to deliver the incoming message, when the next node is dead or unavailable.
- FIG. 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention.
- FIG. 2 is an illustration of supporting guaranteed multi-point message delivery in a distributed data grid in accordance with various embodiments of the invention.
- FIG. 3 is an illustration of creating a chain request message in a distributed data grid in accordance with various embodiments of the invention.
- FIG. 4 is an illustration of handling a chain request message in a distributed data grid in accordance with various embodiments of the invention.
- FIG. 5 is an illustration of supporting consistent message delivery when a primary owner node of a partition becomes unavailable in a distributed data grid in accordance with various embodiments of the invention.
- FIG. 6 illustrates an exemplary flow chart for supporting guaranteed multi-point message delivery in a distributed data grid in accordance with an embodiment of the invention.
- Described herein are systems and methods that can support guaranteed multi-point message delivery in a distributed data grid.
- a “distributed data grid”, “data grid cluster”, or “data grid”, is a system comprising a plurality of computer servers which work together to manage information and related operations, such as computations, within a distributed or clustered environment.
- the data grid cluster can be used to manage application objects and data that are shared across the servers.
- a data grid cluster should have low response time, high throughput, predictable scalability, continuous availability and information reliability. As a result of these capabilities, data grid clusters are well suited for use in computational intensive, stateful middle-tier applications.
- Some examples of data grid clusters can store the information in-memory to achieve higher performance, and can employ redundancy in keeping copies of that information synchronized across multiple servers, thus ensuring resiliency of the system and the availability of the data in the event of server failure.
- Coherence provides replicated and distributed (partitioned) data management and caching services on top of a reliable, highly scalable peer-to-peer clustering protocol.
- An in-memory data grid can provide the data storage and management capabilities by distributing data over a number of servers working together.
- the data grid can be middleware that runs in the same tier as an application server or within an application server. It can provide management and processing of data and can also push the processing to where the data is located in the grid.
- the in-memory data grid can eliminate single points of failure by automatically and transparently failing over and redistributing its clustered data management services when a server becomes inoperative or is disconnected from the network. When a new server is added, or when a failed server is restarted, it can automatically join the cluster and services can be failed back over to it, transparently redistributing the cluster load.
- the data grid can also include network-level fault tolerance features and transparent soft re-start capability.
- the functionality of a data grid cluster is based on using different cluster services.
- the cluster services can include root cluster services, partitioned cache services, and proxy services.
- each cluster node can participate in a number of cluster services, both in terms of providing and consuming the cluster services.
- Each cluster service has a service name that uniquely identifies the service within the data grid cluster, and a service type, which defines what the cluster service can do.
- the services can be either configured by the user, or provided by the data grid cluster as a default set of services.
- FIG. 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention.
- a data grid cluster 100 e.g. an Oracle Coherence data grid, includes a plurality of cluster nodes 101 - 106 having various cluster services 111 - 116 running thereon. Additionally, a cache configuration file 110 can be used to configure the data grid cluster 100 .
- a distributed data grid can support guaranteed multi-point message delivery, which can provide consistency in delivering messages among different cluster nodes in the distributed data grid.
- FIG. 2 is an illustration of supporting consistent message delivery in a distributed data grid in accordance with various embodiments of the invention.
- a distributed data grid 200 can include a plurality of cluster nodes, such as node A-D 201 - 204 .
- a cluster node e.g. node A 201
- the cluster node A 201 can use a message facility 211 to configure and deliver a message to different cluster nodes in the distributed data grid 200 , such as nodes B-D 202 - 204 .
- the cluster node A 201 can deliver the message to the recipient nodes B-D 202 - 204 in a particular order, e.g. based on a list from node B 202 to node C 203 then to node D 204 .
- the message facility 211 on the cluster node A 201 can be used to assign the order to the recipient nodes B-D 202 - 204 in the distributed data grid 200 .
- each recipient node B-D 202 - 204 in the list can keep track of how the message is to be delivered down the list. For example, a recipient node, e.g. node B 202 , can detect that node C 203 , which is the next node in the list, is dead or temporarily unavailable. Then, the node B 202 can skip the node C 203 , and deliver the message directly to node D 204 , which is the node next to the node C 203 on the list.
- the cluster node A 201 can receive an incoming message from a client 210 .
- the incoming message can include a request from the client 210 , and the client 210 may expect a response from the distributed data grid 200 .
- a response message can be sent back to the client in the reverse direction, e.g. from node D 204 to node C 203 then to node B 202 before reaching node A 201 .
- the distributed data grid 200 can provide the response message to the client 210 , after the incoming message is delivered to the recipient nodes B-D 202 - 204 and processed in the distributed data grid 200 .
- the guaranteed multi-point message delivery feature can be used for managing partition backups.
- a cluster node e.g. the node A 201
- nodes B-D are backup nodes for the node A 201 .
- the cluster node A 201 can deliver the message to each recipient node B-D 202 - 204 separately, or in parallel (as shown in dotted line in FIG. 2 ).
- the distributed data grid 200 may not be able to maintain a consistent view that the value of x equals to 2.
- This alternative approach can cause inconsistency, in terms of determining the value of x at a later time point, among the different cluster nodes A-D 201 - 204 in the distributed data grid 200 .
- the guaranteed multi-point message delivery feature can be used complimentarily with a partition versioning feature supported in the distributed data grid 200 .
- the distributed data grid 200 can bring the state of the newly added node E 205 current, based on the partition versioning feature, so that the node E 205 can start receive new messages based on the guaranteed multi-point message delivery feature, as described above.
- the guaranteed multi-point message delivery feature can be used complimentarily with a poll model that is supported in the distributed data grid 200 for processing incoming messages asynchronizingly.
- FIG. 3 is an illustration of creating a chain request message in a distributed data grid in accordance with various embodiments of the invention.
- a cluster node A 301 in the distributed data grid 300 can create a chain request message 320 , e.g. using a messaging facility 311 on the cluster node A 301 .
- the chain request message 320 can be either initiated by the cluster node A 301 , or be created based on an incoming message 310 received by the cluster node A 301 .
- the chain request message 320 can include an internal data structure 321 that stores the information about a list of recipient nodes that the chain request message 320 will be delivered to.
- the cluster node A 301 can deliver the chain request message 320 to the different cluster nodes in the distributed data grid 300 , e.g. nodes B-C 302 - 203 , for further processing.
- FIG. 4 is an illustration of handling a chain request message in a distributed data grid in accordance with various embodiments of the invention.
- a cluster node e.g. node B 402
- the distributed data grid 400 can receive a chain request message 420 that contains a list of recipient nodes in an internal data structure 421 .
- a messaging facility 412 in the cluster node B 402 can track the delivery of the chain request message 420 to the rest of the recipient nodes in the list 421 . For example, when the cluster node B 402 detects that node C 403 is dead or is unavailable, the cluster node B 402 can access the internal data structure 421 for the list of recipient nodes and find out that node D 204 is the next node following node C 203 . Therefore, the cluster node B 402 can deliver the chain request message 420 to node D 404 accordingly.
- FIG. 5 is an illustration of supporting consistent message delivery when a primary owner node of a partition becomes unavailable in a distributed data grid in accordance with various embodiments of the invention.
- a distributed data grid 500 can include a plurality of cluster nodes, such as node A-D 501 - 504 .
- the partition can be maintained on the primary owner node A 501 and each backup node B-D 502 - 504 .
- the distributed data grid 500 can guarantee that the message is delivered to the rest of nodes C-D 503 - 504 , and can ensure a consistent view that the value of x is equal to 2.
- the distributed data grid 500 maintains the consistent view that the value of x is equal to 1.
- the distributed data grid 500 can ensure a consistent view of the value of x client 510 in either case.
- the distributed data grid 500 can provide a new primary owner node, e.g. cluster node E 505 , which can continue maintaining the partition and handle subsequent incoming messages from the client 510 .
- a new primary owner node e.g. cluster node E 505
- FIG. 6 illustrates an exemplary flow chart for supporting guaranteed multi-point message delivery in a distributed data grid in accordance with an embodiment of the invention.
- a messaging facility on a cluster node in the distributed data grid can receive an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid.
- the messaging facility can be configured to deliver the incoming message to the plurality of nodes according to an order in a list.
- a node in the plurality of nodes can skip a next node in the list for delivering the incoming message, when the next node is dead or unavailable.
- the present invention may be conveniently implemented using one or more conventional general purpose or specialized digital computer, computing device, machine, or microprocessor, including one or more processors, memory and/or computer readable storage media programmed according to the teachings of the present disclosure.
- Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art.
- the present invention includes a computer program product which is a storage medium or computer readable medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention.
- the storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- Quality & Reliability (AREA)
- Data Mining & Analysis (AREA)
- Databases & Information Systems (AREA)
- Signal Processing (AREA)
- Computer Networks & Wireless Communication (AREA)
- Computer Security & Cryptography (AREA)
- Environmental & Geological Engineering (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Software Systems (AREA)
- Computer And Data Communications (AREA)
- Mathematical Physics (AREA)
- Multi Processors (AREA)
- Data Exchanges In Wide-Area Networks (AREA)
Abstract
A system and method can support guaranteed multi-point message delivery in a distributed data grid. A messaging facility in the distributed data grid can receive an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid. The messaging facility can deliver the incoming message to the plurality of nodes according to an order in a list. Furthermore, a node in the plurality of nodes operates to skip a next node in the list to deliver the incoming message, when the next node is dead or unavailable.
Description
- This application claims priority on U.S. Provisional Patent Application No. 61/714,100, entitled “SYSTEM AND METHOD FOR SUPPORTING A DISTRIBUTED DATA GRID IN A MIDDLEWARE ENVIRONMENT,” by inventors Robert H. Lee, Gene Gleyzer, Charlie Helin, Mark Falco, Ballav Bihani and Jason Howes, filed Oct. 15, 2012, which application is herein incorporated by reference.
- A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
- The current application hereby incorporates by reference the material in the following patent applications:
- U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR PROVIDING PARTITION PERSISTENT STATE CONSISTENCY IN A DISTRIBUTED DATA GRID,” by inventors Robert H. Lee and Gene Gleyzer, filed ______ (Attorney Docket No.: ORACL-05359US0).
- U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR PROVIDING TRANSIENT PARTITION CONSISTENCY IN A DISTRIBUTED DATA GRID,” by inventors Robert H. Lee and Gene Gleyzer, filed ______(Attorney Docket No.: ORACL-05359US1).
- U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR SUPPORTING ASYNCHRONOUS MESSAGE PROCESSING IN A DISTRIBUTED DATA GRID,” by inventor Gene Gleyzer, filed ______ (Attorney Docket No.: ORACL-05360US0).
- U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR SUPPORTING OUT-OF-ORDER MESSAGE PROCESSING IN A DISTRIBUTED DATA GRID,” by inventors Mark Falco and Gene Gleyzer, filed ______ (Attorney Docket No.: ORACL-05364US0).
- 1. Field of Invention
- The present invention is generally related to computer systems, and is particularly related to a distributed data grid.
- 2. Background
- Modern computing systems, particularly those employed by larger organizations and enterprises, continue to increase in size and complexity. Particularly, in areas such as Internet applications, there is an expectation that millions of users should be able to simultaneously access that application, which effectively leads to an exponential increase in the amount of content generated and consumed by users, and transactions involving that content. Such activity also results in a corresponding increase in the number of transaction calls to databases and metadata stores, which have a limited capacity to accommodate that demand.
- This is the general area that embodiments of the invention are intended to address.
- Described herein are systems and methods that can support guaranteed multi-point message delivery in a distributed data grid. A messaging facility in the distributed data grid can receive an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid. The messaging facility can deliver the incoming message to the plurality of nodes according to an order in a list. Furthermore, a node in the plurality of nodes operates to skip a next node in the list to deliver the incoming message, when the next node is dead or unavailable.
-
FIG. 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention. -
FIG. 2 is an illustration of supporting guaranteed multi-point message delivery in a distributed data grid in accordance with various embodiments of the invention. -
FIG. 3 is an illustration of creating a chain request message in a distributed data grid in accordance with various embodiments of the invention. -
FIG. 4 is an illustration of handling a chain request message in a distributed data grid in accordance with various embodiments of the invention. -
FIG. 5 is an illustration of supporting consistent message delivery when a primary owner node of a partition becomes unavailable in a distributed data grid in accordance with various embodiments of the invention. -
FIG. 6 illustrates an exemplary flow chart for supporting guaranteed multi-point message delivery in a distributed data grid in accordance with an embodiment of the invention. - Described herein are systems and methods that can support guaranteed multi-point message delivery in a distributed data grid.
- In accordance with an embodiment, as referred to herein a “distributed data grid”, “data grid cluster”, or “data grid”, is a system comprising a plurality of computer servers which work together to manage information and related operations, such as computations, within a distributed or clustered environment. The data grid cluster can be used to manage application objects and data that are shared across the servers. Preferably, a data grid cluster should have low response time, high throughput, predictable scalability, continuous availability and information reliability. As a result of these capabilities, data grid clusters are well suited for use in computational intensive, stateful middle-tier applications. Some examples of data grid clusters, e.g., the Oracle Coherence data grid cluster, can store the information in-memory to achieve higher performance, and can employ redundancy in keeping copies of that information synchronized across multiple servers, thus ensuring resiliency of the system and the availability of the data in the event of server failure. For example, Coherence provides replicated and distributed (partitioned) data management and caching services on top of a reliable, highly scalable peer-to-peer clustering protocol.
- An in-memory data grid can provide the data storage and management capabilities by distributing data over a number of servers working together. The data grid can be middleware that runs in the same tier as an application server or within an application server. It can provide management and processing of data and can also push the processing to where the data is located in the grid. In addition, the in-memory data grid can eliminate single points of failure by automatically and transparently failing over and redistributing its clustered data management services when a server becomes inoperative or is disconnected from the network. When a new server is added, or when a failed server is restarted, it can automatically join the cluster and services can be failed back over to it, transparently redistributing the cluster load. The data grid can also include network-level fault tolerance features and transparent soft re-start capability.
- In accordance with an embodiment, the functionality of a data grid cluster is based on using different cluster services. The cluster services can include root cluster services, partitioned cache services, and proxy services. Within the data grid cluster, each cluster node can participate in a number of cluster services, both in terms of providing and consuming the cluster services. Each cluster service has a service name that uniquely identifies the service within the data grid cluster, and a service type, which defines what the cluster service can do. Other than the root cluster service running on each cluster node in the data grid cluster, there may be multiple named instances of each service type. The services can be either configured by the user, or provided by the data grid cluster as a default set of services.
-
FIG. 1 is an illustration of a data grid cluster in accordance with various embodiments of the invention. As shown inFIG. 1 , a data grid cluster 100, e.g. an Oracle Coherence data grid, includes a plurality of cluster nodes 101-106 having various cluster services 111-116 running thereon. Additionally, a cache configuration file 110 can be used to configure the data grid cluster 100. - In accordance with various embodiments of the invention, a distributed data grid can support guaranteed multi-point message delivery, which can provide consistency in delivering messages among different cluster nodes in the distributed data grid.
-
FIG. 2 is an illustration of supporting consistent message delivery in a distributed data grid in accordance with various embodiments of the invention. As shown inFIG. 2 , a distributed data grid 200 can include a plurality of cluster nodes, such as node A-D 201-204. - A cluster node,
e.g. node A 201, can be either the originator of an internal message, or a recipient of an internal message. Additionally, thecluster node A 201 can also be a recipient of an incoming message from aclient 210. Thecluster node A 201 can use amessage facility 211 to configure and deliver a message to different cluster nodes in the distributed data grid 200, such as nodes B-D 202-204. - The
cluster node A 201 can deliver the message to the recipient nodes B-D 202-204 in a particular order, e.g. based on a list fromnode B 202 tonode C 203 then tonode D 204. Here, themessage facility 211 on thecluster node A 201 can be used to assign the order to the recipient nodes B-D 202-204 in the distributed data grid 200. - Furthermore, each recipient node B-D 202-204 in the list can keep track of how the message is to be delivered down the list. For example, a recipient node,
e.g. node B 202, can detect thatnode C 203, which is the next node in the list, is dead or temporarily unavailable. Then, thenode B 202 can skip thenode C 203, and deliver the message directly tonode D 204, which is the node next to thenode C 203 on the list. - As shown in
FIG. 2 , thecluster node A 201 can receive an incoming message from aclient 210. The incoming message can include a request from theclient 210, and theclient 210 may expect a response from the distributed data grid 200. Then, a response message can be sent back to the client in the reverse direction, e.g. fromnode D 204 tonode C 203 then tonode B 202 before reachingnode A 201. Finally, the distributed data grid 200 can provide the response message to theclient 210, after the incoming message is delivered to the recipient nodes B-D 202-204 and processed in the distributed data grid 200. - In accordance with various embodiments of the invention, the guaranteed multi-point message delivery feature can be used for managing partition backups. For example, a cluster node, e.g. the
node A 201, can be the owner of a partition, while nodes B-D are backup nodes for thenode A 201. Here, the partition can define that the value of a property x is equal to 1 (“x=1”), which can be maintained on both the primaryowner node A 201 and each backup node B-D 202-204. - Then, the
node A 201 can receive a message from theclient 210, which changes the value of the property x to 2 (“x=2”). Thus, thecluster node A 201 may propagate the message, “x=2”, to each backup node B-D 202-204. - As shown in
FIG. 2 , thecluster node A 201 can deliver the message, “x=2”, to the recipient nodes B-D 202-204 in order. Furthermore, when thenode B 202 detects that thenode C 203 is dead or unavailable, thenode B 202 can deliver the message, “x=2”, to thenode D 204 directly, while skipping thenode C 203. Thus, the distributed data grid 200 can maintain a consistent view that the value of x equals to 2. - On the other hand, an alternative approach is that the
cluster node A 201 can deliver the message to each recipient node B-D 202-204 separately, or in parallel (as shown in dotted line inFIG. 2 ). - Unlike the guaranteed multi-point message delivery feature as described in the above, this alternative approach can be problematic in the scenario when a cluster node,
e.g. node C 203 is dead or become unavailable. Since the delivery of the message, “x=2”, to thecluster node C 203 may not go through, the value of x on node C may remain to be 1 without notice. - Thus, using the alternative approach, the distributed data grid 200 may not be able to maintain a consistent view that the value of x equals to 2. This alternative approach can cause inconsistency, in terms of determining the value of x at a later time point, among the different cluster nodes A-D 201-204 in the distributed data grid 200.
- Furthermore, such inconsistency may not be resolved until a full synchronization is performed in the distributed data grid 200. The full synchronization in the distributed data grid 200 can be costly, since it may require the distributed data grid 200 to stop providing services.
- Additionally, the guaranteed multi-point message delivery feature can be used complimentarily with a partition versioning feature supported in the distributed data grid 200. For example, when a
new node E 205 is added into the distributed data grid 200, the distributed data grid 200 can bring the state of the newly addednode E 205 current, based on the partition versioning feature, so that thenode E 205 can start receive new messages based on the guaranteed multi-point message delivery feature, as described above. - Additional descriptions of various embodiments of using partition versioning feature in a distributed data grid 200 are provided in U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR PROVIDING PARTITION PERSISTENT STATE CONSISTENCY IN A DISTRIBUTED DATA GRID”, filed ______, which application is herein incorporated by reference.
- Furthermore, the guaranteed multi-point message delivery feature can be used complimentarily with a poll model that is supported in the distributed data grid 200 for processing incoming messages asynchronizingly.
- Additional descriptions of various embodiments of supporting asynchronized message processing in a distributed data grid 200 are provided in U.S. patent application Ser. No. ______, entitled “SYSTEM AND METHOD FOR SUPPORTING ASYNCHRONIZED MESSAGE PROCESSING IN A DISTRIBUTED DATA GRID”, filed ______, which application is herein incorporated by reference.
-
FIG. 3 is an illustration of creating a chain request message in a distributed data grid in accordance with various embodiments of the invention. As shown inFIG. 3 , acluster node A 301 in the distributeddata grid 300 can create achain request message 320, e.g. using amessaging facility 311 on thecluster node A 301. - The
chain request message 320 can be either initiated by thecluster node A 301, or be created based on anincoming message 310 received by thecluster node A 301. Thechain request message 320 can include aninternal data structure 321 that stores the information about a list of recipient nodes that thechain request message 320 will be delivered to. - As shown in
FIG. 3 , thecluster node A 301 can deliver thechain request message 320 to the different cluster nodes in the distributeddata grid 300, e.g. nodes B-C 302-203, for further processing. -
FIG. 4 is an illustration of handling a chain request message in a distributed data grid in accordance with various embodiments of the invention. As shown inFIG. 4 , a cluster node,e.g. node B 402, in the distributeddata grid 400 can receive achain request message 420 that contains a list of recipient nodes in aninternal data structure 421. - A
messaging facility 412 in thecluster node B 402 can track the delivery of thechain request message 420 to the rest of the recipient nodes in thelist 421. For example, when thecluster node B 402 detects thatnode C 403 is dead or is unavailable, thecluster node B 402 can access theinternal data structure 421 for the list of recipient nodes and find out thatnode D 204 is the next node followingnode C 203. Therefore, thecluster node B 402 can deliver thechain request message 420 to node D 404 accordingly. -
FIG. 5 is an illustration of supporting consistent message delivery when a primary owner node of a partition becomes unavailable in a distributed data grid in accordance with various embodiments of the invention. As shown inFIG. 5 , a distributed data grid 500 can include a plurality of cluster nodes, such as node A-D 501-504. - Initially, a partition can define that the value of a property x is equal to 1 (“x=1”). The partition can be maintained on the primary
owner node A 501 and each backup node B-D 502-504. Then, thenode A 501 can receive a message from theclient 510, which changes the value of the property x to 2 (“x=2”). Thus, themessaging facility 511 in thecluster node A 501 can propagate the message, “x=2”, to each backup node B-D 202-204. - As shown in
FIG. 5 , when theclient 510 detects that the primaryowner node A 501 is dead or unavailable, the message, “x=2”, can be either having already been delivered to thecluster node B 502, or undelivered. - When the message, “x=2”, has already been delivered to the
cluster node B 502, the distributed data grid 500 can guarantee that the message is delivered to the rest of nodes C-D 503-504, and can ensure a consistent view that the value of x is equal to 2. On the other hand when the message, “x=2”, is undelivered before the primaryowner node A 501 becomes unavailable, the distributed data grid 500 maintains the consistent view that the value of x is equal to 1. Thus, the distributed data grid 500 can ensure a consistent view of the value ofx client 510 in either case. - Then, the distributed data grid 500 can provide a new primary owner node, e.g.
cluster node E 505, which can continue maintaining the partition and handle subsequent incoming messages from theclient 510. - Alternatively, the
cluster node A 501 can deliver the message, “x=2”, to each recipient node B-D 502-504 separately, or in parallel (as shown in dotted line inFIG. 5 ). Unlike the guaranteed multi-point message delivery feature as described in the above, this alternative approach can be problematic in the scenario when the primaryowner node A 501 is dead or become unavailable. Since the delivery of the message, “x=2”, to the different cluster nodes B-D 502-504 may not go through, the value of x on the cluster nodes B-D 502-504 is not guaranteed to be consistent. -
FIG. 6 illustrates an exemplary flow chart for supporting guaranteed multi-point message delivery in a distributed data grid in accordance with an embodiment of the invention. As shown inFIG. 6 , atstep 601, a messaging facility on a cluster node in the distributed data grid can receive an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid. Then, atstep 602, the messaging facility can be configured to deliver the incoming message to the plurality of nodes according to an order in a list. Furthermore, atstep 603, a node in the plurality of nodes can skip a next node in the list for delivering the incoming message, when the next node is dead or unavailable. - The present invention may be conveniently implemented using one or more conventional general purpose or specialized digital computer, computing device, machine, or microprocessor, including one or more processors, memory and/or computer readable storage media programmed according to the teachings of the present disclosure. Appropriate software coding can readily be prepared by skilled programmers based on the teachings of the present disclosure, as will be apparent to those skilled in the software art.
- In some embodiments, the present invention includes a computer program product which is a storage medium or computer readable medium (media) having instructions stored thereon/in which can be used to program a computer to perform any of the processes of the present invention. The storage medium can include, but is not limited to, any type of disk including floppy disks, optical discs, DVD, CD-ROMs, microdrive, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, DRAMs, VRAMs, flash memory devices, magnetic or optical cards, nanosystems (including molecular memory ICs), or any type of media or device suitable for storing instructions and/or data.
- The foregoing description of the present invention has been provided for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations will be apparent to the practitioner skilled in the art. The embodiments were chosen and described in order to best explain the principles of the invention and its practical application, thereby enabling others skilled in the art to understand the invention for various embodiments and with various modifications that are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalence.
Claims (20)
1. A method for supporting consistent message delivery in a distributed data grid operating on one or more microprocessors, comprising:
receiving an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid;
delivering the incoming message to the plurality of nodes according to an order in a list; and
allowing a node in the plurality of nodes to skip a next node in the list for delivering the incoming message, when the next node is dead or unavailable.
2. The method according to claim 1 , further comprising:
allowing the incoming message to include a request from a client; and
providing a response to the client after the incoming message is delivered to the plurality of nodes in the distributed data grid.
3. The method according to claim 1 , further comprising:
allowing the node to deliver the incoming message to a node that is next to the node that is dead or unavailable in the list.
4. The method according to claim 1 , further comprising:
allowing each node in the plurality of nodes to keep track of the incoming message as the incoming message is delivered accordingly to the list.
5. The method according to claim 1 , further comprising:
creating a chain request message that stores information about the list in an internal data structure based on the incoming message.
6. The method according to claim 5 , further comprising:
allowing a node in the distributed data grid to obtain the information about the list from the internal data structure in the chain request message.
7. The method according to claim 1 , further comprising:
allowing a first node in the list to be a primary owner of a partition and other nodes in the list of nodes to be backup nodes for the first node.
8. The method according to claim 7 , further comprising:
propagating an update of the partition from the primary owner of the partition to the backup nodes.
9. The method according to claim 1 , further comprising:
performing a full synchronization on the plurality of nodes in the distributed data grid.
10. The method according to claim 1 , further comprising:
configuring a new node in the distributed data grid to receive the incoming message.
11. A system for supporting guaranteed multi-point delivery in a distributed data grid, comprising:
one or more microprocessors;
a messaging facility in the distributed data grid running on the one or more microprocessors, wherein the messaging facility operates to perform the steps of
receiving an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid;
delivering the incoming message to the plurality of nodes according to an order in a list; and
allowing a node in the plurality of nodes to skip a next node in the list for delivering the incoming message, when the next node is dead or unavailable.
12. The system according to claim 11 , wherein:
the incoming message includes a request from a client, and wherein a response is provided to the client after the incoming message is delivered to the plurality of nodes in the distributed data grid.
13. The system according to claim 11 , wherein:
the node operates to deliver the incoming message to a node that is next to the node that is dead or unavailable in the list.
14. The system according to claim 11 , wherein:
each node in the plurality of nodes operates to keep track of the incoming message as it is delivered accordingly to the list.
15. The system according to claim 11 , wherein:
a first node in the plurality of nodes operates to create a chain request message that stores information about the list in an internal data structure based on the incoming message
16. The system according to claim 15 , wherein:
a second node in the distributed data grid operates to obtain the information about the list from the internal data structure in the chain request message.
17. The system according to claim 11 , wherein:
a first node in the list operates to be a primary owner of a partition and other nodes in the list of nodes operates to be backup nodes for the first node.
18. The system according to claim 17 , wherein:
an update of the partition is propagated from the primary owner of the partition to the backup nodes.
19. The system according to claim 11 , wherein:
a new node in the distributed data grid is confiured to receive the incoming message.
20. A non-transitory machine readable storage medium having instructions stored thereon that when executed cause a system to perform the steps of:
receiving an incoming message that is adaptive to be delivered to a plurality of nodes in the distributed data grid;
delivering the incoming message to the plurality of nodes according to an order in a list; and
allowing a node in the plurality of nodes to skip a next node in the list for delivering the incoming message, when the next node is dead or unavailable.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/671,369 US20140108532A1 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting guaranteed multi-point delivery in a distributed data grid |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201261714100P | 2012-10-15 | 2012-10-15 | |
US13/671,369 US20140108532A1 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting guaranteed multi-point delivery in a distributed data grid |
Publications (1)
Publication Number | Publication Date |
---|---|
US20140108532A1 true US20140108532A1 (en) | 2014-04-17 |
Family
ID=50476346
Family Applications (11)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/671,433 Active 2033-03-10 US9787561B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting a selection service in a server environment |
US13/671,369 Abandoned US20140108532A1 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting guaranteed multi-point delivery in a distributed data grid |
US13/671,376 Active US8930316B2 (en) | 2012-10-15 | 2012-11-07 | System and method for providing partition persistent state consistency in a distributed data grid |
US13/671,441 Active US8874811B2 (en) | 2012-10-15 | 2012-11-07 | System and method for providing a flexible buffer management interface in a distributed data grid |
US13/671,356 Active US8930409B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting named operations in a distributed data grid |
US13/671,479 Active 2033-04-26 US9548912B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting smart buffer management in a distributed data grid |
US13/671,481 Active 2033-06-26 US9083614B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting out-of-order message processing in a distributed data grid |
US13/671,395 Active US8898680B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting asynchronous message processing in a distributed data grid |
US13/671,408 Active US8954391B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting transient partition consistency in a distributed data grid |
US13/671,420 Active 2033-12-02 US9246780B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting port multiplexing in a server environment |
US15/688,151 Active US10050857B2 (en) | 2012-10-15 | 2017-08-28 | System and method for supporting a selection service in a server environment |
Family Applications Before (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/671,433 Active 2033-03-10 US9787561B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting a selection service in a server environment |
Family Applications After (9)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/671,376 Active US8930316B2 (en) | 2012-10-15 | 2012-11-07 | System and method for providing partition persistent state consistency in a distributed data grid |
US13/671,441 Active US8874811B2 (en) | 2012-10-15 | 2012-11-07 | System and method for providing a flexible buffer management interface in a distributed data grid |
US13/671,356 Active US8930409B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting named operations in a distributed data grid |
US13/671,479 Active 2033-04-26 US9548912B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting smart buffer management in a distributed data grid |
US13/671,481 Active 2033-06-26 US9083614B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting out-of-order message processing in a distributed data grid |
US13/671,395 Active US8898680B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting asynchronous message processing in a distributed data grid |
US13/671,408 Active US8954391B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting transient partition consistency in a distributed data grid |
US13/671,420 Active 2033-12-02 US9246780B2 (en) | 2012-10-15 | 2012-11-07 | System and method for supporting port multiplexing in a server environment |
US15/688,151 Active US10050857B2 (en) | 2012-10-15 | 2017-08-28 | System and method for supporting a selection service in a server environment |
Country Status (1)
Country | Link |
---|---|
US (11) | US9787561B2 (en) |
Families Citing this family (29)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9292569B2 (en) * | 2012-10-02 | 2016-03-22 | Oracle International Corporation | Semi-join acceleration |
US9787561B2 (en) | 2012-10-15 | 2017-10-10 | Oracle International Corporation | System and method for supporting a selection service in a server environment |
US9288225B1 (en) * | 2013-04-17 | 2016-03-15 | Ca, Inc. | Server port sharing based on shared socket |
US9894143B1 (en) * | 2013-11-06 | 2018-02-13 | Amazon Technologies, Inc. | Pre-processing and processing pipeline for queue client |
US20150163179A1 (en) * | 2013-12-09 | 2015-06-11 | Hewlett-Packard Development Company, L.P. | Execution of a workflow that involves applications or services of data centers |
US9898414B2 (en) | 2014-03-28 | 2018-02-20 | Oracle International Corporation | Memory corruption detection support for distributed shared memory applications |
US10133489B2 (en) * | 2014-09-16 | 2018-11-20 | Oracle International Corporation | System and method for supporting a low contention queue in a distributed data grid |
CN104410607B (en) * | 2014-11-06 | 2018-05-22 | 上海汽车集团股份有限公司 | Engine investigating method based on thread pool |
CN105808790B (en) * | 2014-12-29 | 2019-04-05 | 联合汽车电子有限公司 | Digital data recording system for new-energy automobile power drive system malfunction elimination |
US10069938B1 (en) * | 2015-03-30 | 2018-09-04 | EMC IP Holding Company LLC | Returning identifiers in default query responses |
CN105118012A (en) * | 2015-08-25 | 2015-12-02 | 杭州环特环保设备有限公司 | Urban management intelligent terminal service platform and usage method thereof |
US10164106B2 (en) | 2016-12-29 | 2018-12-25 | Taiwan Semiconductor Manufacturing Co., Ltd. | Semiconductor device and a method for fabricating the same |
US10282115B2 (en) | 2017-04-13 | 2019-05-07 | International Business Machines Corporation | Object synchronization in a clustered system |
CN107506403B (en) * | 2017-08-03 | 2020-06-23 | 国网江苏省电力公司信息通信分公司 | Gridding information system and data synchronous changing method |
CN107508881A (en) * | 2017-08-21 | 2017-12-22 | 叶军 | A kind of service locking method for distributed file system |
CN109660569B (en) * | 2017-10-10 | 2021-10-15 | 武汉斗鱼网络科技有限公司 | Multitask concurrent execution method, storage medium, device and system |
US10467139B2 (en) | 2017-12-29 | 2019-11-05 | Oracle International Corporation | Fault-tolerant cache coherence over a lossy network |
US10452547B2 (en) | 2017-12-29 | 2019-10-22 | Oracle International Corporation | Fault-tolerant cache coherence over a lossy network |
CN108416005A (en) * | 2018-02-28 | 2018-08-17 | 湖北邮电规划设计有限公司 | A kind of telecommunications grid data processing method and system |
CN108810071A (en) * | 2018-03-30 | 2018-11-13 | 广东睿江云计算股份有限公司 | A method of the Apache Server cluster load balance based on process migration |
CN108733823A (en) * | 2018-05-22 | 2018-11-02 | 浪潮软件股份有限公司 | A kind of system and method that ORACLE database intranet and extranet data double-ways are synchronized |
US11509531B2 (en) | 2018-10-30 | 2022-11-22 | Red Hat, Inc. | Configuration sharing and validation for nodes in a grid network |
CN109495587B (en) * | 2018-12-25 | 2021-08-24 | 思必驰科技股份有限公司 | Language model synchronization method and system for voice conversation platform |
US11126365B2 (en) * | 2019-03-11 | 2021-09-21 | Commvault Systems, Inc. | Skipping data backed up in prior backup operations |
CN111338914A (en) * | 2020-02-10 | 2020-06-26 | 华为技术有限公司 | Fault notification method and related equipment |
US12111842B2 (en) * | 2021-03-15 | 2024-10-08 | Nvidia Corporation | Low latency communications for nodes in replication relationships |
US11687565B2 (en) * | 2021-03-22 | 2023-06-27 | Sap Se | Asynchronous data replication in a multiple availability zone cloud platform |
US11874765B2 (en) | 2021-05-28 | 2024-01-16 | International Business Machines Corporation | Buffer pool management |
CN115767448B (en) * | 2022-12-07 | 2024-06-18 | 京东科技信息技术有限公司 | Message sending method, device, equipment and storage medium |
Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20050249215A1 (en) * | 2004-02-19 | 2005-11-10 | Kelsey Richard A | Directing packets in a mesh network |
US7058055B2 (en) * | 2001-07-17 | 2006-06-06 | Smartmatic Corp. | Method of device-to-device communications in hybrid distributed device control networks |
US7139925B2 (en) * | 2002-04-29 | 2006-11-21 | Sun Microsystems, Inc. | System and method for dynamic cluster adjustment to node failures in a distributed data system |
US20090116484A1 (en) * | 2005-09-08 | 2009-05-07 | Matsushita Electric Industrial Co., Ltd. | Parallelizing Peer-to-Peer Overlays Using Multi-Destination Routing |
US20090177914A1 (en) * | 2002-02-22 | 2009-07-09 | Richard A. Angell | Clustering Infrastructure System and Method |
US20100265945A1 (en) * | 2009-04-20 | 2010-10-21 | Yigal Bejerano | Method and apparatus for fault-resilient multicast and unicast in transport networks |
US7936753B1 (en) * | 2007-11-30 | 2011-05-03 | Qlogic, Corporation | Method and system for reliable multicast |
US20110225121A1 (en) * | 2010-03-11 | 2011-09-15 | Yahoo! Inc. | System for maintaining a distributed database using constraints |
US20110225120A1 (en) * | 2010-03-11 | 2011-09-15 | Yahoo! Inc. | System for maintaining a distributed database using leases |
US20120113896A1 (en) * | 2010-11-10 | 2012-05-10 | Telcordia Technologies, Inc. | Skip Ahead Routing in Wireless Ad Hoc Networks |
US8249072B2 (en) * | 2009-03-12 | 2012-08-21 | Oracle America, Inc. | Scalable interface for connecting multiple computer systems which performs parallel MPI header matching |
US8271980B2 (en) * | 2004-11-08 | 2012-09-18 | Adaptive Computing Enterprises, Inc. | System and method of providing system jobs within a compute environment |
US8290919B1 (en) * | 2010-08-27 | 2012-10-16 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US20120278398A1 (en) * | 2011-04-27 | 2012-11-01 | Lowekamp Bruce B | System and method for reliable distributed communication with guaranteed service levels |
US8392368B1 (en) * | 2010-08-27 | 2013-03-05 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US20130128726A1 (en) * | 2006-05-17 | 2013-05-23 | Rajant Corporation | System and method for packet delivery backtracking |
US20130262632A1 (en) * | 2008-03-20 | 2013-10-03 | Tajitshu Transfer Limited Liability Company | Redundant data forwarding storage |
US8612386B2 (en) * | 2011-02-11 | 2013-12-17 | Alcatel Lucent | Method and apparatus for peer-to-peer database synchronization in dynamic networks |
US20140016457A1 (en) * | 2011-03-31 | 2014-01-16 | Telefonaktiebolaget L M Ericsson (Publ) | Technique for Operating a Network Node |
US8768981B1 (en) * | 2010-08-27 | 2014-07-01 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US20140219209A1 (en) * | 2011-10-13 | 2014-08-07 | Fujitsu Limited | Node device and communication method |
Family Cites Families (66)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5301350A (en) * | 1989-10-10 | 1994-04-05 | Unisys Corporation | Real time storage/retrieval subsystem for document processing in banking operations |
US5014265A (en) | 1989-11-30 | 1991-05-07 | At&T Bell Laboratories | Method and apparatus for congestion control in a data network |
US5561785A (en) * | 1992-10-29 | 1996-10-01 | International Business Machines Corporation | System for allocating and returning storage and collecting garbage using subpool of available blocks |
US5784698A (en) | 1995-12-05 | 1998-07-21 | International Business Machines Corporation | Dynamic memory allocation that enalbes efficient use of buffer pool memory segments |
US6151688A (en) * | 1997-02-21 | 2000-11-21 | Novell, Inc. | Resource management in a clustered computer system |
US6070202A (en) | 1998-05-11 | 2000-05-30 | Motorola, Inc. | Reallocation of pools of fixed size buffers based on metrics collected for maximum number of concurrent requests for each distinct memory size |
US6236999B1 (en) | 1998-11-05 | 2001-05-22 | Bea Systems, Inc. | Duplicated naming service in a distributed processing system |
US6826597B1 (en) | 1999-03-17 | 2004-11-30 | Oracle International Corporation | Providing clients with services that retrieve data from data sources that do not necessarily support the format required by the clients |
WO2002010917A1 (en) | 2000-07-27 | 2002-02-07 | Bea Systems, Inc. | System and method for concentration and load-balancing of requests |
EP1265445A3 (en) * | 2001-06-08 | 2007-05-30 | The Distribution Systems Research Institute | Terminal-to-terminal communication connection control system for IP full service |
US6823437B2 (en) | 2002-07-11 | 2004-11-23 | International Business Machines Corporation | Lazy deregistration protocol for a split socket stack |
US7124211B2 (en) | 2002-10-23 | 2006-10-17 | Src Computers, Inc. | System and method for explicit communication of messages between processes running on different nodes in a clustered multiprocessor system |
ITMI20022347A1 (en) | 2002-11-06 | 2004-05-07 | Pasqua Roberto Della | COMMUNICATION METHOD WITH RESPONSE TIME REDUCED IN |
US7444620B2 (en) | 2003-02-28 | 2008-10-28 | Bea Systems, Inc. | Systems and methods for a common runtime container framework |
US8621031B2 (en) * | 2003-04-29 | 2013-12-31 | Oracle International Corporation | Method and apparatus using connection pools in communication networks |
US7594236B2 (en) * | 2004-06-28 | 2009-09-22 | Intel Corporation | Thread to thread communication |
US7610400B2 (en) | 2004-11-23 | 2009-10-27 | Juniper Networks, Inc. | Rule-based networking device |
US8849781B2 (en) | 2004-12-13 | 2014-09-30 | Sap Ag | Adaptive software component techniques |
US8019855B2 (en) | 2004-12-22 | 2011-09-13 | Lg Electronics Inc. | Method and apparatus interfacing between an application and a library of a master for network managing |
US20060230128A1 (en) | 2004-12-22 | 2006-10-12 | Lg Electronics Inc. | Method and apparatus interfacing for controlling a device between an application and a library of a master on home network |
US7945676B2 (en) | 2005-03-10 | 2011-05-17 | International Business Machines Corporation | Processing requests transmitted using a first communication protocol directed to an application that uses a second communication protocol |
US7801912B2 (en) | 2005-12-29 | 2010-09-21 | Amazon Technologies, Inc. | Method and apparatus for a searchable data service |
US8464265B2 (en) * | 2006-04-03 | 2013-06-11 | Secure64 Software | Method and system for reallocating computational resources using resource reallocation enabling information |
CN102684988B (en) | 2006-04-26 | 2015-02-11 | 日本电信电话株式会社 | Load control device and method thereof |
US8095625B2 (en) | 2007-02-28 | 2012-01-10 | Red Hat, Inc. | Directory server plug-in call ordering |
US7904493B2 (en) * | 2007-03-30 | 2011-03-08 | Sap Ag | Method and system for object age detection in garbage collection heaps |
US20080301135A1 (en) | 2007-05-29 | 2008-12-04 | Bea Systems, Inc. | Event processing query language using pattern matching |
US8001553B2 (en) * | 2007-06-25 | 2011-08-16 | Microsoft Corporation | Aggregate computer system via coupling of computing machines |
US7707513B2 (en) * | 2008-02-06 | 2010-04-27 | Oracle International Corporation | System for streaming data from server to multiple slave browser windows in client computer through a master browser window |
DK2248324T3 (en) * | 2008-02-20 | 2019-01-07 | Nabto Aps | PROCEDURE AND SYSTEM TO PROVIDE CONNECTIVITY BETWEEN CLIENTS CONNECTED TO THE INTERNET. |
EP2096793B1 (en) | 2008-02-28 | 2014-11-19 | Alcatel Lucent | Systems and methods for automatic configuration of customer premises equipment |
US7792879B2 (en) * | 2008-03-11 | 2010-09-07 | Intel Corporation | Efficient heap utilization and partitioning |
JP4586873B2 (en) | 2008-03-28 | 2010-11-24 | セイコーエプソン株式会社 | Socket management apparatus and method |
US8990812B2 (en) | 2008-07-07 | 2015-03-24 | Infosys Limited | Task decomposition with throttled message processing in a heterogeneous environment |
US8261323B2 (en) | 2008-07-11 | 2012-09-04 | International Business Machines Corporation | Managing logical sockets |
US8412768B2 (en) | 2008-07-11 | 2013-04-02 | Ball Gaming, Inc. | Integration gateway |
US9176754B2 (en) | 2008-07-16 | 2015-11-03 | Google Inc. | Method and system for executing applications using native code modules |
US8201176B2 (en) * | 2008-08-06 | 2012-06-12 | International Business Machines Corporation | Detecting the starting and ending of a task when thread pooling is employed |
US8028113B2 (en) | 2008-08-15 | 2011-09-27 | International Business Machines Corporation | Methods and systems for deadlock-free allocation of memory |
US8854680B2 (en) | 2008-09-11 | 2014-10-07 | Konica Minolta Laboratory U.S.A., Inc. | Systems and methods for optimal memory allocation units |
US8286194B2 (en) * | 2008-11-19 | 2012-10-09 | International Business Machines Corporation | Coupling state aware systems |
CN101739298B (en) | 2008-11-27 | 2013-07-31 | 国际商业机器公司 | Shared cache management method and system |
US20120144025A1 (en) | 2008-12-23 | 2012-06-07 | Telefonaktiebolaget L.M. Ericsson (Publ) | Method and an Arrangement For Enabling User Traffic Classification Configuration |
US20110004701A1 (en) * | 2008-12-29 | 2011-01-06 | Debashish Panda | Provisioning highly available services for integrated enterprise and communication |
US20100287216A1 (en) | 2009-05-07 | 2010-11-11 | Tatu Ylonen Oy Ltd | Grouped space allocation for copied objects |
US8495643B2 (en) | 2009-06-30 | 2013-07-23 | International Business Machines Corporation | Message selection based on time stamp and priority in a multithreaded processor |
US20110137991A1 (en) * | 2009-12-01 | 2011-06-09 | Lester Paul Russell | Systems and methods for management and collaboration in a private network |
US8539192B2 (en) * | 2010-01-08 | 2013-09-17 | International Business Machines Corporation | Execution of dataflow jobs |
US9009112B2 (en) | 2010-03-15 | 2015-04-14 | Microsoft Technology Licensing, Llc | Reorganization of data under continuous workload |
US8259571B1 (en) * | 2010-03-26 | 2012-09-04 | Zscaler, Inc. | Handling overlapping IP addresses in multi-tenant architecture |
US9116946B2 (en) | 2010-04-02 | 2015-08-25 | Scalebase Inc. | System and method for interacting with a plurality of data sources |
CN102906706A (en) | 2010-05-24 | 2013-01-30 | 索尼电脑娱乐公司 | Information processing device and information processing method |
US8667171B2 (en) * | 2010-05-28 | 2014-03-04 | Microsoft Corporation | Virtual data center allocation with bandwidth guarantees |
US8687636B1 (en) | 2010-06-02 | 2014-04-01 | Marvell Israel (M.I.S.L) Ltd. | Extended policy control list keys having backwards compatibility |
US9164806B2 (en) | 2011-01-28 | 2015-10-20 | Oracle International Corporation | Processing pattern framework for dispatching and executing tasks in a distributed computing grid |
US9081839B2 (en) | 2011-01-28 | 2015-07-14 | Oracle International Corporation | Push replication for use with a distributed data grid |
US9262229B2 (en) | 2011-01-28 | 2016-02-16 | Oracle International Corporation | System and method for supporting service level quorum in a data grid cluster |
US9477536B2 (en) | 2011-05-16 | 2016-10-25 | Oracle International Corporation | System and method for providing a messaging application program interface |
US8447793B2 (en) * | 2011-06-04 | 2013-05-21 | International Business Machines Corporation | Efficient remembered set for region-based garbage collectors |
WO2012173541A1 (en) * | 2011-06-16 | 2012-12-20 | Telefonaktiebolaget L M Ericsson (Publ) | A client and a server and methods thereof for data management relating to shared data storage |
US9436748B2 (en) | 2011-06-23 | 2016-09-06 | Simplivity Corporation | Method and apparatus for distributed configuration management |
US9621409B2 (en) | 2011-09-15 | 2017-04-11 | Oracle International Corporation | System and method for handling storage events in a distributed data grid |
US8918602B2 (en) | 2011-09-19 | 2014-12-23 | International Business Machines Corporation | Dynamically altering time to live values in a data cache |
US9519520B2 (en) | 2011-10-25 | 2016-12-13 | Viasat, Inc. | Federated, policy-driven service meshes for distributed software systems |
US9123010B2 (en) * | 2012-06-05 | 2015-09-01 | Apple Inc. | Ledger-based resource tracking |
US9787561B2 (en) | 2012-10-15 | 2017-10-10 | Oracle International Corporation | System and method for supporting a selection service in a server environment |
-
2012
- 2012-11-07 US US13/671,433 patent/US9787561B2/en active Active
- 2012-11-07 US US13/671,369 patent/US20140108532A1/en not_active Abandoned
- 2012-11-07 US US13/671,376 patent/US8930316B2/en active Active
- 2012-11-07 US US13/671,441 patent/US8874811B2/en active Active
- 2012-11-07 US US13/671,356 patent/US8930409B2/en active Active
- 2012-11-07 US US13/671,479 patent/US9548912B2/en active Active
- 2012-11-07 US US13/671,481 patent/US9083614B2/en active Active
- 2012-11-07 US US13/671,395 patent/US8898680B2/en active Active
- 2012-11-07 US US13/671,408 patent/US8954391B2/en active Active
- 2012-11-07 US US13/671,420 patent/US9246780B2/en active Active
-
2017
- 2017-08-28 US US15/688,151 patent/US10050857B2/en active Active
Patent Citations (21)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7058055B2 (en) * | 2001-07-17 | 2006-06-06 | Smartmatic Corp. | Method of device-to-device communications in hybrid distributed device control networks |
US20090177914A1 (en) * | 2002-02-22 | 2009-07-09 | Richard A. Angell | Clustering Infrastructure System and Method |
US7139925B2 (en) * | 2002-04-29 | 2006-11-21 | Sun Microsystems, Inc. | System and method for dynamic cluster adjustment to node failures in a distributed data system |
US20050249215A1 (en) * | 2004-02-19 | 2005-11-10 | Kelsey Richard A | Directing packets in a mesh network |
US8271980B2 (en) * | 2004-11-08 | 2012-09-18 | Adaptive Computing Enterprises, Inc. | System and method of providing system jobs within a compute environment |
US20090116484A1 (en) * | 2005-09-08 | 2009-05-07 | Matsushita Electric Industrial Co., Ltd. | Parallelizing Peer-to-Peer Overlays Using Multi-Destination Routing |
US20130128726A1 (en) * | 2006-05-17 | 2013-05-23 | Rajant Corporation | System and method for packet delivery backtracking |
US7936753B1 (en) * | 2007-11-30 | 2011-05-03 | Qlogic, Corporation | Method and system for reliable multicast |
US20130262632A1 (en) * | 2008-03-20 | 2013-10-03 | Tajitshu Transfer Limited Liability Company | Redundant data forwarding storage |
US8249072B2 (en) * | 2009-03-12 | 2012-08-21 | Oracle America, Inc. | Scalable interface for connecting multiple computer systems which performs parallel MPI header matching |
US20100265945A1 (en) * | 2009-04-20 | 2010-10-21 | Yigal Bejerano | Method and apparatus for fault-resilient multicast and unicast in transport networks |
US20110225120A1 (en) * | 2010-03-11 | 2011-09-15 | Yahoo! Inc. | System for maintaining a distributed database using leases |
US20110225121A1 (en) * | 2010-03-11 | 2011-09-15 | Yahoo! Inc. | System for maintaining a distributed database using constraints |
US8290919B1 (en) * | 2010-08-27 | 2012-10-16 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US8392368B1 (en) * | 2010-08-27 | 2013-03-05 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US8768981B1 (en) * | 2010-08-27 | 2014-07-01 | Disney Enterprises, Inc. | System and method for distributing and accessing files in a distributed storage system |
US20120113896A1 (en) * | 2010-11-10 | 2012-05-10 | Telcordia Technologies, Inc. | Skip Ahead Routing in Wireless Ad Hoc Networks |
US8612386B2 (en) * | 2011-02-11 | 2013-12-17 | Alcatel Lucent | Method and apparatus for peer-to-peer database synchronization in dynamic networks |
US20140016457A1 (en) * | 2011-03-31 | 2014-01-16 | Telefonaktiebolaget L M Ericsson (Publ) | Technique for Operating a Network Node |
US20120278398A1 (en) * | 2011-04-27 | 2012-11-01 | Lowekamp Bruce B | System and method for reliable distributed communication with guaranteed service levels |
US20140219209A1 (en) * | 2011-10-13 | 2014-08-07 | Fujitsu Limited | Node device and communication method |
Non-Patent Citations (2)
Title |
---|
Network Working Group, RFC 4728, The Dynamic Source Routing Protocol (DSR) for Mobile Ad Hoc Networks for IPv4 (Feb. 2007) * |
Tanenbaum, Andrew S., Computer Networks, Fourth Edition, 2003, Prentice Hall PTR, Chapter 5.2.7 Broadcast Routing * |
Also Published As
Publication number | Publication date |
---|---|
US9787561B2 (en) | 2017-10-10 |
US20140108533A1 (en) | 2014-04-17 |
US9548912B2 (en) | 2017-01-17 |
US20140108358A1 (en) | 2014-04-17 |
US20140108347A1 (en) | 2014-04-17 |
US8930316B2 (en) | 2015-01-06 |
US20140108645A1 (en) | 2014-04-17 |
US8898680B2 (en) | 2014-11-25 |
US8954391B2 (en) | 2015-02-10 |
US8874811B2 (en) | 2014-10-28 |
US20140108450A1 (en) | 2014-04-17 |
US20140108681A1 (en) | 2014-04-17 |
US20140108760A1 (en) | 2014-04-17 |
US20140108664A1 (en) | 2014-04-17 |
US20140109110A1 (en) | 2014-04-17 |
US10050857B2 (en) | 2018-08-14 |
US9246780B2 (en) | 2016-01-26 |
US9083614B2 (en) | 2015-07-14 |
US20170359240A1 (en) | 2017-12-14 |
US8930409B2 (en) | 2015-01-06 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US20140108532A1 (en) | System and method for supporting guaranteed multi-point delivery in a distributed data grid | |
US10817478B2 (en) | System and method for supporting persistent store versioning and integrity in a distributed data grid | |
US10423643B2 (en) | System and method for supporting resettable acknowledgements for synchronizing data in a distributed data grid | |
JP6615761B2 (en) | System and method for supporting asynchronous calls in a distributed data grid | |
US10243869B2 (en) | System and method for providing a distributed queue in a distributed data grid | |
CN105830029B (en) | system and method for supporting adaptive busy-wait in a computing environment | |
WO2015179092A1 (en) | System and method for supporting a distributed data structure in a distributed data grid |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: ORACLE INTERNATIONAL CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:LEE, ROBERT H.;GLEYZER, GENE;REEL/FRAME:029636/0029 Effective date: 20130107 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |